Abstract
When taking photos in the outdoors, e.g., playgrounds, gardens and zoos, fences are often inevitable to interfere with the perception of the background scenes. Thus it is an interesting topic on removing fences from a single image, which is dubbed image de-fencing. Generally, image de-fencing can be tackled within the two-stage framework, i.e., first detecting fences and then restoring background image. In existing methods, detecting fences mask cannot guarantee accuracy due to the diverse patterns of fences. Meanwhile, state-of-the-art supervised learning-based restoration methods usually fail in well handling inaccurate fences mask, yielding artifacts in restored background image. In this paper, we propose a semi-supervised framework for removing fences from a single image, where a simple yet effective recurrent network is proposed for supervised learning to detect fences mask and unsupervised learning is employed for robust restoration of background image. Specifically, we propose a recurrent network for fences mask detection, where convolutional LSTM is adopted for progressive detection. Instead of strict fences mask detection using standard \(\ell _1\) loss, we adopt an asymmetric loss to make detected fences mask be more inclusive of true fences. Then motivated by the success of deep image prior (DIP) for several image restoration tasks, we propose to employ unsupervised learning of three DIP networks for modeling background image, fences layer and fences mask, respectively. Moreover, we propose to adopt Laplacian smoothness loss function for refining the detected fences mask, making the restoration of background image be more robust to detection errors of fences mask. Experimental results validate the effectiveness of our semi-supervised image de-fencing approach.
The first author is a second year master student.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Barnes, C., Shechtman, E., Finkelstein, A., Goldman, D.B.: PatchMatch: a randomized correspondence algorithm for structural image editing. ACM Trans. Graph. (ToG) 28, 24 (2009)
Criminisi, A., Pérez, P., Toyama, K.: Region filling and object removal by exemplar-based image inpainting. IEEE Trans. Image Process. 13(9), 1200–1212 (2004)
Divyanshu, G., Shorya, J., Utkarsh, T., Pratik, C., Wang, L.: Fully automated image de-fencing using conditional generative adversarial networks. arXiv preprint arXiv:1908.06837 (2019)
Du, C., Kang, B., Xu, Z., Dai, J., Nguyen, T.: Accurate and efficient video de-fencing using convolutional neural networks and temporal information. In: 2018 IEEE International Conference on Multimedia and Expo (ICME), pp. 1–6. IEEE (2018)
Farid, M.S., Mahmood, A., Grangetto, M.: Image de-fencing framework with hybrid inpainting algorithm. SIViP 10(7), 1193–1201 (2016). https://doi.org/10.1007/s11760-016-0876-7
Gandelsman, Y., Shocher, A., Irani, M.: “Double-DIP”: unsupervised image decomposition via coupled deep-image-priors, June 2019
Guo, S., Yan, Z., Zhang, K., Zuo, W., Zhang, L.: Toward convolutional blind denoising of real photographs. In: 2019 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2019)
Jonna, S., Nakka, K.K., Sahay, R.R.: My camera can see through fences: a deep learning approach for image de-fencing. In: 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), pp. 261–265. IEEE (2015)
Jonna, S., Nakka, K.K., Sahay, R.R.: Deep learning based fence segmentation and removal from an image using a video sequence. In: Hua, G., Jégou, H. (eds.) ECCV 2016. LNCS, vol. 9915, pp. 836–851. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-49409-8_68
Jonna, S., Voleti, V.S., Sahay, R.R., Kankanhalli, M.S.: A multimodal approach for image de-fencing and depth inpainting. In: 2015 Eighth International Conference on Advances in Pattern Recognition (ICAPR), pp. 1–6. IEEE (2015)
Khasare, V.S., Sahay, R.R., Kankanhalli, M.S.: Seeing through the fence: image de-fencing using a video sequence. In: 2013 IEEE International Conference on Image Processing, pp. 1351–1355. IEEE (2013)
Kumar, V., Mukherjee, J., Mandal, S.K.D.: Image defencing via signal demixing. In: Proceedings of the Tenth Indian Conference on Computer Vision, Graphics and Image Processing, p. 11. ACM (2016)
Levin, A., Lischinski, D., Weiss, Y.: A closed-form solution to natural image matting. IEEE Trans. Pattern Anal. Mach. Intell. 30(2), 228–242 (2007)
Liu, Y., Belkina, T., Hays, J., Lublinerman, R.: Image defencing. In: Proceedings of CVPR (2008)
Paszke, A., et al.: PyTorch: an imperative style, high-performance deep learning library. In: Advances in Neural Information Processing Systems, pp. 8024–8035 (2019)
Pathak, D., Krahenbuhl, P., Donahue, J., Darrell, T., Efros, A.A.: Context encoders: feature learning by inpainting. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2536–2544 (2016)
Shi, X., Chen, Z., Wang, H., Yeung, D.Y., Wong, W.K., Woo, W.c.: Convolutional LSTM network: a machine learning approach for precipitation nowcasting. In: Advances in Neural Information Processing Systems, pp. 802–810 (2015)
Ulyanov, D., Vedaldi, A., Lempitsky, V.: Deep image prior. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9446–9454 (2018)
Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600–612 (2004)
Xie, C., et al.: Image inpainting with learnable bidirectional attention maps. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 8858–8867 (2019)
Xie, J., Xu, L., Chen, E.: Image denoising and inpainting with deep neural networks. In: Advances in Neural Information Processing Systems, pp. 341–349 (2012)
Yan, Z., Li, X., Li, M., Zuo, W., Shan, S.: Shift-Net: image inpainting via deep feature rearrangement. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision – ECCV 2018. LNCS, vol. 11218, pp. 3–19. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01264-9_1
Zhou, B., Lapedriza, A., Khosla, A., Oliva, A., Torralba, A.: Places: a 10 million image database for scene recognition. IEEE Trans. Pattern Anal. Mach. Intell. 40, 1452–1464 (2017)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Shang, W., Zhu, P., Ren, D. (2020). Semi-supervised Learning to Remove Fences from a Single Image. In: Peng, Y., et al. Pattern Recognition and Computer Vision. PRCV 2020. Lecture Notes in Computer Science(), vol 12305. Springer, Cham. https://doi.org/10.1007/978-3-030-60633-6_7
Download citation
DOI: https://doi.org/10.1007/978-3-030-60633-6_7
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-60632-9
Online ISBN: 978-3-030-60633-6
eBook Packages: Computer ScienceComputer Science (R0)