Elsevier

Displays

Volume 69, September 2021, 102028
Displays

Image inpainting based on deep learning: A review

https://doi.org/10.1016/j.displa.2021.102028Get rights and content

Highlights

  • Classify image inpainting methods based on deep learning from a new perspective.

  • Summarizes the current research status in the field of image inpainting.

  • Select some representative image inpainting methods for comparison and analysis.

  • The research direction and development trend of image inpainting are prospected.

Abstract

Image inpainting aims to restore the pixel features of damaged parts in incomplete image and plays a key role in many computer vision tasks. Image inpainting technology based on deep learning is a major current research hotspot. To deeply understand related methods and technologies, this article combs and summarizes the latest research status in this field. Firstly, we summarize inpainting methods of different types of neural network structure based on deep learning, then analyze and study important technical improvement mechanisms. In addition, various algorithms are comprehensively reviewed from the aspects of model network structure and restoration methods. And we select some representative image inpainting methods for comparison and analysis. Finally, the current problems of image inpainting are summarized, and the future development trend and research direction are prospected.

Introduction

Image inpainting is a technology that aims to restore the damaged part of pixel features in the incomplete image, and then reconstruct and generate high-quality and deep semantic approximation to the original image. In recent years, the implementation of artificial intelligence scientific research and deep learning related technologies has achieved vigorous development along with the substantial increase in computer computing power, which has brought important promotion and improvement to science technology and the quality of human life. Image inpainting technology based on deep learning plays an important role in many computer vision applications [1] (such as target removal in image editing technology, old photo restoration, defective cultural relics and font restoration, facial restoration, etc.) and has become a major research hotspot in computer vision.

In traditional image inpainting technology, the related methods are mostly machine learning algorithms based on statistical probability. Marcelo Bertalmio et al. [2] proposed a Markov Random Field (MRF) image inpainting algorithm on the basis of structure migration mapping statistics and multi-directional features for large-scale damaged image restoration. The inpainting algorithm is mainly used for target removal, which can better maintain the continuity of repaired image structure and the consistency between adjacent pixels. Shen et al. [3] proposed an improved sparse representation inpainting algorithm in the light of similar matching blocks, which achieved good restoration effects in the inpainting of color damaged images with multiple damaged shapes in a small area. Tsai et al. [4] proposed a matrix completion method with automatic rank estimation based on low-rank decomposition is used to extract restored high-quality images from images with different degrees of low sampling rate. Considering the above, Bertalmio et al. [5] introduced conjugate gradient method based on riemannian manifold to optimize matrix completion and combined convolution neural network to preprocess sample images. The method of block processing is adopted to further save operation space and improve the quality of restored images. These methods have made improvements on traditional machine learning algorithms to promote image inpainting effects in different ways. However, compared with the deep learning image inpainting technology, the restored images generated by the traditional methods when processing large image data in damaged areas often lack semantic consistency and texture structure coherence.

Around 2014, with the rise of deep learning, image inpainting technology has been deeply applied in the field of computer vision. Many researchers have continuously carried out in-depth research on the problem of high-quality image inpainting at the level of generating semantic understanding [6], [7], [8], and then a large number of classical image inpainting methods based on deep learning have emerged [9], [10], [11], [12], [13], [14], [15], [16], [17], [18], Some scholars have also summarized the work in this field. Recent work [1], [19] summarized the image inpainting technology based on deep learning. Omar Elharrouss et al. [1] divided the image inpainting model methods proposed in some classic papers into three categories from a global perspective, namely, sequence-based methods, CNN-based methods and GAN-based methods. Qiang et al. [19] summarized the main image inpainting methods based on deep learning in recent years and classified the existing methods into three network structure types of image inpainting methods based on convolutional autoencoder network, generative adversarial network and recurrent neural network according to the inpainting network structure.

In the past few years, deep learning has made great breakthroughs in the field of image inpainting. A hybrid network model based on the combination of autoencoder and Generative Adversarial Network (GAN) [20], [21], [22], [23], an improved autoencoder based on attention mechanism [24], [25], [26], [27], [28], [29], and improved shared codec network layer based on coarse-to-fine network [30], [31], [32], [33], [34], [35] have emerged, which gradually repair damaged images at the semantic level. Based on the above work, this paper makes a more comprehensive and detailed summary of image inpainting related network models based on deep learning in recent years, aiming to provide a more comprehensive and in-depth learning perspective for subsequent research in related fields.

Section snippets

Image inpainting tasks

Current image inpainting research mainly includes tasks such as repairing rectangular block mask, irregular mask, target removal, denoising, remove watermark, remove text, remove scratches, and coloring of old photos [20], [26], [35], [36], [37], [38], [79]. The example effects of above the 8 inpainting tasks are shown in Fig. 1:

Traditional image inpainting

Traditional image inpainting, mainly divided into diffusion-based methods [2], [39], [40], [41] and patch-based methods [42], [43], [44], [45], [46].

Diffusion-based

Single-stage inpainting

The approaches related to single-stage inpainting can classified into two categories: single result inpainting and pluralistic inpainting approaches.

Image inpainting datasets

Currently, due to it is impossible to collect a large number of paired real damaged images, researchers often choose suitable image data set when performing image inpainting experiments, then add corresponding masks to the original data. The most widely used masks mainly include rectangular shaped hole and irregular mask, rectangular shaped hole usually added by experimenters in the center of the image or scattered with multiple small rectangular masks.

One of the most widely used is a testing

Discussion and analysis

Since the birth of the two major generative models of VAE and GAN, various deep learning network models based on generative models have continuously emerged, leading the vigorous development of the entire computer vision [[80], [90], [91], [94], [95], [96]]. Comparing and summarizing the above various types of representative image inpainting methods, we can find:

  • (1)

    In network selection, the image inpainting method based on convolution neural network is still the mainstream method of deep learning

Conclusions

At present, image inpainting technology has become an important branch in field of vision research. Deep learning image inpainting based on generation network gradually become mainstream method. Researchers have continuously innovated and made great progress in generation model selection, network structure design, introduction of prior guidance, discriminator optimization, loss function optimization, etc. However, the following problems still need to be solved urgently:

  • (1)

    The current image

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgment

This work was supported by the National Science Foundation of China (Grant No. 61901436) and the Key Research Program of the Chinese Academy of Sciences (Grant No. XDPB22).

References (96)

  • H. Qin et al.

    Binary neural networks: A survey

    Pattern Recogn.

    (2020)
  • X. Bai et al.

    Edwin Robert Hancock: Adaptive hash retrieval with kernel based similarity

    Pattern Recogn.

    (2018)
  • O. Elharrouss et al.

    Image inpainting: a review

    Neural Process. Lett.

    (2020)
  • Marcelo Bertalmio, Guillermo Sapiro, Vincent Caselles, Coloma Ballester, Image inpainting, in: Proceedings of the 27th...
  • J.H. Shen et al.

    Euler’s elastica and curvaturebased inpainting

    SIAM J. Appl. Math.

    (2003)
  • A. Tsai et al.

    Curve evolution implementation of the Mumford-Shah functional for image segmentation, denoising, interpolation, and magnification

    IEEE Trans. Image Process.

    (2001)
  • M. Bertalmio, A.L. Bertozzi, G. Sapiro, Navier-stokes, fluid dynamics, and image and video inpainting, in: Proceedings...
  • X. Ning et al.

    BULDP: biomimetic uncorrelated locality discriminant projection for feature extraction in face recognition

    IEEE Trans Image Process

    (May 2018)
  • Xin Ning, Ke Gong, Weijun Li, Liping Zhang, Xiao Bai, Shengwei Tian, Feature refinement and filter network for person...
  • Zongyu Guo, Zhibo Chen, Tao Yu, Jiale Chen, Sen Liu, Progressive image inpainting with full-resolution residual...
  • Rui Xu, Minghao Guo, Jiaqi Wang, Xiaoxiao Li, Bolei Zhou, Chen Change Loy, Texture memory-augmented deep patch-based...
  • Håkon Hukkelås, Frank Lindseth, Rudolf Mester, Image inpainting with learnable feature imputation,...
  • Chao Yang, Xin Lu, Zhe Lin, Eli Shechtman, Oliver Wang, Hao Li, High-resolution image inpainting using multi-scale...
  • D. Kim, S. Woo, J.Y. Lee, et al., Deep video inpainting, in: 2019 IEEE/CVF Conference on Computer Vision and Pattern...
  • Y. Wang, X. Tao, X. Qi, et al., Image inpainting via generative multi-column convolutional neural networks, in:...
  • Ya-Liang Chang, Zhe Liu, Kuan-Ying Lee, Winston Hsu, Learnable gated temporal shift module for deep video inpainting,...
  • U.S.M. Nadim et al.

    Global and local attention-based free-form image inpainting

    Sensors

    (2020)
  • Avisek Lahiri, Arnav Jain, Prabir Biswas, Pabitra Mitra, Improving consistency and correctness of sequence inpainting...
  • Avisek Lahiri, Sourav Bairagya, Sutanu Bera, Siddhant Haldar, Prabir Biswas, Lightweight modules for efficient deep...
  • Z.P. Qiang et al.

    Survey on deep learning image inpainting methods

    J. Image Graph.

    (2019)
  • Deepak Pathak, Philipp Krahenbuhl, Jeff Donahue, Trevor Darrell, Alexei A. Efros, Context encoders: Feature learning by...
  • Satoshi Iizuka, Edgar Simo-Serra, Hiroshi Ishikawa. Globally and locally consistent image completion, ACM TOG 36(4)...
  • Hongyu Liu, Bin Jiang, Yibing Song, Wei Huang, Chao Yang, Rethinking image inpainting via a mutual encoder-decoder with...
  • Y. Song et al.

    Contextual-based image inpainting: Infer, match, and translate

    ECCV

    (2018)
  • X. Ning et al.

    Real-time 3D face alignment using an encoder-decoder network with an efficient deconvolution layer

    IEEE Signal Process Lett.

    (2020)
  • X. Ning et al.

    JWSAA: Joint weak saliency and attention aware for person re-identification

    Neurocomputing

    (2020)
  • Guilin Liu, Fitsum A. Reda, Kevin J. Shih, Ting-Chun Wang, Andrew Tao, Bryan Catanzaro, Image inpainting for irregular...
  • Chaohao Xie, Shaohui Liu, Chao Li, Ming-Ming Cheng, Wangmeng Zuo, Xiao Liu, Shilei Wen, Errui Ding, Image inpainting...
  • Y. Zeng, J. Fu, H. Chao, B. Guo, Learning pyramid-context encoder network for high-quality image inpainting, in: CVPR,...
  • Jingyuan Li, Wang Ning, Lefei Zhang, Bo Du, Dacheng Tao, Recurrent feature reasoning for image inpainting, 2020,...
  • Jiahui Yu, Zhe Lin, Jimei Yang, Xiaohui Shen, Xin Lu, Thomas S. Huang, Generative image inpainting with contextual...
  • Jiahui Yu, Zhe Lin, Jimei Yang, Xiaohui Shen, Xin Lu, Thomas S. Huang, Free-form image inpainting with gated...
  • Hongyu Liu, Bin Jiang, Yi Xiao, Chao Yang, Coherent semantic attention for image inpainting, in: Proc. ICCV, 2019. 3,...
  • Min-cheol Sagong, Yong-goo Shin, Seung-wook Kim, Seung Park, Sung-jea Ko, Pepsi: Fast image inpainting with parallel...
  • Yu Zeng, Zhe Lin, Jimei Yang, Jianming Zhang, Eli Shechtman, Huchuan Lu, High-resolution image inpainting with...
  • Zili Yi, Qiang Tang, Shekoofeh Azizi, Daesik Jang, Zhan Xu, Contextual residual aggregation for ultra high-resolution...
  • Wenchao Du, Hu Chen, Hongyu Yang, Learning invariant representation for unsupervised image restoration, 2020,...
  • Ziyu Wan, Bo Zhang, Dongdong Chen, Pan Zhang, Dong Chen, Jing Liao, Fang Wen, Bringing old photos back to life, 2020,...
  • D. Ulyanov et al.

    Deep Image Prior

    Int. J. Comput. Vision

    (2020)
  • C. Ballester et al.

    Filling-in by joint interpolation of vector fields and gray levels

    IEEE Trans. Image Process.

    (2001)
  • Anat Levin, Assaf Zomet, Yair Weiss, Learning how to inpaint from global image statistics, in: null, 305. IEEE,...
  • S. Esedoglu, J. Shen, Digital inpainting based on the mumford–shah–euler image model, Eur. J. Appl. Math. 13(4) (2002)...
  • M. Bertalmio et al.

    Simultaneous structure and texture image inpainting

    IEEE Trans. Image Process.

    (2003)
  • Antonio Criminisi, Patrick Perez, Kentaro Toyama, Object removal by exemplar-based inpainting, in: Computer Vision and...
  • A. Criminisi et al.

    ́ Region filling and object removal by exemplar-based image inpainting

    IEEE Trans. Image Process.

    (2004)
  • C. Barnes et al.

    Patchmatch: A randomized correspondence algorithm for structural image editing

    ACM Trans. Graphics (ToG)

    (2009)
  • J.-B. Huang, S.B. Kang, N. Ahuja, J. Kopf, Image completion using planar structure guidance, ACM Trans. Graphics (TOG)...
  • S. Darabi, E. Shechtman, C. Barnes, D.B. Goldman, P. Sen, Image melding: Combining inconsistent images using...
  • Cited by (97)

    • THE FRACTIONAL LAPLACIAN BASED IMAGE INPAINTING

      2024, Inverse Problems and Imaging
    View all citing articles on Scopus

    This paper was recommended for publication by Prof G Guangtao Zhai.

    View full text