Deep Learning-Based Image and Video Inpainting: A Survey

Weize Quan, Jiaxi Chen, Yanli Liu, Dong Ming Yan*, Peter Wonka

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

22 Scopus citations

Abstract

Image and video inpainting is a classic problem in computer vision and computer graphics, aiming to fill in the plausible and realistic content in the missing areas of images and videos. With the advance of deep learning, this problem has achieved significant progress recently. The goal of this paper is to comprehensively review the deep learning-based methods for image and video inpainting. Specifically, we sort existing methods into different categories from the perspective of their high-level inpainting pipeline, present different deep learning architectures, including CNN, VAE, GAN, diffusion models, etc., and summarize techniques for module design. We review the training objectives and the common benchmark datasets. We present evaluation metrics for low-level pixel and high-level perceptional similarity, conduct a performance evaluation, and discuss the strengths and weaknesses of representative inpainting methods. We also discuss related real-world applications. Finally, we discuss open challenges and suggest potential future research directions.

Original languageEnglish (US)
Pages (from-to)2367-2400
Number of pages34
JournalInternational Journal of Computer Vision
Volume132
Issue number7
DOIs
StatePublished - Jul 2024

Keywords

  • Content generation
  • Deep learning
  • Image inpainting
  • Video inpainting

ASJC Scopus subject areas

  • Software
  • Computer Vision and Pattern Recognition
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Deep Learning-Based Image and Video Inpainting: A Survey'. Together they form a unique fingerprint.

Cite this