arXiv Analytics

Sign in

arXiv:2411.05705 [cs.CV]AbstractReferencesReviewsResources

Image inpainting enhancement by replacing the original mask with a self-attended region from the input image

Kourosh Kiani, Razieh Rastgoo, Alireza Chaji, Sergio Escalera

Published 2024-11-08Version 1

Image inpainting, the process of restoring missing or corrupted regions of an image by reconstructing pixel information, has recently seen considerable advancements through deep learning-based approaches. In this paper, we introduce a novel deep learning-based pre-processing methodology for image inpainting utilizing the Vision Transformer (ViT). Our approach involves replacing masked pixel values with those generated by the ViT, leveraging diverse visual patches within the attention matrix to capture discriminative spatial features. To the best of our knowledge, this is the first instance of such a pre-processing model being proposed for image inpainting tasks. Furthermore, we show that our methodology can be effectively applied using the pre-trained ViT model with pre-defined patch size. To evaluate the generalization capability of the proposed methodology, we provide experimental results comparing our approach with four standard models across four public datasets, demonstrating the efficacy of our pre-processing technique in enhancing inpainting performance.

Related articles: Most relevant | Search more
arXiv:1905.03556 [cs.CV] (Published 2019-05-09)
Cycle-IR: Deep Cyclic Image Retargeting
arXiv:1211.2881 [cs.CV] (Published 2012-11-13, updated 2012-11-28)
Deep Attribute Networks
arXiv:2106.01863 [cs.CV] (Published 2021-06-03)
Robust Reference-based Super-Resolution via C2-Matching