Full Text

Turn on search term navigation

© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.

Abstract

Although image inpainting based on the generated adversarial network (GAN) has made great breakthroughs in accuracy and speed in recent years, they can only process low-resolution images because of memory limitations and difficulty in training. For high-resolution images, the inpainted regions become blurred and the unpleasant boundaries become visible. Based on the current advanced image generation network, we proposed a novel high-resolution image inpainting method based on multi-scale neural network. This method is a two-stage network including content reconstruction and texture detail restoration. After holding the visually believable fuzzy texture, we further restore the finer details to produce a smoother, clearer, and more coherent inpainting result. Then we propose a special application scene of image inpainting, that is, to delete the redundant pedestrians in the image and ensure the reality of background restoration. It involves pedestrian detection, identifying redundant pedestrians and filling in them with the seemingly correct content. To improve the accuracy of image inpainting in the application scene, we proposed a new mask dataset, which collected the characters in COCO dataset as a mask. Finally, we evaluated our method on COCO and VOC dataset. the experimental results show that our method can produce clearer and more coherent inpainting results, especially for high-resolution images, and the proposed mask dataset can produce better inpainting results in the special application scene.

Details

Title
High-Resolution Image Inpainting Based on Multi-Scale Neural Network
Author
Sun, Tingzhu 1 ; Fang, Weidong 2   VIAFID ORCID Logo  ; Chen, Wei 3   VIAFID ORCID Logo  ; Yao, Yanxin 4 ; Bi, Fangming 1 ; Wu, Baolei 1 

 School of Computer Science and Technology, China University of Mining Technology, Xuzhou 221000, Jiangsu, China; [email protected] (T.S.); [email protected] (F.B.); [email protected] (B.W.); Mine Digitization Engineering Research Center of the Ministry of Education, China University of Mining and Technology, Xuzhou 221116, Jiangsu, China 
 Key Laboratory of Wireless Sensor Network & Communication, Shanghai Institute of Micro-system and Information Technology, Chinese Academy of Sciences, Shanghai 201800, China; [email protected] 
 School of Computer Science and Technology, China University of Mining Technology, Xuzhou 221000, Jiangsu, China; [email protected] (T.S.); [email protected] (F.B.); [email protected] (B.W.); Mine Digitization Engineering Research Center of the Ministry of Education, China University of Mining and Technology, Xuzhou 221116, Jiangsu, China; School of Earth and Space Sciences, Peking University, Beijing 100871, China 
 School of Communication and Information Engineering, Beijing Information Science & Technology University, Beijing 100101, China; [email protected] 
First page
1370
Publication year
2019
Publication date
2019
Publisher
MDPI AG
e-ISSN
20799292
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
2548429996
Copyright
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.