期刊文献+

基于密集连接块U-Net的语义人脸图像修复 被引量:9

Semantic face image inpainting based on U-Net with dense blocks
下载PDF
导出
摘要 针对人脸图像在待修复缺损面积较大时,现有方法的修复存在图像语义理解不合理、边界不连贯等视觉瑕疵的问题,提出基于密集连接块的U-Net结构的端到端图像修复模型,以实现对任意模板的语义人脸图像的修复。首先,采用生成对抗网络思想,生成器采用密集连接块代替U-Net中的普通卷积模块,以捕捉图像中缺损部分的语义信息并确保前面层的特征被再利用;然后,使用跳连接以减少通过下采样而造成的信息损失,从而提取图像缺损区域的语义;最后,通过引入对抗损失、内容损失和局部总变分(TV)损失这三者的联合损失函数来训练生成器,确保了修复边界和周围真实图像的视觉一致,并通过Hinge损失来训练判别器。所提模型和GLC、DF、门控卷积(GC)在人脸数据集CelebA-HQ上进行了对比。实验结果表明,所提模型能有效提取人脸图像语义信息,修复结果具有自然过渡的边界和清晰的局部细节。相较性能第二的GC,所提模型对中心模板修复的结构相似性(SSIM)和峰值信噪比(PSNR)分别提高了5.68%和7.87%,Frechet Inception距离(FID)降低了7.86%;对随机模板修复的SSIM和PSNR分别提高了7.06%和4.80%,FID降低了6.85%。 When the areas to be inpainted in the face image are large,there are some visual defects caused by the inpainting of the existing methods,such as unreasonable image semantic understanding and incoherent boundary.To solve this problem,an end-to-end image inpainting model of U-Net structure based on dense blocks was proposed to achieve the inpainting of semantic face of any mask.Firstly,the idea of generative adversarial network was adopted.In the generator,the convolutional layers in U-Net were replaced with dense blocks to capture the semantic information of the missing regions of the image and to make sure the features of the previous layers were reused.Then,the skip connections were adopted to reduce the information loss caused by the down-sampling,so as to extract the semantics of the missing regions.Finally,by introducing the joint loss function combining adversarial loss,content loss and local Total Variation(TV)loss to train the generator,the visual consistency between the inpainted boundary and the surrounding real image was ensured,and Hinge loss was used to train the discriminator.The proposed model was compared with Globally and Locally Consistent image completion(GLC),Deep Fusion(DF)and Gated Convolution(GC)on CelebA-HQ face dataset.Experimental results show that the proposed model can effectively extract the semantic information of face images,and its inpainting results have the boundaries with natural transition and clear local details.Compared with the second-best GC,the proposed model has the Structure SIMilarity index(SSIM)and Peak Signal-to-Noise Ratio(PSNR)increased by 5.68%and 7.87%respectively,while the Frechet Inception Distance(FID)decreased by 7.86%for the central masks;and has the SSIM and PSNR increased by 7.06%and 4.80%respectively while the FID decreased by 6.85%for the random masks.
作者 杨文霞 王萌 张亮 YANG Wenxia;WANG Meng;ZHANG Liang(School of Science,Wuhan University of Technology,Wuhan Hubei 430070,China;Tus College of Digit,Guangxi University of Science and Technology,Liuzhou Guangxi 545006,China)
出处 《计算机应用》 CSCD 北大核心 2020年第12期3651-3657,共7页 journal of Computer Applications
基金 国家自然科学基金资助项目(61573012) 国家留学基金资助项目(201906955038) 柳州科技计划项目(2018DH10503)。
关键词 语义图像修复 生成对抗网络 密集连接块 损失函数 局部总变分 编码器-解码器 semantic image inpainting generative adversarial network dense block loss function local Total Variation(TV) encoder-decoder
  • 相关文献

参考文献4

二级参考文献17

共引文献18

同被引文献70

引证文献9

二级引证文献31

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部