期刊文献+

基于残差卷积注意力网络的视频修复 被引量:4

Video inpainting based on residual convolution attention network
下载PDF
导出
摘要 视频修复旨在填补视频中的缺失区域,由于很难精确保持修复内容的时空一致性,故视频修复仍具有挑战性。针对现有视频修复中存在的修复结果语义信息不连续,出现视频模糊和时间伪影,以及网络设计越来越复杂,网络整体速度变慢的问题,本文提出了一种基于残差网络的卷积注意力网络(RCAN)用以视频修复。通过将自注意力机制和全局注意力机制引入进残差网络,增强网络对所有输入帧的时空特征的学习能力,并采用时空对抗损失函数进行优化,提高视频修复的质量。同时网络还能够高度自由地定义层数和参数量,提高网络的实际应用能力。实验结果表明,该网络在DAVIS和YouTube-VOS数据集上取得了PSNR为30.68 dB,SSIM为0.961,FID为0.113的平均修复结果,基本符合实际场景对模型的修复质量要求,为视频修复提供了一种新思路。 Video inpainting,which aims at filling in missing regions of a video,remains challenging due to the difficulty of preserving the precise spatial and temporal coherence of video contents.In order to solve the problems of discontinuous semantic information,video blurriness and temporal artifact,and more and more complex network design,the overall speed of the network becoming slow,this paper proposes a residual convolution attention network(RCAN)for video inpainting.By introducing the self-attention mechanism and the global attention mechanism into the residual network,the ability of the network to learn the spatio-temporal features of all input frames is enhanced.This method proposes a spatial-temporal adversarial loss function to optimize RCAN,which improves the quality of video inpainting.At the same time,the network can define the number of layers and parameters with a high degree of freedom to improve the practical application ability of the network.Experimental results show that the network can achieve an average inpainting result in that the PSNR is 30.68 dB,the SSIM is 0.961,and the FID is 0.113 on DAVIS and YouTube-VOS data sets.This method meets the inpainting quality requirements of the actual scene on the model and provides a new idea for video inpainting.
作者 李德财 严群 姚剑敏 林志贤 董泽宇 LI De-cai;YAN Qun;YAO Jian-min;LIN Zhi-xian;DONG Ze-yu(College of Physics and Information Engineering, Fuzhou University, Fuzhou 350108, China;Jinjiang RichSence Electronic Technology Company Limited, Jinjiang 362200, China)
出处 《液晶与显示》 CAS CSCD 北大核心 2022年第1期86-96,共11页 Chinese Journal of Liquid Crystals and Displays
基金 国家重点研发计划(No.2016YFB0401503) 广东省科技重大专项(No.2016B090906001) 福建省科技重大专项(No.2014HZ003-1) 广东省光信息材料与技术重点实验室开放基金(No.2017B030301007)。
关键词 深度学习 视频修复 自注意力机制 残差网络 生成对抗网络 deep learning video inpainting self-attention mechanism residual networks generative adversarial networks
  • 相关文献

参考文献5

二级参考文献28

  • 1赵悟翔,王琼华,李大海,陶宇虹,王芳宁.阶梯光栅多视点自由立体显示的子像素排列[J].四川大学学报(工程科学版),2009,41(6):216-218. 被引量:8
  • 2WEXLER Y, SHECHTMAN E, IRANI M. Space-time completion of video[ J]. IEEE Transactions on Pattern Analysis and Machine In- telligence, 2007, 29(3): 463-476.
  • 3JIA Y T, HU S M, MARTIN R R. Video completion using tracking and fragment merging[ J]. The Visual Computer, 2005, 21 (8/9/ 10): 601-610.
  • 4SHIRATORI T, MATSUSHITA Y, TANG X, et al . Video completion by motion field transfer[ C]// Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE, 2006, 1:411-418.
  • 5GRANADOS M, TOMPKIN J, KIM K, et al. How not to be seen- object removal from videos of crowded scenes[ C]// Eurographics 2012: Computer Graphics Forum. Oxford: Blackwell Publishing, 2012:219-228.
  • 6PATWARDHAN K A, SAPIRO G, BERTALMIO M. Video inpainting under constrained camera motion[ J]. IEEE Transactions on Image Processing, 2007, 16(2) : 545 -553.
  • 7SHEN Y, LU F, CAO X, et al. Video completion for perspective camera under constrained motion[ C]// Proceedings of the 18th International Conference on Pattern Recognition. Piscataway: IEEE, 2006, 3:63 - 66.
  • 8GRANADOS M, KIM K I, TOMPKIN J, et al. Background inpainting for videos with dynamic objects and a free-moving camera[M]. Heidelberg: Springer, 2012:682-695.
  • 9AGARWALA A, HERTZMANN A, SALESIN D H, et al. Keyframe-based tracking for rotoscoping and animation [ J]. ACM Transactions on Graphics, 2004, 23(3): 584-591.
  • 10LIU C, FREEMAN W T, ADELSON E H, et al. Human-assisted motion annotation[ C]// Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE, 2008:1-8.

共引文献26

同被引文献35

引证文献4

二级引证文献4

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部