期刊文献+

基于通道域注意力机制的特征融合方式 被引量:2

Feature fusion methods based on channel domain attention mechanism
下载PDF
导出
摘要 针对在基于卷积神经网络的图像处理领域内,大部分特征融合只是通过Add或者Concat操作进行特征叠加或特征拼接而不能很好地将有效特征进行融合的问题,对Add和Concat特征融合引入通道域的注意力机制,设计了4种可学习的特征融合方式:A-Cat、B-Cat、A-Add和B-Add.为了验证方法的有效性,选择YOLOv3-Tiny作为baseline,在Pascal VOC2007数据集上进行测试.结果表明:A-Cat比原Concat的mAP提高了0.76%,比B-Cat提高了1.49%;A-Add比原Add的mAP提高了0.34%,比B-Add提高了1.41%.基于注意力机制的特征融合方式可以通过学习不同特征的重要程度并据此进行特征融合,有效地提升网络的性能. Feature fusion is widely used during image processing based on convolution neural network.Most methods only use add or Concat to fuse features,and they can not be fused effectively.Therefore,the channel domain attention mechanism is introduced into the feature fusion of Add and Concat,and four learning feature fusion methods are designed:A-Cat,B-Cat,A-Add and B-Add.In order to verify the validity of the method,YOLOv3-Tiny was selected to be baseline,and it was tested on Pascal VOC2007 data set.The test results show that the mAP of A-Cat is improved 0.76%,B-Cat is improved 1.49%,A-Add is improved 0.34%and B-Add is improved 1.41%.Experiments show that feature fusion based on attention mechanism can effectively improve network performance by learning the importance of different features and fusing them accordingly.
作者 罗大为 方建军 刘艳霞 LUO Da-wei;FANG Jian-jun;LIU Yan-xia(Beijing Key Laboratory of Information Service Engineering,Beijing Union University,Beijing 100101,China;College of Urban Rail Transit and Logistics,Beijing Union University,Beijing 100101,China)
出处 《东北师大学报(自然科学版)》 北大核心 2021年第3期44-48,共5页 Journal of Northeast Normal University(Natural Science Edition)
基金 国家自然科学基金资助项目(61602041) 北京联合大学人才强校优选计划项目(BPHR2017CZ07).
关键词 Add操作 Concat操作 特征图 通道域 注意力机制 Add operation Concat operation feature maps channel domain attention mechanism
  • 相关文献

参考文献4

二级参考文献14

  • 1Stauffer C and Grimson W E L.Adaptive background mixture models for real-time tracking[C].Proceedings IEEE Computer Society Conference on Computer Vision and Pattern Recognition,Fort Collins,America 1999,2:246-252.
  • 2Qi Bin,Ghazal Mohammed,and Amer Aishy.Robust global motion estimation oriented to video object segmentation[J].IEEE Transactions on Image Processing,2008,17(6):958-967.
  • 3Chert Yue-meng.A joint approach to global motion estimation and motion segmentation from a coarsely sampled motion vector field[J].IEEE Transactions on Circuits and Systems for Video Technology,2011,21(9):1316-1328.
  • 4Itti L and Koch C.Computational modeling of visual attention[J].Nature Reviews Neuroscience,2001,2(3):193-203.
  • 5Fang Yu-ming,Lin Wei-si,Lau Chiew Tong,et al.A visual attention model combining top-down and bottom-up mechanisms for salient object detection[C].Proceedings IEEE International Conference on Acoustics,Speech and Signal Processing,Prague,Czech Republic,2011:1293-1296.
  • 6Ozkei Motoyuki,Kashiwagi Yasuhiro,Inot Top-down visual attention control based on a paticle flit,for human-interactive robots[C].Proceedings Internatiomas Conference on Human System Interactions,Yokohama,Japse,2011:188-194.
  • 7Ma Yu-Fei,Hua Xian-Sheng,and Lu Lie.A generic franmwork of user attention model and its application in video summarization[J].IEEE Transactions on Multimedia,2005,7(5):907-919.
  • 8Han Jun-wei.Object segmentation from consumer video:a unified framework based on visual attention[J].IEEE Transactions on Consumer Electronics,2009,55(3):1597-1605.
  • 9Verri A and Pggio T.Motion field and optical flow:qualitative Properties[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,1989,11(5):490-498.
  • 10曹诗雨,刘跃虎,李辛昭.基于Fast R-CNN的车辆目标检测[J].中国图象图形学报,2017,22(5):671-677. 被引量:63

共引文献107

同被引文献11

引证文献2

二级引证文献4

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部