期刊文献+

基于注意力的双流CNN的行为识别 被引量:8

Action recognition of two-stream CNN based on attention
下载PDF
导出
摘要 针对现今网络不能充分融合视频的时空信息,提出一种基于注意力的双流CNN与DU-DLSTM的识别模型。采用Opencv提取视频帧和相应的光流特征图,空间流网络解码相应的光流特征图得到空间注意力增强向量,解码图像序列得到原始图像时间维的特征向量,作为时间流网络的输入。将两个网络的输出特征加权融合后输入DU-DLSTM(单双向结构的长短时记忆网络)模块,利用Softmax最大似然函数完成行为识别任务。提出方法具有很好的鲁棒性,在KTH数据集上达到98.9%的识别精度。 To solve the problem that the space-time information of video cannot be fully integrated in today’s network,a recognition model of two-stream CNN networks based on attention and DU-DLSTM was proposed.The video frame and the correspon-ding optical flow feature map were extracted using Opencv.The spatial attention enhancement vector was obtained by decoding the corresponding optical flow feature map through the spatial flow network,and the feature vector of the original image time dimension was obtained by decoding the image sequence,which was used as the input of the time flow network.The output features of the two networks were weighted and merged into DU-DLSTM module.The maximum likelihood function of Softmax was used to complete the task of action recognition.The proposed method has good robustness and achieves a high recognition accuracy of 98.9%on KTH datasets.
作者 马翠红 王毅 毛志强 MA Cui-hong;WANG Yi;MAO Zhi-qiang(College of Electrical Engineering,North China University of Science and Technology,Tangshan 063210,China)
出处 《计算机工程与设计》 北大核心 2020年第10期2903-2906,共4页 Computer Engineering and Design
基金 国家自然科学基金项目(61171058)。
关键词 光流特征 注意力机制 双流网络 单双向结构的长短时记忆网络 行为识别 optical flow feature attention mechanism two-stream networks DU-DLSTM action recognition
  • 相关文献

参考文献1

二级参考文献29

  • 1Text Retrieval Conference (TREC) [Online], available: http://trec.nist.gov/, April 5, 2016.
  • 2National Institute of Standards and Technology (NIST) [Online], available: http://www.nist.gov/index.html, April 5, 2016.
  • 3TREC Video Retrieval Evaluation (TRECVID) [Online], available: http://www-nlpir.nist.gov/projects/trecvid/, Ap- ril 5, 2016.
  • 4Dollar P, Wojek C, Schiele B, Perona P. Pedestrian detec- tion: an evaluation of the state of the art. IEEE Transac- tions on Pattern Analysis and Machine Intelligence, 2012, 34(4): 743--761.
  • 5Benenson R, Omran M, Hosang J, Schiele B. Ten years of pedestrian detection, what have we learned? In: Proceed- ings of the 12th European Conference on Computer Vision. Zurich, Switzerland: Springer, 2014. 613-627.
  • 6Dalal N, Triggs B. Histograms of oriented gradients for hu- man detection. In: Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recog- nition. San Diego, USA: IEEE, 2005. 886-893.
  • 7Felzenszwalb P, McAllester D, Ramanan D. A discrimina- tively trained, multiscale, deformable part model. In: Pro- ceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition. Anchorage, Alaska, USA: IEEE, 2008. 1-8.
  • 8Ouyang W, Wang X. Joint deep learning for pedestrian de- tection. In: Proceedings of the 2013 IEEE International Conference on Computer Vision. Sydney, Australia: IEEE, 2013. 2056-2063.
  • 9Luo P, Tian Y, Wang X, Tang X. Switchable deep network for pedestrian detection. In: Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition. Columbus, Ohio, USA: IEEE, 2014. 899-906.
  • 10Hosang J, Omran M, Benenson R, Schiele B. Taking a deeper look at pedestrians. In: Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recog- nition. Boston, USA: IEEE, 2015. 4073-4082.

共引文献15

同被引文献55

引证文献8

二级引证文献33

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部