期刊文献+

视觉注意机制下结合语义特征的行人检测 被引量:2

Semantic feature-based visual attention model for pedestrian detection
原文传递
导出
摘要 目的为研究多场景下的行人检测,提出一种视觉注意机制下基于语义特征的行人检测方法。方法首先,在初级视觉特征基础上,结合行人肤色的语义特征,通过将自下而上的数据驱动型视觉注意与自上而下的任务驱动型视觉注意有机结合,建立空域静态视觉注意模型;然后,结合运动信息的语义特征,采用运动矢量熵值计算运动显著性,建立时域动态视觉注意模型;在此基础上,以特征权重融合的方式,构建时空域融合的视觉注意模型,由此得到视觉显著图,并通过视觉注意焦点的选择完成行人检测。结果选用标准库和实拍视频,在Matlab R2012a平台上,进行实验验证。与其他视觉注意模型进行对比仿真,本文方法具有良好的行人检测效果,在实验视频上的行人检测正确率达93%。结论本文方法在不同的场景下具有良好的鲁棒性能,能够用于提高现有视频监控系统的智能化性能。 Objective Pedestrian detection under video surveiUance systems has always been a hot topic in computer vision research. These systems are widely used in train stations, airports, large commercial plazas, and other public places. How- ever, pedestrian detection remains difficult because of complex backgrounds. Given its development in recent years, the visual attention mechanism has attracted increasing attention in object detection and tracking research, and previous studies have achieved substantial progress and breakthroughs. We propose a novel pedestrian detection method based on the seman- tic features under the visual attention mechanism. Method The proposed semantic feature-based visual attention model is a spatial-temporal model that consists of two parts: the static visual attention model and the motion visual attention model. The static visual attention model in the spatial domain is constructed by combining bottom-up with top-down attention guid- ance. Based on the characteristics of pedestrians, the bottom-up visual attention model of Itti is improved by intensifying the orientation vectors of elementary visual features to make the visual saliency map suitable for pedestrian detection. In terms of pedestrian attributes, skin color is selected as a semantic feature for pedestrian detection. The regional and Ganssian models are adopted to construct the skin color model. Skin feature-based visual attention guidance is then proposed to com- plete the top-down process. The bottom-up and top-down visual attentions are linearly combined using the proper weights obtained from experiments to construct the static visual attention model in the spatial domain. The spatial-temporal visual attention model is then constructed via the motion features in the temporal domain. Based on the static visual attention mod- el in the spatial domain, the frame difference method is combined with optical flowing to detect motion vectors. Filtering is applied to process the field of motion vectors. The saliency of motion vectors can be evaluated via motion entropy to make the selected motion feature more suitable for the spatial-temporal visual attention model. Result Standard datasets and prac- tical videos are selected for the experiments. The experiments are performed on a MATLAB R2012a platform. The experi- mental results show that our spatial-temporal visual attention model demonstrates favorable robustness under various scenes, including indoor train station surveillance videos and outdoor scenes with swaying leaves. Our proposed model outperforms the visual attention model of hti, the graph-based visual saliency model, the phase spectrum of quaternion Fourier transform model, and the motion channel model of Liu in terms of pedestrian detection. The proposed model achieves a 93% accura- cy rate on the test video. Conclusion This paper proposes a novel pedestrian method based on the visual attention mecha- nism. A spatial-temporal visual attention model that uses low-level and semantic features is proposed to calculate the sali- ency map. Based on this model, the pedestrian targets can be detected through focus of attention shifts. The experimental results verify the effectiveness of the proposed attention model for detecting pedestrians.
出处 《中国图象图形学报》 CSCD 北大核心 2016年第6期723-733,共11页 Journal of Image and Graphics
基金 国家自然科学基金项目(1008-GAA14033) 中国民航总局科技基金项目(1004-14000202) 中国民航信息技术科研基地开放基金项目(1004-ZBA12016) 南京航空航天大学理工融合项目(1008-56XZA15009)~~
关键词 行人检测 视觉注意模型 语义特征 显著图 肤色 运动矢量熵值 people detection visual attention model semantic features saliency map skin color motion entropy
  • 相关文献

参考文献21

  • 1Ren Z X, Gao S H, Chia L T, et al. Region-based saliency de- tection and its application in object recognition[ J]. IEEE Trans- actions on Circuits and Systems for Video Technology, 2014, 24(5) : 769-779. [ DOI: 10. ll09/TCSVT. 2013. 2280096].
  • 2Mahadevan V, Vasconcelos N. Biologically inspired object track- ing using center-surround saliency mechanisms [ J ]. IEEE Trans- actions on Pattern Analysis and Machine Intelligence, 2013, 35(3) : 541-554. [DOI: 10. 1109/TPAMI. 2012.98].
  • 3Chang K Y, Liu T L, Chert H T, et al. Fusing generic object- ness and visual saliency for salient object detection [ C ]//Pro- ceedings of IEEE International Conference on Computer Vision. Barcelona: IEEE, 2011 : 914-921. [DOI : 10. 1109/ICCV. 2011. 6126333 ].
  • 4Itti L, Koch C. Computational modelling of visual attention[J]. Nature Reviews Neuroscience, 2001, 2 (3): 194-203. [ DOI: 10. 1038/35058500].
  • 5Itti L. Models of bottom-up and top-down visual attention [ D ]. California: California Institute of Technology, 2000.
  • 6Itti L, Koch C, Niebur E. A model of saliency-based visual at- tention for rapid scene analysis [J].IEEE Transactions on Pat- tern Analysis and Machine Intelligence, 1998, 20 ( 11 ) : 1254- 1259. [ DOI : 10.1109/34. 730558 ].
  • 7Seholkopf B, Platt J, Hofmann T. Graph-based visual saliency [ C ]//Proceedings of the 2006 Conference on Advanees in Neu- ral Information Processing Systems 19. London: MIT Press, 2007 : 545-552.
  • 8Hou X D, Zhang L Q. Saliency detection : a spectral residual ap- proach [ C ]//Proceedings of IEEE Conference on Computer Vi- sion and Pattern Recognition. Minneapolis, MN : IEEE, 2007 : 1-8. [DOI: 10. l109/CVPR. 2007. 383267 ].
  • 9Guo C L, Ma Q, Zhang L M. Spatio-temporal saliency detection using phase spectrum of quatemion fourier transform [ C ]//Pro- ceedings of IEEE Conference on Computer Vision and Pattern Recognition. Anchorage, AK: IEEE, 2008: 1-8. [DOI: 10. 1109/CVPR. 2008. 4587715 ].
  • 10Einhauser W, Spain M, Perona P. Objects predict fixations bet- ter than early saliency [ J ]. Journal of Vision, 2008, 8 ( 14 ) : 1-26. [DOI: 10. 1167/8.14.18].

同被引文献19

引证文献2

二级引证文献13

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部