期刊文献+

自然环境视频中基于显著鲁棒轨迹的行为识别 被引量:3

Action recognition from unconstrained videos via salient and robust trajectory
原文传递
导出
摘要 目的人类行为识别是计算机视觉领域的一个重要研究课题。由于背景复杂、摄像机抖动等原因,在自然环境视频中识别人类行为存在困难。针对上述问题,提出一种基于显著鲁棒轨迹的人类行为识别算法。方法该算法使用稠密光流技术在多尺度空间中跟踪显著特征点,并使用梯度直方图(HOG)、光流直方图(HOF)和运动边界直方图(MBH)特征描述显著轨迹。为了有效消除摄像机运动带来的影响,使用基于自适应背景分割的摄像机运动估计技术增强显著轨迹的鲁棒性。然后,对于每一类特征分别使用Fisher Vector模型将一个视频表示为一个Fisher向量,并使用线性支持向量机对视频进行分类。结果在4个公开数据集上,显著轨迹算法比Dense轨迹算法的实验结果平均高1%。增加摄像机运动消除技术后,显著鲁棒轨迹算法比显著轨迹算法的实验结果平均高2%。在4个数据集(即Hollywood2、You Tube、Olympic Sports和UCF50)上,显著鲁棒轨迹算法的实验结果分别是65.8%、91.6%、93.6%和92.1%,比目前最好的实验结果分别高1.5%、2.6%、2.5%和0.9%。结论实验结果表明,该算法能够有效地识别自然环境视频中的人类行为,并且具有较低的时间复杂度。 Background In the past few years, we have witnessed a great success of social networks and multimedia technolo- gies, leading to the generation of vast amount of Interact videos. To organize these videos and to provide value-added services to users, human activities from videos should be automatically recognized. A number of research studies have focused on this challenging topic. Objective Human action recognition is a significant research topic in computer vision. The recognition of human actions from unconstrained videos is difficult because of complex background and camera motion. A robust and salient trajectory-based approach is proposed to address such problem. Method Dense optical flow is utilized to track the scale invar- iant feature transform keypoints at multiple spatial scales. The histogram of oriented gradient, histogram of optical flow, and motion boundary histogram are employed to depict the trajectory efficiently. To eliminate the influence of camera motions, a camera motion estimation approach based on adaptive background segmentation is utilized to improve the robustness of trajec- tory. The Fisher vector model is utilized to compute one Fisher vector over a complete video for each descriptor separately, and the linear support vector machine is enaployed for classification. Result The average improvement of salient trajectory al- gorithm over dense trajectory algorithm is 1% on four challenging datasets. After utilizing the camera motion elimination ap- proach, the average experimental result over salient trajectory is improved by 2%. The state-of-the-art results on four datasets (i. e., Hollywood2, ~ouTube, Olympic Sports and UCF50), the proposed algorithm obtains 65.8%, 91.6% , 93.6%, 92. 1% , and the state-of-the-art results have been improved by 1.5% , 2. 6% , 2. 5% , 0. 9% respectively. Conclusion Experi- mental results on four challenging datasets demonstrate that the proposed algorithm can effectively recognize human actions from unconstrained videos in a more computationally efficient manner compared with a number of state-of-the-art approaches.
作者 易云 王瀚漓
出处 《中国图象图形学报》 CSCD 北大核心 2015年第2期245-253,共9页 Journal of Image and Graphics
基金 国家自然科学基金项目(61102059,61472281) 中央高校基本科研业务费专项资金(0800219158,0800219270) 上海市曙光计划(12SG23)
关键词 行为识别 显著轨迹 摄像机运动消除 action recognition salient trajectory camera motion elimination Fisher vector
  • 相关文献

参考文献23

  • 1Marszalek M, Laptev I, Schmid C. Actions in context[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Miami, USA: IEEE, 2009:2929-2936. [DOI:10.1109/CVPR.2009.5206557].
  • 2Ballas N, Delezoide B, Preteux F. Trajectory signature for action recognition in video[C]//Proceedings of the ACM International Conference on Multimedia. Nara, Japan: ACM, 2012: 1429-1432. [DOI: 10.1145/2393347.2396511].
  • 3Jain M, Jégou H, Bouthemy P. Better exploiting motion for better action recognition[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Portland, USA: IEEE, 2013:2555-2562. [DOI: 10.1109/CVPR.2013.330].
  • 4Liang X, Lin L, Cao L. Learning latent spatio-temporal compositional model for human action recognition[C]//Proceedings of the ACM International Conference on Multimedia. Barcelona, Spain: ACM, 2013:263-272. [DOI: 10.1145/2502081.2502089].
  • 5Oneata D, Verbeek J, Schmid C. Action and event recognition with Fisher vectors on a compact feature set[C]//Proceedings of the IEEE International Conference on Computer Vision. Sydney, Australia: IEEE, 2013:1817-1824. [DOI: 10.1109/ICCV.2013.228].
  • 6Wang H, Kl?ser A, Schmid C, et al. Dense trajectories and motion boundary descriptors for action recognition[J]. International Journal of Computer Vision, 2013, 103(1): 60-79. [DOI: 10.1007/s11263-012-0594-8].
  • 7Farneb?ck G. Two-frame motion estimation based on polynomial expansion[C]//Proceedings of the Image Analysis. Halmstad, Sweden: Springer, 2003: 363-370. [DOI: 10.1007/3-540-45103-X_50].
  • 8Dalal N, Triggs B. Histograms of oriented gradients for human detection[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. San Diego, USA: IEEE, 2005, 1: 886-893. [DOI: 10.1109/CVPR.2005.177].
  • 9Laptev I, Marszalek M, Schmid C, et al. Learning realistic human actions from movies[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Anchorage, USA: IEEE, 2008: 1-8. [DOI: 10.1109/CVPR.2008.4587756].
  • 10Dalal N, Triggs B, Schmid C. Human detection using oriented histograms of flow and appearance[C]//Proceedings of the European Conference on Computer Vision. Graz, Austria: Springer, 2006: 428-441. [DOI: 10.1007/11744047_33].

同被引文献34

  • 1Feng Zhiquan,Yang Bo,Li Yi,et al.Real-time Oriented Behavior-driven 3D Freehand Tracking for Direct Interaction[J].Pattern Recognition,2013,46(2):590-608.
  • 2Tran D D,Le T L.Abnormal Event Detection Using Multimedia Information for Monitoring System[C]//Proceedings of IEEE Conference on Communications and Electronics.Washington D.C.,USA:IEEE Press,2014:490-495.
  • 3Li Qingwu,Cheng Haisu,Zhou Yan,et al.Road Vehicle Monitoring System Based on Intelligent Visual Internet of Things[J].Journal of Sensors,2015,151:1042-1052.
  • 4Laptev I,Lindeberg T.On Space-time Interest Points[J].International Journal of Computer Vision,2005,64(2/3):107-123.
  • 5Marszalek M,Laptev I,Schmid C.Actions in Context[C]//Proceedings of IEEE International Conference on Computer Vision and Pattern Recognition.Washington D.C.,USA:IEEE Press,2009:2929-2936.
  • 6Liu Jingen,Luo Jiebo,Shah M.Recognizing Realistic Actions from Videos“in the Wild”[C]//Proceedings of IEEE International Conference on Computer Vision and Pattern Recognition.Washington D.C.,USA:IEEE Press,2009:1996-2003.
  • 7Chakraborty B,Holte M B,Moeslund T B,et al.A Selective Spatio-temporal Interest Point Detector for Human Action Recognition in Complex Scenes[C]//Proceedings of IEEE International Conference on Computer Vision.Washington D.C.,USA:IEEE Press,2011:1776-1783.
  • 8Chen Yuanbo,Li Zhixuan,Guo Xin,et al.A Spatio-temporal Interest Point Detector Based on Vorticity for Action Recognition[C]//Proceedings of IEEE Conference on Multimedia and Expo Workshops.Washington D.C.,USA:IEEE Press,2013:365-374.
  • 9Wang Heng,Klser A,Schmid C,et al.Dense Trajectories and Motion Boundary Descriptors for Action Recognition[J].International Journal of Computer Vision,2013,103(1):60-79.
  • 10O’Hara S,Draper B.Scalable Action Recognition with a Subspace Forest[C]//Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.Washington D.C.,USA:IEEE Press,2012:1210-1217.

引证文献3

二级引证文献15

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部