摘要
提出一种融合时间及目标之间空间语义信息的视频运动目标交互行为识别方法,即基于目标之间空间语义的变化规律识别其交互行为类别。不同于传统的语义事件建模方法,首先根据运动目标跟踪结果,基于其运动方向以及建立目标之间的空间关系(拓扑关系和方向关系)模型,提出一种提取人目标之间空间语义(前面、后面、背对、面对以及左右)的方法;然后基于空间语义的变化规律建立随机文法规则;最后采用随机文法器识别九种常见的两人交互行为。该方法无需训练样本,实验结果验证了方法的有效性及优越性。
A method for recognizing human-human interaction is proposed based on the spatial-temporal semantic information.Different from traditional methods to model the interactions,this framework achieves recognizing activities based on the transformation of the spatial semantic meaning.First,with detection and tracking results,the spatial semantic meaning(front,back,face to face,back to back,and left or right) between the persons are extracted based on motion directions and spatial relationships(including topological and directional relations).Then,stochastic context-free grammar is used to recognize interactions that the rules are learned based on the transformation of spatial semantics.Extensive experiments have been executed to validate the effectiveness of the proposed approach,and the method can recognize the interactions without additional training.
出处
《光学学报》
EI
CAS
CSCD
北大核心
2012年第5期145-151,共7页
Acta Optica Sinica
基金
国家973计划(2010CB327900)
国家自然科学基金(61001176)资助课题
关键词
机器视觉
交互行为识别
空间关系
空间语义
随机文法
machine vision
interaction recognition
spatial relationship
spatial semantic meaning
stochastic context-free grammar