期刊文献+

双分支特征融合网络的步态识别算法 被引量:7

Dual branch feature fusion network based gait recognition algorithm
原文传递
导出
摘要 目的在步态识别算法中,基于外观的方法准确率高且易于实施,但对外观变化敏感;基于模型的方法对外观变化更加鲁棒,但建模困难且准确率较低。为了使步态识别算法在获得高准确率的同时对外观变化具有更好的鲁棒性,提出了一种双分支网络融合外观特征和姿态特征,以结合两种方法的优点。方法双分支网络模型包含外观和姿态两条分支,外观分支采用Gait Set网络从轮廓图像中提取外观特征;姿态分支采用5层卷积网络从姿态骨架中提取姿态特征。在此基础上构建特征融合模块,融合外观特征和姿态特征,并引入通道注意力机制实现任意尺寸的特征融合,设计的模块结构使其能够在融合过程中抑制特征中的噪声。最后将融合后的步态特征应用于识别行人身份。结果实验在CASIA-B(Institute of Automation,Chinese Academy of Sciences,Gait Dataset B)数据集上通过跨视角和不同行走状态两种实验设置与目前主流的步态识别算法进行对比,并以Rank-1准确率作为评价指标。在跨视角实验设置的MT(medium-sample training)划分中,该算法在3种行走状态下的准确率分别为93.4%、84.8%和70.9%,相比性能第2的算法分别提升了1.4%、0.5%和8.4%;在不同行走状态实验设置中,该算法在两种行走状态下的准确率分别为94.9%和90.0%,获得了最佳性能。结论在能够同时获取外观数据和姿态数据的场景下,该算法能够有效地融合外观信息和姿态信息,在获得更丰富的步态特征的同时降低了外观变化对步态特征的影响,提高了步态识别的性能。 ObjectiveGait is a kind of human walking pattern,which is one of the key biometric features for person identification.As a non-contact and long-distance recognition way to capture human identity information,gait recognition has been developed in video surveillance and public security.Gait recognition algorithms can be segmented into two mainstreams like appearance-based methods and the model-based methods.The appearance-based methods extract gait from a sequence of silhouette images in common.However,the appearance-based methods are basically affected by appearance changes like non-rigid clothing deformation and background clutters.Different from the appearance-based methods,the model-based methods commonly leverage body structure or motion prior to model gait pattern and more robust to appearance variations.Actually,it is challenged to identify a universal model for gait description,and the previous pre-defined models can be constrained in certain scenarios.Recent model-based methods are focused on deep learning-based pose estimation to model key-points of human body.But the estimated pose model constrains the redundant noises in subject to pose estimators and occlusion.In summary,the appearance-based methods are based visual features description while the model-based methods tend to describe a semantic level-based motion and structure.We aim to design a novel approach for gait recognition beyond the existed two methods mentioned above and improve gait recognition ability via the added appearance features and pose features.Methodwe design a dual-branch network for gait recognition.The input data are fed into a dual-branch network to extract appearance features and pose features each.Then,the two kinds of features are merged into the final gait features in the context of feature fusion module.In detail,we adopt an optimal network Gait Set as the appearance branch to extract appearance features from silhouette images and design a two-stream convolutional neural network(CNN)to extract pose features from pose key-points based on the position information and motion information.Meanwhile,a squeeze-and-excitation feature fusion module(SEFM)is designed to merge two kinds of features via the weights of two kinds of features learning.In the squeeze step,appearance feature maps and pose feature maps are integrated via pooling,concatenation,and projection.In the excitation step,we obtain the weighted feature maps of appearance and pose via projection and Hadamard product.The two kinds of feature maps are down-sampled and concatenated into the final gait feature in accordance with adaptive weighting.To verify the appearance features and pose features,we design two variants of SEFM in related to SEFM-A and SEFM-P further.The SEFM module merges appearance features and pose features in mutual;the SEFM-A module merges pose features into appearance features and appearance features remain unchanged;the SEFM-P module merges appearance features into pose features and no pose features changed.Our algorithm is based on Pytorch and the evaluation is carried out on database CASIA(Institute of Automation,Chinese Academy of Sciences)Gait Dataset B(CASIA-B).We adopt the Alpha Pose algorithm to extract pose key-points from origin RGB videos,and use silhouette images obtained.In each iteration of the training process,we randomly select 16 subjects and select 8 random samples of each subject further.Every sample of them contains a sub-sequence of 30 frames.Consequently,each batch has 3840 image-skeleton pairs.We adopt the Adam optimizer to optimize the network for 60000 iterations.The initial learning rate is set to 0.0002 for the pose branch,and 0.0001 for the appearance branch and the SEFM,and then the learning rate is cut10 times at the 45000-th iteration.ResultWe first verify the effectiveness of the dual-branch network and feature fusion modules.Our demonstration illustrates that our dual-branch network can enhance performance and there is a clear complementary effect between appearance features and pose features.The Rank-1 accuracies of five feature fusion modules like SEFM,SEFM-A,SEFM-P,Concatenation,and multi-modal transfer module(MMTM)are 83.5%,81.9%,93.4%,92.6%and 79.5%,respectively.These results demonstrate that appearance features are more discriminative because there are noises existed in pose features.Our SEFM-P is capable to merge two features in the feature fusion procedure via noises suppression.Then,we compare our methods to advanced gait recognition methods like CNNs,event-based gait recognition(EV-Gait),Gait Set,and Pose Gait.We conduct the experiments with two protocols and evaluate the rank-1 accuracy of three walking scenarios in the context of normal walking,bag-carrying,and coat-wearing.Our method archives the best performance in all experimental protocols.Our three scenarios-based rank-1 accuracies are reached 93.4%,84.8%,and 70.9%in protocol 1.The results of protocol 2 are obtained by 95.7%,87.8%,77.0%,respectively.Comparing to the second-best method of Gait Set,the rank-1 accuracies in the context of coat-wearing walking scenario are improved by 8.4%and 6.6%.ConclusionWe harness a novel gait recognition network based on the fusions of appearance features and pose features.Our analyzed results demonstrated that our method can develop two kinds of features and the appearance variations is more robust,especially for clothing changes scenario.
作者 徐硕 郑锋 唐俊 鲍文霞 Xu Shuo;Zheng Feng;Tang Jun;Bao Wenxia(School of Electronics and Information Engineering,Anhui University,Hefei 230601,China;College of Engineering,Southern University of Science and Technology,Shenzhen 518055,China)
出处 《中国图象图形学报》 CSCD 北大核心 2022年第7期2263-2273,共11页 Journal of Image and Graphics
基金 国家自然科学基金项目(61772032) 国家重点研发计划资助(SQ2018YFC080102) 安徽省重点研发计划资助(202004a7020050)。
关键词 生物特征识别 步态识别 特征融合 双分支网络 SE模块 人体姿态估计 步态轮廓图像 biometric recognition gait recognition feature fusion two-branch network squeeze-and-excitation module human body pose estimation gait silhouette images
  • 相关文献

参考文献1

二级参考文献89

  • 1彭彰,吴晓娟,杨军.基于肢体长度参数的多视角步态识别算法[J].自动化学报,2007,33(2):210-213. 被引量:10
  • 2Little J,Boyd J E.Recognizing People by Their Gait:The Shape of Motion.Videre:Journal of Computer Vision Research,1998,1(2):1-32.
  • 3Tanawongsuwan R,Bobick A.Performance Analysis of TimeDistance Gait Parameters under Different Speeds//Proc of the4th International Conference on Audio-and Video-Based Biometric Person Authentication.Guildford,UK,2003:715-724.
  • 4Cuntoor N,Kale A,Chellappa R.Combining Multiple Evidences for Gait Recognition//Proc of the International Conference on Acoustics,Speech and Signal Processing.Hong Kong,China,2003,III:33-36.
  • 5Chalidabhongse T,Kruger V,Chellappa R.The UMD Database for Human Identification at a Distance.Technical Report.College Park,USA:University of Maryland,2001.
  • 6Gross R,Shi J.The CMU Motion of Body(MoBo)Database.Technical Report,CMU-RI-TR-01-18.Pittsburgh,USA:Carnegie Mellon University,2001.
  • 7Nixon M,Carter J,Shutler J,et al.Experimental Plan for Automatic Gait Recognition.Technical Report.Southampton,UK:University of Southampton,2001.
  • 8Sarkar S.The Human ID Gait Challenge Problem:Data Sets,Performance and Analysis.IEEE Trans on Pattern Analysis and Machine Intelligence,2005,27(2):162-177.
  • 9Wang Liang,Tan Tieniu.Silhouette Analysis-Based Gait Recognition for Human Identification.IEEE Trans on Pattern Analysis and Machine Intelligence,2003,25(12):1505-1518.
  • 10Yu Shiqi,Tan Daoliang,Tan Tieniu.A Framework for Evaluating the Effect of View Angle,Clothing and Carrying Condition on Gait Recognition//Proc of the18th International Conference on Pattern Recognition.Hong Kong,China,2006:441-444.

共引文献72

同被引文献38

引证文献7

二级引证文献5

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部