期刊文献+

小数据样本深度迁移网络自发表情分类 被引量:9

Classification of small spontaneous expression database based on deep transfer learning network
原文传递
导出
摘要 目的相较于传统表情,自发表情更能揭示一个人的真实情感,在国家安防、医疗等领域有巨大的应用潜力。由于自发表情具有诱导困难、样本难以采集等特殊性,因此数据样本较少。为判别自发表情的种类,结合在越来越多的场景得到广泛应用的神经网络学习方法,提出基于深度迁移网络的表情种类判别方法。方法为保留原始自发表情图片的特征,即使在小数据样本上也不使用数据增强技术,并将光流特征3维图像作为对比样本。将样本置入不同的迁移网络模型中进行训练,然后将经过训练的同结构的网络组合成同构网络并输出结果,从而实现自发表情种类的判别。结果实验结果表明本文方法在不同数据库上均表现出优异的自发表情分类判别特性。在开放的自发表情数据库CASME、CASMEⅡ和CAS(ME)~2上的测试平均准确率分别达到了94. 3%、97. 3%和97. 2%,比目前最好测试结果高7%。结论本文将迁移学习方法应用于自发表情种类的判别,并对不同网络模型以及不同种类的样本进行比较,取得了目前最优的自发表情种类判别的平均准确率。 Objective Expression is important in human-computer interaction.As a special expression,spontaneous expression features shorter duration and weaker intensity in comparison with traditional expressions.Spontaneous expressions can reveal a person’s true emotions and present immense potential in detection,anti-detection,and medical diagnosis.Therefore,identifying the categories of spontaneous expression can make human-computer interaction smooth and fundamentally change the relationship between people and computers.Given that spontaneous expressions are difficult to be induced and collected,the scale of a spontaneous expression dataset is relatively small for training a new deep neural network.Only ten thousand spontaneous samples are present in each database.The convolutional neural network shows excellent performance and is thus widely used in a large number of scenes.For instance,the approach is better than the traditional feature extraction method in the aspect of improving the accuracy of discriminating the categories of spontaneous expression.Method This study proposes a method on the basis of different deep transfer network models for discriminating the categories of spontaneous expression.To preserve the characteristics of the original spontaneous expression,we do not use the technique of data enhancement to reduce the risk of convergence.At the same time,training samples,which comprise three-dimensional images that are composed of optical flow and grayscale images,are compared with the original RGB images.The threedimensional image contains spatial information and temporal displacement information.In this study,we compare three network models with different samples.The first model is based on Alexnet that only changes the number of output layer neurons that is equal to the number of categories of spontaneous expression.Then,the network is fine-tuned to obtain the best training and testing results by fixing the parameters of different layers several times.The second model is based on Inception V3.Two fully connected layers whose neuron numbers are equal to 512 and the number of spontaneous expression categories,respectively,are added to the output results.Thus,we only need to fine-tune the parameters of the two layers.Network depth increases with a reduction of the number of parameters due to the 3×3 convolution kernel replacing the 7×7 convolution kernel.The third model is based on Inception-ResNet-v2.Similar to the first model,we only change the number of output layer neurons.Finally,the isomorphic network model is proposed to identify the categories of spontaneous expression.The model is composed of two transfer learning networks of the same type that are trained by different samples and then takes the maximum as the final output.The isomorphic network makes decisions with high accuracy because the same output of the isomorphic network is infinitely close to the standard answer.From the perspective of probability,we take the maximum of different outputs as a prediction value.Result Experimental results indicate that the proposed method exhibits excellent classification performance on different samples.The single network output clearly shows that the features extracted from RGB images are as effective as the features extracted from the three-dimensional images of optical flow.This result indicates that spatiotemporal features extracted by the optical flow method can be replaced by features that are extracted from the deep neural network.Simultaneously,the method shows that at a certain degree,features extracted from the neural network can replace the lost information and features,such as the temporal features of RGB images or color features of OF+images.The high average accuracy of a single network indicates that it has good testing performance on each dataset.Networks with high complexity perform well because the samples of spontaneous expression can train the deep transfer learning network effectively.The proposed models achieve state-of-the-art performance and an average accuracy of over96%.After analyzing the result of the isomorphic network model,we know that its expression is not better than that of a single network in some cases because a single network has a high confidence degree in discriminating the categories of spontaneous expression and thus,the isomorphic network cannot easily improve the average accuracy.The Titan Xp used for this research was donated by the NVIDIA Corporation.Conclusion Compared with traditional expression,spontaneous expression is able to change subtly and extract features in a difficult manner.In the study,different transfer learning networks are applied to discriminate the categories of spontaneous expression.Concurrently,the testing accuracies of different networks,which are trained by different kinds of samples,are compared.Experimental results show that in contrast to traditional methods,deep learning has obvious advantages in spontaneous expression feature extraction.The findings also prove that deep network can extract complete features from spontaneous expression and that it is robust on different databases because of its good testing results.In the future,we will extract spontaneous expressions directly from videos and identify the categories of spontaneous expression with high accuracy by removing distracting occurrences,such as blinking.
作者 付晓峰 吴俊 牛力 Fu Xiaofeng;Wu Jun;Niu Li(School of Computer Science and Technology,Hangzhou Dianzi Universty,Hangzhou 310018. China)
出处 《中国图象图形学报》 CSCD 北大核心 2019年第5期753-761,共9页 Journal of Image and Graphics
基金 国家自然科学基金项目(61672199 61572161) 浙江省科技计划项目--2018年度重点研发计划项目(2018C01030) 浙江省自然科学基金项目(Y1110232)~~
关键词 自发表情 迁移学习 分类 神经网络 同构网络 spontaneous expression transfer learning classification neural networks isomorphic network
  • 相关文献

参考文献2

二级参考文献86

  • 1杨常清,王孝通,李博,金良安.基于特征光流的角点匹配快速算法[J].光电工程,2006,33(4):85-88. 被引量:4
  • 2薛雨丽,毛峡,张帆.BHU人脸表情数据库的设计与实现[J].北京航空航天大学学报,2007,33(2):224-228. 被引量:20
  • 3Belhumeur P N, Hespanha J P, Kriegman D J. Eigenfaces: vs. fisherfaces: recognition using class specific linear projection [ J ]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1997, 19(7) : 711-720.
  • 4Sire T, Baker S, Bsat M. The CMU pose, illumination, and expression (PIE) database [ A ] . In: Proceedings of the IEEE International Conference on Automatic Face and Gesture Recognition [C] , Washington, DC, USA, 2002: 46-51.
  • 5Martinez A M, Benavente R. The AIR face database [ R]. Technical Report 24, The Computer Vision Center (CVC), Barcelona, Spain, 1998.
  • 6Hwang B W, Rob M C, Lee S W. Performance evaluation of face recognition algorithms on Asian face database [ A ]. In: Proceedings of the Sixth IEEE International Conference on Automatic Face and Gesture Recognition [ C ], Seoul, South Korea, 2004 : 278-283.
  • 7Gau W, Cao B, Sban S, et al. The CAS-PEAL large-scale chinese face database and baseline evaluations [ J ]. IEEE Transactions on Systems, Man and Cybernetics, Part A, 2008, 38( 1 ) : 149-161.
  • 8Pantic M, Rothkrantz L J M. Automatic analysis of facial expressions: the state of the art [ J ]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2000, 22 ( 12 ) : 1424-1446.
  • 9Fasel B, Luettin J. Automatic facial expression analysis: a survey [ J]. Pattern Recognition, 2003, 36 ( 1 ) : 259-275.
  • 10Essa I, Pentland A. Coding, analysis, interpretation, and recognition of facial expressions [ J ]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1997, 19(7): 757-763.

共引文献64

同被引文献66

引证文献9

二级引证文献32

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部