期刊文献+

支持表情细节的语音驱动人脸动画 被引量:2

Towards Expressively Speech-Driven Facial Animation
下载PDF
导出
摘要 针对语音驱动人脸动画中如何生成随语音运动自然呈现的眨眼、抬眉等表情细节以增强虚拟环境的沉浸感的问题,提出一种可以合成表情细节的语音驱动人脸动画方法.该方法分为训练与合成2个阶段.在训练阶段,首先对富有表情的三维人脸语音运动捕获数据特征进行重采样处理,降低训练数据量以提升训练效率,然后运用隐马尔可夫模型(HMM)学习表情人脸语音运动和同步语音的关系,同时统计经过训练的HMM在训练数据集上的合成余量;在合成阶段,首先使用经过训练的HMM从新语音特征中推断与之匹配的表情人脸动画,在此基础上,根据训练阶段计算的合成余量增加表情细节.实验结果表明,文中方法比已有方法计算效率高,合成的表情细节通过了用户评价验证. In order to synthesize facial expression details such as eye blinking and eyebrow-lifting, this paper presents a new speech-driven facial animation method. This method contains two phases, training and synthesizing phases. During the training phase, 3D facial animation data is firstly re sampled to improve training efficiency. Then a hidden Markov model (HMM) is utilized to study the correlation between the expressive facial animation features and the synchronized speeches. At the same time, statistical data of the synthetic residuals is collected from the trained HMM. During the synthesizing phase, firstly, the trained HMM is used to estimate the matching expressive facial animation from the novel input speech features. Secondly, based on the estimated animation, expression details are synthesized using the collected residuals. Numerical and user study experiments show that this method outperforms conventional approaches both in the efficiency and animation quality.
出处 《计算机辅助设计与图形学学报》 EI CSCD 北大核心 2013年第6期890-899,共10页 Journal of Computer-Aided Design & Computer Graphics
基金 国家自然科学基金(60970086) 国家自然科学基金重点项目(U0935003)
关键词 人脸表情动画 隐马尔可夫模型 expressive facial animation hidden Markov model (HMM)
  • 相关文献

参考文献16

  • 1Mehrabian A. Silent messages[M]. Belmont: Wadsworth, 1971.
  • 2CassellJ, Pelachaud C, Badler N, et al. Animated conversation: rule-based generation of facial expression, gesture &. spoken intonation for multiple conversational agent[CJ !/Computer Graphics Proceedings, Annual Conference Series, ACM SlGGRAPH. New York: ACM Press. 1994: 413-420.
  • 3Sifakis E, Selle A, Robinson-Mosher A, et al. Simulating speech with a physics-based facial muscle model[CJ II Proceedings of the ACM SIGGRAPH/Eurographics Symposium on Computer Animation. Aire-Ia-Ville , Eurographics Association Press, 2006: 261-270.
  • 4Cao v, Faloutsos r. Pighin F. Unsupervised learning for speech motion editing[CJ II Proceedings of the ACM SIGGRAPH/Eurographics Symposium on Computer Animation. Aire-la - Ville: Eurographics Association Press, 2003: 225-231.
  • 5Deng Z, Neumann U, LewisJ P, et al. Expressive facial animation synthesis by learning speech co-articulation and expression spaces[J]. IEEE Transactions on Visualization and Computer Graphics, 2006, 12(6): 1523-1534.
  • 6Chuang E S, Deshpande H, Bregler C. Facial expression space learning[CJ IIProceedings of the 10th Pacific Conference on Computer Graphics and Applications. Washington DC: IEEE Computer Society Press, 2002: 68-76.
  • 7Blanz V, Basso C, Poggio T, et al . Reanimating faces in images and video[J]. Computer Graphics Forum, 2003, 22 (3): 641-650.
  • 8Brand M. Voice puppetry[CJ II Computer Graphics Proceedings, Annual Conference Series, ACM SIGGRAPH. New York: ACM Press, 1999: 21-28.
  • 9Busso C, Narayanan S S. Interrelation between speech and facial gestures in emotional utterances: a single subject study [J]. IEEE Transactions on Audio, Speech, and Language Processing, 2007, 15(8): 2331-2347.
  • 10Iu E, LeeJ. Expressive facial gestures from motion capture data[J]. Computer Graphics Forum, 2008, 27(2): 381-388.

二级参考文献13

  • 1孙权森,曾生根,杨茂龙,王平安,夏德深.基于典型相关分析的组合特征抽取及脸像鉴别[J].计算机研究与发展,2005,42(4):614-621. 被引量:29
  • 2贺云辉,赵力,邹采荣.一种基于KCCA的小样本脸像鉴别方法[J].应用科学学报,2006,24(2):140-144. 被引量:8
  • 3Mehrabian A.Silent messages[OL].[2010-07-07].http://en.wikipedia.org/wiki/Albert_Mehrabian.
  • 4Busso C,Narayanan S S.Interrelation between speech and facial gestures in emotional utterances:a single subject study[J].IEEE Transactions on Audio,Speech,and Language Processing,2007,15(8):2331-2347.
  • 5Busso C,Deng Z G,Neumann U,et al.Natural head motion synthesis driven by acoustic prosodic features[J].Computer Animation and Virtual Worlds,2005,16(3/4):283-290.
  • 6Deng Z G,Neumann U,Lewis J P,et al.Expressive facial animation synthesis by learning speech co-articulation and expression spaces[J].IEEE Transactions on Visualization and Computer Graphics,2006,12(6):1523-1534.
  • 7Chuang E S,Deshpande H,Bregler C.Facial expression space learning[C] //Proceedings of the 10th Pacific Conference on Computer Graphics and Applications.Los Alamitos:IEEE Computer Society Press,2002:68-76.
  • 8Cao Y,Faloutsos P,Pighin F.Unsupervised learning for speech motion editing[C] //Proceedings of ACM SIGGRAPH/Eurographics Symposium on Computer Animation.Aire-la-Ville:Eurographics Association Press,2003:225-231.
  • 9Brand M.Voice puppetry[C] //Proceedings of the 26th Annual Conference on Computer Graphics and Interactive Techniques.New York:ACM Press,1999:21-28.
  • 10Hotelling H.Relations between two sets of variates[J].Biometrika,1936,28(3/4):321-377.

共引文献3

同被引文献29

  • 1Chuang E, Bregler C. Performance driven facial animation us- ing blendshape interpolation[R]. Santa Clara: Stanford Univer- sity. Computer Science Department, 2002.
  • 2Zhao H, Ta C L. Subtle facial animation transfer from 2D vid- eos to 3d faces with Laptacian deformation[OL]. [2014-09-08]. http://www.cs.ust.hk/-taicl/proj ects/ExpSpeech.pdf.
  • 3Garrido P, Valgaerts L, Wu C L, et al. Reconstructing detailed dynamic face geometry from monocular video[J]. ACM Trans- actions on Graphics, 2013, 32(6): Article No. 158.
  • 4Zeng M, Liang L, Liu X G, et al. Video-driven state-aware fa- cial animation[J]. Computer Animation & Virtual Worlds, 2012, 23(3/4): 167-178.
  • 5Dornaika F, Davoine E On appearance based lace and facial action tracking[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2006, 16(9): 1107-1124.
  • 6Cao C, Weng Y L, Zhou S, et al. FaceWarehouse: a 3D facial expression database for visual computing[J]. IEEE Transactions on Visualization and Computer Graphics, 2014,20(3): 413-425.
  • 7Chai J X, Xiao J, Hodgins J. Vision-based control of 3D facial animation[C] //Proceedings of the ACM SIGGRAPH/Euro- graphics Symposium on Computer Animation. Aire-la-Ville: Eurographics Association Press, 2003:193-206.
  • 8Rhee T, Hwang Y, Kiln J D, et al. Real-time facial animation from live video tracking[C] //Proceedings of ACM SIGGRAPH/Eurographics Symposium on Computer Anima- tion. New York: ACM Press, 2011 : 215-224.
  • 9Deng Z G, Chiang P Y, Fox P, et al. Animating blendshape fac- es by cross-mapping motion capture data[C] //Proceedings of the Symposium on Interactive 3D Graphics and Games. New York: ACM Press, 2006:43-48.
  • 10Zhang L, Snavely N, Curless B, et al. Spacetime faces: high resolution capture for modeling and animation[J]. ACM Trans- actions on Graphics, 2004, 23(3): 548-558.

引证文献2

二级引证文献7

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部