期刊文献+

虚拟人“双簧”—与语音同步的三维人脸动画的研究 被引量:1

COLLABORATE TWO-VIRTUAL HUMAN SHOW—ON 3D FACIAL ANIMATION SYNCHRONISED WITH SPEECH
下载PDF
导出
摘要 为了有效地合成人脸语音动画,提出一种与语音同步的三维人脸口型与表情动画合成的方法。首先,根据对人脸运动解剖学的研究,构建一个基于肌肉模型与微分几何学模型的三维人脸控制模型,通过数据结构的形式去控制肌肉模型和微分几何学模型实现人脸的运动,从而实现各种口型和表情运动的变化;然后,充分考虑汉语的发音特征,提出一种基于几何描述的并且符合汉语发音习惯的协同发音模型,从而产生逼真的三维人脸语音动画。仿真结果和实验对比表明,采用该方法可以得到符合汉语发音习惯的三维人脸口型动画,且合成的三维人脸表情较为自然,逼真。 In order to effectively synthesise the 3D face and speech animation,we proposed a method of animation synthesis with speech synchronisation for 3D mouth shapes and facial expressions. First,according to the research on anatomy of facial movement,we built a 3D face control model,which is based on muscle model and differential geometry model,it controls the muscle model and differential geometry model in the form of data structure to realise facial movement,so that implements the variation of mouth shapes and expression movements.Then,fully considering the characteristics of Chinese pronunciation,we presented a coarticulation model,which is based on geometric description and compliant with Chinese pronunciation habit,so that the vivid 3D face animation with speech could be generated. Finally,it was indicated by the comparison between simulation result and experiment that to apply this method could realise the 3D facial mouth shapes animation in accordance with the habits of Chinese pronunciation,and the synthesised 3D facial expressions were more natural and vivid.
出处 《计算机应用与软件》 CSCD 2015年第8期145-149,173,共6页 Computer Applications and Software
基金 国家自然科学基金面上项目(61371165) 计算机软件新技术国家重点实验室开放课题(KFKT2013B22) 浙江大学CAD&CG国家重点实验室开放课题(A1416) 四川省动漫研究中心2012年度科研项目(DM201204)
关键词 肌肉模型 微分几何学 协同发音建模 语音动画 Muscle model Differential geometry Coarticulation modelling Speech animation
  • 相关文献

参考文献17

  • 1Wang L J,Han W,Soong F K. High quality lip-sync animation for 3Dphoto-realistic talking head [ C ]//Proceedings of IEEE InternationalConference Acoustics, Speech, and Signal Processing, 2012 : 4529-4532.
  • 2Han W,Wang L J,Soong F K,et al. Improved minimum converted traj-ectory error training for real-time speech-to-lips conversion[ C ]//Pro-ceedings of IEEE International Conference Acoustics,Speech,and Sig-nal Processing,2012:4513 -4516.
  • 3Zoric G, Forchheimer G, Pandzic I S. On creating multimodal virtualhumans-real time speech driven facial gesturing[ J]. Multimedia Toolsand Applications,2011,54( 1 ) : 165 ~ 179.
  • 4Deena S,AGalata. Speech-driven facial animation using a sharedGaussian process latent variable model [ C ] //Proceedings of 5 th Inter-national Symposium Advances in Visual Computing,2009 :89 ~ 100.
  • 5Deena S,Hou S, AGalata. Visual speech synthesis by modeling coartic-ulation dynamics using a non-parametric switching state-space model[C ]//Proceedings of International Conference Multimodal Interfacesand the workshop on Machine Learning for Multimodal Interaction,2010:1 -8.
  • 6Mariooryad S, Busso C. Generating human-like behaviors using jointspeech-driven models for conversational agents( J ]. IEEE Transactionson Audio, Speech, and Language Processing, 2012,20 ( 8 ) : 2329-2340.
  • 7Jia J,Zhang S,Meng F B,et al. Emotional audio-visual speech synthe-sis based on PAD[ J]. IEEE Transactions on Audio,Speech,and Lan-guage Processing,2011,19(3) :570 -582.
  • 8Wu Z Y,Meng H M,Yang H W ,et al. Modeling the expressivity of in-put text semantics for chinese text-to-speech synthesis in a spoken dia-log system [ J ]. IEEE Transactions on Audio, Speech, and LanguageProcessing,2009,17(8) :1567-1576.
  • 9Zhang S,Wu Z Y,Meng H M,et al. Facial expression synthesis basedon emotion dimensions for affective talking avatar [ J ]. Smart Innova-tion ,Systems and Technologies ,2010(1) :109 —132.
  • 10Tang H,Chu S,Hasegawa-Johnson M,et al. Partially supervised speak-er clustering^ J] . IEEE Transactions on Pattern Analysis and MachineIntelligence,2012,34(5) :959 -971.

二级参考文献28

  • 1曹剑芬.普通话双音子和三音子结构系统代表语料集[J].语言文字应用,1997(1):62-70. 被引量:7
  • 2徐向华,朱杰,郭强.汉语连续语音识别中的分级聚类算法的研究和应用[J].信号处理,2004,20(5):497-500. 被引量:2
  • 3王志明,蔡莲红,艾海舟.基于数据驱动方法的汉语文本-可视语音合成(英文)[J].软件学报,2005,16(6):1054-1063. 被引量:16
  • 4S Baek, B kim, K Lee. 3D Face Model Reconstruction From Single 2D Frontal hnage [ J ]. VRCAL, 2009:96-98.
  • 5S F Wang, S H Lai. Reconstucting 3D Face Model with Associated Expression Deformation from a Single Face Image via Constructing a Low- Dimensional Expression Deformation Manifold [ J ]. PAMI, 2011,33(10) : 2115-2120.
  • 6X Fan, Q Peng, M Zhong. 3D face reconstruction from single 2D image based on robust facial feature, points extraction and generic wire frame model [J]. Communications and Mobile Computing, 2010 : 396 -400.
  • 7Z Chen, J Hou, D Zhang, X Qin. An annotation rule extraction algorithm for image retrieval [ J ]. Pattern Recognition Letters, 2012,33(10) :1257-1268.
  • 8F Hu, Y Lin, B Zou, M Zhang. Individual 3D Face Generation Based on Candide-3 for Face Recognition [ J ]. Image and Signal Processing, 2008:646-648.
  • 9TANG Hao,FU Yun,TU Ji-lin. Humanoid audio-visual avatar with emotive text-to-speech synthesis[J].IEEE Transactions on Multimedia,2008,(06):969-981.
  • 10TANG Hao,FU Yun,TU Ji-lin. EAVA:a 3 D emotive audio-visual avatar[A].Washington,DC:IEEE Computer Society,2008.1-6.

共引文献15

同被引文献9

引证文献1

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部