期刊文献+

手语合成中的多模式行为协同韵律模型 被引量:9

Multi-Model Behavior Synchronizing Prosody Model in Sign Language Synthesis
下载PDF
导出
摘要 利用大量真实多模式行为数据进行学习训练、获取单模式行为的韵律模型以及多模式行为之间的协同关联模型的方法,来实现虚拟人多模式行为之间的协同.重点给出了多模式行为的韵律模型描述,同时给出基于手语韵律参数与语音韵律特征融合的协同控制韵律模型以及韵律参数获取方法,并运用于多模式行为协同控制中,取得了较好的实验结果.与传统的规则法相比,该学习方法更能刻画多模式之间协同关联的复杂性,更好地实现虚拟人多模式行为合成的逼真性. This paper proposes a multi-model behavior synchronizing prosody model and its application to Chinese sign language synthesis. Based on huge realistic multi-model behavior training data, the authors adopt learning the prosody mode for each single channel behavior data and further synchronizing relation model of all models, and present the framework for the multi-mod- el synchronization in virtual human synthesis, including models of sign language, speech, facial expression and lip movement and so on. The formal description of multi-model prosody model is demonstrated in detail. Comparing to traditional regularity approaches, the learning based ap- proach in this paper is more adequate to express complicatedly the multi-model synchronizing rela- tionship, and to synthesize realistically the multi-model behavior of the virtual human. As the example, the synchronizing prosody model involving sign language prosody parameters and speech prosody parameters is given. The authors design an approach to compute the prosody parameters and apply it to control the virtual human's multi-model behavior synchronously. Experiments based on the Coss ("863" speech material library) and Chinese sign language library show that the multi-model behavior synchronizing prosody model works well. It enhances the recognition rate of synthetic sign language by 5.94%.
出处 《计算机学报》 EI CSCD 北大核心 2006年第5期822-827,共6页 Chinese Journal of Computers
基金 国家自然科学基金(60303018 60403037) 北京市科技新星计划项目基金(2005B54) 北京工业大学多媒体与智能软件技术实验室开放课题基金联合资助.
关键词 手语合成 多模式 韵律模型 sign language synthesis multi model prosody model
  • 相关文献

参考文献8

  • 1Cassell J,Vilhjalmsson H,Bickmore T..BEAT:The behavior expression animation toolkit.In:Proceedings of the International Conference on Computer Graphics and Interactive Techniques (SIGGRAPH'2001),2001,205~216
  • 2Kshirsagar S,Thalmann N,Vuilleme A,Thalmann D,Kamyab K,Mamdani E..Avatar markup language.In:Proceedings of 8th Eurographics Workshop on Virtual Environments,2002,169~177
  • 3Brand M,Hertzmann A..Style machines.In:Proceedings of the International Conference on Computer Graphics and Interactive Techniques (SIGGRAPH2000),2000,183~192
  • 4Li Yan,Wang Tian-Shu,Shum Heung-Yeung.Motion texture:A two level statistical model for character motion synthesis.In:Proceedings of the International Conference on Computer Graphics and Interactive Techniques (SIGGRAPH2002),2002,465~472
  • 5Badler N,Allbeck J,Zhao Li-Wei,Byun M..Representing and parameterizing agent behaviors.In:Proceedings of the Computer Animation 2002 Geneva,Switzerland,2002,133~144
  • 6Cassell Justine,Pelachaud Catherine et al.Animated conversation:Rule-based generation of facial expression,gesture & spoken intonation for multiple conversational agents.In:Proceedings of the 21st Annual Conference on Computer Graphics and Interactive Techniques (SIGGRAPH94),1994,413~420
  • 7Joint Research & Development Laboratory for Advanced Computer and Communication Technologies.The Version 2.0 of Chinese Sign Language Synthesis System Achieve Success in Four Deaf Schools of Beijing.http://sy.jdl.ac.cn/yjjz/dt3.asp
  • 8Chen Yi-Qiang,Gao Wen,Zhu Ting-Shao,Ling Charles.Learning prosodic patterns for mandarin text to speech.The Journal of Intelligent Information System,2002,19(1):95~109

同被引文献60

引证文献9

二级引证文献238

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部