期刊文献+
共找到1篇文章
< 1 >
每页显示 20 50 100
MusicFace: Music-driven expressive singing face synthesis
1
作者 Pengfei liu Wenjin Deng +5 位作者 hengda li Jintai Wang Yinglin Zheng Yiwei Ding Xiaohu Guo Ming Zeng 《Computational Visual Media》 SCIE EI CSCD 2024年第1期119-136,共18页
It remains an interesting and challenging problem to synthesize a vivid and realistic singing face driven by music. In this paper, we present a method for this task with natural motions for the lips, facial expression... It remains an interesting and challenging problem to synthesize a vivid and realistic singing face driven by music. In this paper, we present a method for this task with natural motions for the lips, facial expression, head pose, and eyes. Due to the coupling of mixed information for the human voice and backing music in common music audio signals, we design a decouple-and-fuse strategy to tackle the challenge. We first decompose the input music audio into a human voice stream and a backing music stream. Due to the implicit and complicated correlation between the two-stream input signals and the dynamics of the facial expressions, head motions, and eye states, we model their relationship with an attention scheme, where the effects of the two streams are fused seamlessly. Furthermore, to improve the expressivenes of the generated results, we decompose head movement generation in terms of speed and direction, and decompose eye state generation into short-term blinking and long-term eye closing, modeling them separately. We have also built a novel dataset, SingingFace, to support training and evaluation of models for this task, including future work on this topic. Extensive experiments and a user study show that our proposed method is capable of synthesizing vivid singing faces, qualitatively and quantitatively better than the prior state-of-the-art. 展开更多
关键词 face synthesis SINGING MUSIC generative adversarial network
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部