期刊文献+

结合区分性训练深度神经网络的歌声与伴奏分离方法 被引量:1

A separation method of singing and accompaniment combining discriminative training deep neural network
下载PDF
导出
摘要 针对音乐信号中的歌声与伴奏相互关联难以分离的问题,提出了一种区分性训练深度神经网络(Deep Neural Network,DNN)的音乐分离方法。首先,在DNN模型的基础上同时考虑歌声与伴奏间的重建误差和区分性信息,提出了一种改进的目标函数进行区分性训练;其次,在DNN模型上额外添加一层,引入时频掩蔽对估计出的歌声伴奏进行联合优化,相应的时域信号由傅里叶逆变换获得;最后,验证不同参数设置对分离性能的影响,并与现有的音乐分离方法进行对比.实验结果表明,改进的目标函数和时频掩蔽的引入明显提高了DNN的分离性能,且与现有的音乐分离方法相比分离性能最高提高了4 dB从而证实所提方法是一种有效的音乐分离方法。 For the difficulty of separation between singing and accompaniment in the musical signals, an improved music separation method of based on discriminative training Depth Neural Network(DNN) was proposed. Firstly,based on the DNN model, considering the reconstruction errors and discrimination information between singing and accompaniment, an improved objective function was presented to discriminate the training;Then, an additional layer was added to DNN model, introducing the time-frequency masking to optimize the estimated accompaniment of the song, and the corresponding time-domain signal was obtained by inverse Fourier transform;Finally, the influence of different parameters on the separation performance was verified, and compared it with the existing music separation methods. The experimental results showed that the improved objective function and the introduction of time-frequency masking significantly improved the separation performance of the DNN, and the separation performance was improved about 4 dB compared with other existing music separation methods, thus verifying that the proposed method was an effective music separation algorithm.
作者 张天骐 熊梅 张婷 杨强 ZHANG Tianqi;XIONG Mei;ZHANG Ting;YANG Qiang(Chongqing Key Laboratory of Signal and Information Processing^ Chongqing University of Posts and Telecommunications Chongqing 400065)
出处 《声学学报》 EI CSCD 北大核心 2019年第3期393-400,共8页 Acta Acustica
基金 国家自然科学基金项目(61671095 61371164 61702065 61701067 61771085) 信号与信息处理重庆市市级重点实验室建设项目(CSTC2009CA2003) 重庆市研究生科研创新项目(CYS17219) 重庆市教育委员会科研项目(KJ130524 KJ1600427 KJ1600429)资助
  • 相关文献

参考文献5

二级参考文献24

  • 1TAO Ran,DENG Bing,WANG Yue.Research progress of the fractional Fourier transform in signal processing[J].Science in China(Series F),2006,49(1):1-25. 被引量:99
  • 2邹霞,陈亮,张雄伟.基于Gamma语音模型的语音增强算法[J].通信学报,2006,27(10):118-123. 被引量:11
  • 3Kim Get al. An algorithm that improves speech intelligi- bility in noise for normal-hearing listeners. J Acoust. Soc. Am., 2009; 126:1486--1494.
  • 4Wang D L, Brown G J. Separation of speech from interfer ing sounds based on oscillatory correlation. IEEE Trans Neural Netw., 1999; 10(3): 684--697.
  • 5Ozerov A, F6votte C. Multichannel nonnegative matrix fac- torization in convolutive mixtures for audio source separa- tion. IEEE Transactions on Audio, Speech, and Language Processing, 2010; 18(3): 550--563.
  • 6Ozerov A, Vincent E, Bimbot F. A general flexible frame: work for the handling of prior information in audio source separation. IEEE Transactions on Audio, Speech, and Language Processing, 2012; 20(4): 1118--1133.
  • 7Antoine Liutkus, Zafer Rafii, Roland Badeau, Bryan Pardo, Gael Richard. Adaptive filtering for music/voice separation exploiting the repeating musical structure. In: IEEE International Conference on Acoustics, Speech and Signal Processing, Kyoto, Japan, 2012:53--56.
  • 8Zafer Rafii, Bryan Pardo. Online REPET-SIM for real- time speech enhancement. 38th International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 2013:848--852.
  • 9Rafii Z, Pardo B. Repeating pattern extraction technique (REPET): A simple method for music/voice separation. IEEE Transactions on Audio, Speech, and Language Pro- cessing, 2013; 21(1): 73--84.
  • 10Foote J, Uchihashi S. The beat spectrum: a new approach to rhythm analysis. IEEE International Conference oil Multimedia & Expo, Tokyo, Japan, 2001:1088--1091.

共引文献49

同被引文献14

引证文献1

二级引证文献1

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部