期刊文献+
共找到4篇文章
< 1 >
每页显示 20 50 100
一种基于卷积神经网络的端到端语音分离方法 被引量:12
1
作者 范存航 刘斌 +2 位作者 陶建华 温正棋 易江燕 《信号处理》 CSCD 北大核心 2019年第4期542-548,共7页
大部分的语音分离系统仅仅增强混合的幅值谱(短时傅里叶变换的系数),但是对于相位谱却不做任何处理。然而,最近的研究表明相位信息对于语音分离的质量起着很重要的作用。为了同时利用幅值和相位信息,本文提出了一种有效的端到端分离方... 大部分的语音分离系统仅仅增强混合的幅值谱(短时傅里叶变换的系数),但是对于相位谱却不做任何处理。然而,最近的研究表明相位信息对于语音分离的质量起着很重要的作用。为了同时利用幅值和相位信息,本文提出了一种有效的端到端分离方法。这种方法是直接利用原始语音波行点作为特征,是一种基于编解码器的卷积神经网络结构。跟其他的说话人独立的语音分离系统不同,本文提出的方法其神经网络只输出一个说话人的信号,其他的语音可以由混合语音与网络输出信号的差值获得。我们在TIMIT数据集上验证本文提出的方法。实验结果表明,本文提出的方法明显优于句子级别的排列不变性训练(utterance-level permutation invariant training,uPIT)基线方法,对于信号失真比(signal-to-distortion ratio,SDR)相对提高了16.06%。 展开更多
关键词 说话人独立语音分离 鸡尾酒会问题 端到端 卷积编解码器
下载PDF
Blind Separation of Speech Signals Based on Wavelet Transform and Independent Component Analysis 被引量:4
2
作者 吴晓 何静菁 +2 位作者 靳世久 徐安桃 王伟魁 《Transactions of Tianjin University》 EI CAS 2010年第2期123-128,共6页
Speech signals in frequency domain were separated based on discrete wavelet transform (DWT) and independent component analysis (ICA). First, mixed speech signals were decomposed into different frequency domains by DWT... Speech signals in frequency domain were separated based on discrete wavelet transform (DWT) and independent component analysis (ICA). First, mixed speech signals were decomposed into different frequency domains by DWT and the subbands of speech signals were separated using ICA in each wavelet domain; then, the permutation and scaling problems of frequency domain blind source separation (BSS) were solved by utilizing the correlation between adjacent bins in speech signals; at last, source signals were reconstructed from single branches. Experiments were carried out with 2 sources and 6 microphones using speech signals at sampling rate of 40 kHz. The microphones were aligned with 2 sources in front of them, on the left and right. The separation of one male and one female speeches lasted 2.5 s. It is proved that the new method is better than single ICA method and the signal to noise ratio is improved by 1 dB approximately. 展开更多
关键词 wavelet transform independent component analysis blind source separation
下载PDF
Speech Separation Based on Robust Independent Component Analysis 被引量:1
3
作者 YAO Wen-po WU Min +2 位作者 LIU Tie-bing WANG Jun SHEN Qian 《Chinese Journal of Biomedical Engineering(English Edition)》 2013年第4期169-177,共9页
In this paper, we applied RobustICA to speech separation and made a comprehensive comparison to FastICA according to the separation results. Through a series of speech signal separation test, RobustICA reduced the sep... In this paper, we applied RobustICA to speech separation and made a comprehensive comparison to FastICA according to the separation results. Through a series of speech signal separation test, RobustICA reduced the separation time consumed by FastICA with higher stability, and speeches separated by RobustICA were proved to having lower separation errors. In the 14 groups of speech separation tests, separation time consumed by RobustICA was 3.185 s less than FastICA by nearly 68%. Separation errors of FastICA had a float between 0.004 and 0.02, while the errors of RobustlCA remained around 0.003. Furthermore, compared to FastICA, RobustlCA showed better separation robustness. Experimental results showed that RohustICA was successful to apply to the speech signal separation, and showed superiority to FastlCA in speech separation. 展开更多
关键词 RobustlCA speech separation FASTICA KURTOSIS optimal step size
下载PDF
COMPARISON OFEMA-SYNCHRONIZED AND STAND-ALONE SPEECH BASED ON SPEECH RECOGNITION
4
作者 FANG Qiang 《中国语音学报》 2023年第2期167-176,共10页
Synchronized acoustic-articulatory data is the basis of various applications,such as exploring the fundamental mechanisms of speech production,acoustic to articulatory inversion(AAI),and articulatory to acoustic mappi... Synchronized acoustic-articulatory data is the basis of various applications,such as exploring the fundamental mechanisms of speech production,acoustic to articulatory inversion(AAI),and articulatory to acoustic mapping(AAM).Numerous studies have been conducted based on the synchronized ElectroMagnetic Articulograhy(EMA)data and acoustic data.Hence,it is necessary to make clear whether the EMA-synchronized speech and stand-alone speech are different,and if so,how it affects the performance of the applications that are based on synchronized acoustic-articulatory data.In this study,we compare the differences between EMA-synchronized speech and stand-alone speech from the aspect of speech recognition based on the data of a male speaker.It is found that:i)the general error rate of EMA-synchronized speech is much higher than that of stand-alone speech;ii)apical vowels and apical/blade consonants are more significantly affected by the presence of EMA coils;iii)parts of vowel and consonant tokens are confused with the sounds who use the same articulator or the articulators nearby,such as confusion among apical vowels and confusion among apical and blade consonants;iv)the confusion of labial tokens demonstrates a diverse pattern. 展开更多
关键词 EMA-synchronized speech Stand-alone speech Speech recognition Confusion matrix
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部