摘要
近年来,神经网络语言模型的研究越来越受到学术界的广泛关注。基于长短期记忆(long short-term memory,LSTM)结构的深度神经网络(LSTM-deep neural network,LSTM-DNN)语言模型成为当前的研究热点。在电话交谈语音识别系统中,语料本身具有一定的上下文相关性,而传统的语言模型对历史信息记忆能力有限,无法充分学习语料的相关性。针对这一问题,基于LSTM-DNN语言模型在充分学习电话交谈语料相关性的基础上,将其应用于语音识别系统的重评估过程,并将这一方法与基于高元语言模型、前向神经网络(feed forward neural network,FFNN)以及递归神经网络(recurrent neural network,RNN)语言模型的重评估方法进行对比。实验结果表明,LSTMDNN语言模型在重评估方法中具有最优性能,与一遍解码结果相比,在中文测试集上字错误率平均下降4.1%。
In recent years,the research on the neural network language model has received more and more attention from the academic circles. At present,the neural network language model based on LSTM structure has become a research hotspot. In the speech recognition system,the corpus itself has certain relevance. But the traditional language models have limited memory capacity,and they cannot fully learn the relevance of the corpus. To solve this problem,a novel LSTM-DNN language model is applied to the revaluation of speech recognition,which fully exploits the correlation on a telephone conversation corpus. It is further compared with existing revaluation methods based on language models such as high order language model,feed forward neural network( FFNN) language model and recurrent neural network( RNN) language model. The experimental results show that the performance of LSTM-DNN language model is optimal. Compared to the first pass of decoding,the relative decline in average word error rate is 4. 1% in the Chinese test sets.
出处
《重庆邮电大学学报(自然科学版)》
CSCD
北大核心
2016年第2期180-186,193,共8页
Journal of Chongqing University of Posts and Telecommunications(Natural Science Edition)
基金
国家自然科学基金(10925419
90920302
61072124
11074275
11161140319
91120001
61271426)
中国科学院战略性先导科技专项(XDA06030100
XDA06030500)
国家863计划(2012AA012503)
中科院重点部署项目(KGZD-EW-103-2)~~
关键词
长短期记忆
神经网络语言模型
语音识别
重评估
long short-term memory
neural network language model
speech recognition
revaluation