期刊文献+

基于RNN-RBM语言模型的语音识别研究 被引量:27

A Study of Speech Recognition Based on RNN-RBM Language Model
下载PDF
导出
摘要 近年来深度学习兴起,其在语言模型领域有着不错的成效,如受限玻尔兹曼机(restricted Boltzmann machine,RBM)语言模型等.不同于N-gram语言模型,这些根植于神经网络的语言模型可以将词序列映射到连续空间来评估下一词出现的概率,以解决数据稀疏的问题.此外,也有学者使用递归神经网络来建构语言模型,期望由递归的方式充分利用所有上文信息来预测下一词,进而有效处理长距离语言约束.根据递归受限玻尔兹曼机神经网络(recurrent neural network-restricted Boltzmann machine,RNN-RBM)的基础来捕捉长距离信息;另外,也探讨了根据语言中语句的特性来动态地调整语言模型.实验结果显示,使用RNN-RBM语言模型对于大词汇连续语音识别的效能有相当程度的提升. In the recent years,deep learning is emerging as a new way of multilayer neural networks and back propagation training.Its application in the field of language model,such as restricted Boltzmann machine language model,gets good results.This language model based on neural network can assess the probability of the next word appears according to the word sequence which is mapped to a continuous space.This language model can solve the problem of sparse data.Besides,some scholars are constructing language model making use of recurrent neural network mode in order to make full use of the preceding text to predict the next words.From these models we can sort out the restriction of long-distance dependency in language.This paper attempts to catch the long-distance information based on RNN-RBM.On the other hand,the dynamic adjunction of language model ia analyzed and illustrated according to the language features.The experimental result manifests there are considerable improvement to the efficiency of expanding vocabulary continuing speech recognition using RNN_RBM language model.
出处 《计算机研究与发展》 EI CSCD 北大核心 2014年第9期1936-1944,共9页 Journal of Computer Research and Development
关键词 语音识别 语言模型 神经网络 递归神经网络-受限玻尔兹曼机 关联信息 speech recognition language model neural network recurrent neural network-restricted Boltzmann machine relevance information
  • 相关文献

参考文献8

  • 1Bottou L. Stochastic gradient learning in neural networks [C] //Proc of Neuro Nimes 91, Nimes: EC2, 1991: 687-699.
  • 2Boulanger-Lewandowski N, Bengio Y, Vincent P. Modeling temporal dependencies in high-dimensional sequences Application to polyphonic music generation and transcription [C] //Proc of the 29th Int Conf on Machine Learning. New York: ACM, 2012:590-598.
  • 3Xu W, Rundieky A. Can artificial neural networks learn language models? [C] //Proc of the 6th Int Conf on Spoken I,anguage Processing. Beijing, China: ISCA, 2000.
  • 4Bengio Y, Simard P, Fraseoni P. Learning long-term dependencies with gradient descent is difficult [C]//Proe of IEEE Trans on Neural Networks. Piseataway, NJ: IEEE, 1994:157-166.
  • 5Mikolov T, Kopecky J, Burger L, et al. Neural network based language models for highly inflective languages [C] // Proc of the 34th IEEE Int Conf on Acoustics, Speech and Signal. Piseataway, NJ: IEEE, 2009:4725-4728.
  • 6Bengio Y, Ducharme R, Vincent P, et al. A neural probabilistic language model [J]. Journal of Machine Learning Research, 2003, 3(2): 1137-1155.
  • 7Bengio Y, Frasconi P, Simard P. The problem of learning long-term dependencies in recurrent networks [C] //Proc of IEEE Int Conf on Neural Networks. Piscataway, NJ: IEEE, 1993: 1183-1188.
  • 8Mikolov T, Kombrink S, Deoras A, el al. RNNLM- Recurrent neural network language modeling toolkit [C] // Proc of IEEE Workshop on Automatic Speech Recognition and Understanding. Piscataway, NJ: IEEE, 2011: 5528- 5531.

同被引文献173

引证文献27

二级引证文献669

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部