期刊文献+

基于词向量特征的循环神经网络语言模型 被引量:38

Recurrent Neural Network Language Model Based on Word Vector Features
下载PDF
导出
摘要 循环神经网络语言模型能解决传统N-gram模型中存在的数据稀疏和维数灾难问题,但仍缺乏对长距离信息的描述能力.为此文中提出一种基于词向量特征的循环神经网络语言模型改进方法.该方法在输入层中增加特征层,改进模型结构.在模型训练时,通过特征层加入上下文词向量,增强网络对长距离信息约束的学习能力.实验表明,文中方法能有效提高语言模型的性能. The recurrent neural network language model( RNNLM) solves the problems of data sparseness and dimensionality disaster in traditional N-gram models. However, the original RNNLM is still lack of long dependence due to the vanishing gradient problem. In this paper, an improved method based on contextual word vectors is proposed for RNNLM. To improve the structure of models, a feature layer is added into the input layer. Contextual word vectors are added into the model with feature layer to reinforce the ability of learning long-distance information during the training. Experimental results show that the proposed method effectively improves the performance of RNNLM.
出处 《模式识别与人工智能》 EI CSCD 北大核心 2015年第4期299-305,共7页 Pattern Recognition and Artificial Intelligence
基金 国家863计划项目(No.2012AA011603) 国家自然科学基金项目(No.61175017)资助
关键词 语音识别 语言模型 循环神经网络 词向量 Speech Recognition Language Model Recurrent Neural Network Word Vector
  • 相关文献

参考文献17

  • 1Schwenk H. Continuous Space Language Models. Computer Speech and Language, 2007, 21 (3) : 492-518.
  • 2Bengio Y, Ducharme R, Vincent P, et al. A Neural Probabilistie Language Model. Journal of Machine Learning Research, 2003, 3 : 1137-1155.
  • 3Mikolov T, Karafiett M, Burger L, et al. Recurrent Neural Network Based Language Model//Proc of the 11 th Annual Conference of the International Speech Communication Association. Makuhari, Japan, 2010:1045-1048.
  • 4Mikolov T, Kombrink S, Burget L, et al. Extensions of Recurrent Neural Network Language Model// Proc of the IEEE International Conference on Acoustics , Speech and Signal Processing . Prague ,Czech Republic, 2011 : 5528-5531.
  • 5Bengio Y, Simard P, Frasconi P. Learning Long-Term Dependen- cies with Gradient Descent Is Difficult. IEEE Trans on Neural Net- works, 1994, 5(2): 157-166.
  • 6Son L H, Allauzen A, Yvon F. Measuring the Influence of Long Range Dependencies with Neural Network Language Models//Prec of the NAACL-HLT Workshop : Will We Ever Really Replace the N- gram Model.'? On the Future of Language Modeling for HLT. Man- treal, Canada, 2012:1-10.
  • 7Martens J, Sutskever I. Learning Recurrent Neural Networks with Hessian-Free Optimization [ EB/OL ]. [ 2014 - 02 - 10 ]. http:// www. icml-2011, org/papers/532_icmlpaper, pdf.
  • 8Sundermeyer M, Schltlter R, Ney H. LSTM Neural Networks for Lan- guage Modeling[EB/OL]. [2014-02-10]. http://www-i6, informatik. rwth- aachen, de/publications/download/820/Sundermeycr - 2012. pdf.
  • 9Shi Y, Wiggers P, Jonker C M. Towards Recurrent Neural Networks Language Models with Linguistic and Contextual Features//Proe of the 13th Annual Conference of the International Speech Communica- tion Association. Portland, USA, 2012:1664-1667.
  • 10Auli M, Galley M, Quirk C, et al. Joint Language and Translation Modeling with Recurrent Neural Networks // Proc of the Confe- rence on Empirical Methods in Natural Language Processing. Sea- ttle, USA, 2013:1044-1054.

同被引文献259

引证文献38

二级引证文献870

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部