期刊文献+

融合依存句法和LSTM的神经机器翻译模型 被引量:3

Neural Machine Translation Model Combining Dependency Syntax and LSTM
下载PDF
导出
摘要 针对神经机器翻译模型Transformer缺少语言学知识,以及位置信息编码方式不够灵活等问题,引入依存句法分析和长短时记忆网络LSTM(long short-term memory,LSTM),在神经机器翻译系统中构建源语言句法结构信息,并且利用LSTM的记忆特性获取更加准确的位置信息。给出一个源语言序列,采用依存句法树来将其转化为相应的依存关系矩阵,最后采用CBOW(continuous bag-of-words model,CBOW)模型根据依存关系词和上下文词预测目标词获取词向量。输出的词向量使用LSTM模型训练,将每一个时间步的输出与原序列拼接作为输入序列。实验结果表明:在WMT17汉语-英语语言对的翻译任务中,使用改进后的模型提升了0.93个BLEU点。 For the problem of the lack of linguistic knowledge,and the insufficient flexibility of positional encoding in the neural machine translation model which is called Transformer,this paper introduces dependency syntax analysis and the long short-term memory network(Long short-term memory,LSTM)to construct the Source language syntactic structure information in the neural machine translation system,and uses the memory characteristics of LSTM to obtain more accurate positional information.Given a source language sequence,transform it into the corresponding dependency matrix by dependency syntax tree,and finally use the CBOW(Continuous bag-of-words model,CBOW)model to predict the target word based on the dependency word and the context word to obtain the word embedding.Use the LSTM model to train the output word embedding,and concatenate the output of each step with the original sequence as the input sequence.Experiments show that in the WMT17 Chinese-English language pair translation task,using the improved model improves 0.93BLEU points.
作者 郑鑫 陈海龙 马玉群 王青 ZHENG Xin;CHEN Hailong;MA Yuqun;WANG Qing(School of Computer Science and Technology,Harbin University of Science and Technology,Harbin 150080,China)
出处 《哈尔滨理工大学学报》 CAS 北大核心 2023年第3期20-27,共8页 Journal of Harbin University of Science and Technology
基金 国家自然科学基金面上项目(61772160)。
关键词 神经机器翻译 TRANSFORMER 长短时记忆网络 依存句法 neural machine translation transformer long and short-term memory network dependency syntax
  • 相关文献

参考文献5

二级参考文献6

共引文献219

同被引文献29

引证文献3

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部