期刊文献+

基于ListMLE排序学习方法的机器译文自动评价研究 被引量:2

Research on ListMLE Approach to Ranking for Automatic Machine Translation Evaluation
下载PDF
导出
摘要 机器翻译译文质量的自动评价是推动机器翻译技术快速发展的一条重要途径。该文提出了基于List-MLE排序学习方法的译文自动评价方法。在此基础上,探讨引入刻画译文流利度和忠实度的特征,来进一步提高译文自动评价结果和人工评价结果的一致性。实验结果表明,在评价WMT11德英任务和IWSLT08BTEC CEASR任务上的多个翻译系统的输出译文质量时,该文提出的方法预测准确率高于BLEU尺度和基于RankSVM的译文评价方法。 Automatic evaluation of machine translation plays an important role in promoting the rapid development of machine translation. In this paper, we apply the ListMLE approach to learning to rank for machine translation auto- matic evaluation. In addition, we introduce the features of translation fluency and adequacy to further improve the consistency between the results of the automatic evaluation and human judgments. When assess the translation qual- ity of the submitted system outputs of WMT'll German-English tasks and IWSLT'08 BTEC CE ASR tasks, the ex- perimental results indicate that the predicted accuracy of the proposed approach is higher than the BLEU metric and the one based on RankSVM.
出处 《中文信息学报》 CSCD 北大核心 2013年第4期22-29,共8页 Journal of Chinese Information Processing
基金 国家自然科学基金资助项目(61203313 61272212 61163006) 江西省教育厅自然科学基金资助项目(GJJ12212)
关键词 机器译文评价 排序学习 ListMLE方法 人工评价 自动评价 machine translation evaluation learning to rank ListMLE approach automatic evaluation human eval-uation
  • 相关文献

参考文献28

  • 1Zhifei Li , Chris Callison-Burch, Chris Dyer, et al. An Open Source Toolkit for Parsing-based Machine Trans?lation[CJ/ /Proceedings of the WMT. 2009.
  • 2Callison-Burch Chris, Philipp Koehn, Christof Monz , et al. Findings of the 2011 Workshop on Statistical Machine Translation[CJ/ /Proceedings of the WMT, Edinburgh, Scotland, UK, 2011: 22-64.
  • 3Chang Liu , Daniel Dahlmeier, Hwee Tou Ng. Better Evaluation Metrics Lead to Better Machine Translation[CJ/ /Proceedings of the EMNLP, Edinburgh, Scot?land, UK, 2011: 375-384.
  • 4Kishore Papineni , Salim Roukos, Todd Ward, et al. BLEU: a Method for Automatic Evaluation of Machine Translation[CJ/ /Proceedings of the ACL, Philadel?phia, Pennsylvania, 2002: 311-318.
  • 5Callison-Burch Chris, M Osborne, Philipp Koehn. Re- ?evaluating the role of BLEU in machine translation re?search[CJ/ /Proceedings of EACL, 2006: 249-256.
  • 6David Chiang, Steve DeNeefe, Yee Seng Chan, et al. Decomposability of translation metrics for improved e?valuation and efficient algorithms[CJ/ /Proceedings of the EMNLP, Honolulu, Hawaii, 2008:610-619.
  • 7George Doddington. Automatic Evaluation of Machine Translation Quality Using N-gram Cooccurrence Sta?tistics[CJ//Proceedings of the HLT, San Diego, Cali?fornia, CA, USA, 2002: 138-145,.
  • 8Satanjeev Banerjee, Alon Lavie. METEOR: An Auto?matic Metric for MT Evaluation with Improved Corre?lation with HumanJudgments[CJ/ /Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summariza?tion, Ann Arbor, 2005:65-72.
  • 9I Dan Melamed, Ryan Green,Joseph P Turian. Preci?sion and Recall of Machine Translation[CJ/ /Proceed?ings of the HLT-NAACL, Edmonton, Canada, 2003: 61-63.
  • 10Matthew Snover, Bonnie Dorr , Richard Schwartz, et al. A Study of Translation Edit Rate with Targeted Human Annotation[CJ//Proceedings of AMTA, Cambridge, 2006: 223-231.

同被引文献8

引证文献2

二级引证文献10

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部