期刊文献+

基于融合模型的中文病历文本智能纠错研究

Research on Chinese Medical Records Error Detection Based on Ensemble Model
下载PDF
导出
摘要 目的针对中文医学病历文本,进行错别字智能纠错,以改善中文病历质量,减少诊疗文书差错产生的概率。方法将统计语言模型和基于神经网的预训练模型相融合,进行中文病历文本错别字纠错的训练和验证,最终通过综合指标F1进行模型效果的评估。结果实验结果显示,融合模型的中文病历错别字纠错F1为0.6254,优于单统计语言模型和单预训练模型的F1值0.4813和0.5970。结论基于统计语言模型和预训练模型的融合方法,在中文病历文本错别字纠错方面有较好的效果,对临床病历书写质量的保障有一定的现实辅助意义。 Objective In order to improve the quality of Chinese medical records and reduce the probability of errors in medical documents,this study intends to carry out intelligent error correction for Chinese medical records.Methods Combinedwiththe statistical language model and neural network-based pre-train model,the ensemble model wasused to train and verify the error correction of Chinese medical records.The performance was evaluated by F1,a comprehensivemetric.Results The experimental results showed that F1 of the ensemble model was 0.6254,which was better than that of the single statistical language model and the single pre-train model,with F1 values of 0.4813 and 0.5970 respectively.Conclusion The fusion model based on statistical language model and pre-train model has a good effect in the error correction of Chinese medical record text,and has a certain practical significance for the quality guarantee of clinical medical record writing.
作者 姜会珍 焦雪莹 邹凌伟 许仕杰 朱卫国 JIANG Huizhen;JIAO Xueying;ZOU Lingwei;XU Shijie;ZHU Weiguo(Peking Union Medical College Hospital,Chinese Academy of Medical Sciences&Peking Union Medical College,Beijing100730,China)
出处 《中国卫生信息管理杂志》 2023年第3期448-453,共6页 Chinese Journal of Health Informatics and Management
基金 中国医学科学院医学与健康科技创新工程(项目编号:2021-I2M-1-056)。
关键词 自然语言处理 中文病历 错别字纠错 融合模型 natural language processing Chinese medical record error detection model ensemble
  • 相关文献

参考文献4

二级参考文献30

  • 1吴文健,孙晖,季国忠.电子病历书写时限存在问题及改进[J].中国病案,2020,0(1):9-11. 被引量:21
  • 2Aw A T,Zhang M,Xiao J,et al.A phrase-based statistical model for SMS text normalization[C]//Proceedings of the COLING/ACL on Main conference poster sessions.Association for Computational Linguistics,2006:33-40.
  • 3Kobus C,Yvon F,Damnati G.Normalizing SMS:are two metaphors better than one?[C]//Proceedings of the 22nd International Conference on Computational Linguistics-Volume 1.Association for Computational Linguistics,2008:441-448.
  • 4Han B,Baldwin T.Lexical normalisation of short text messages:Makn sens a#twitter[C]//Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics:Human Language Technologies-Volume 1.Association for Computational Linguistics,2011:368-378.
  • 5Liu F,Weng F,Jiang X.A broad-coverage normalization system for social media language[C]//Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics:Long Papers-Volume 1.Association for Computational Linguistics,2012:1035-1044.
  • 6Han B,Cook P,Baldwin T.Automatically constructing a normalisation dictionary for microblogs[C]//Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning.Association for Computational Linguistics,2012:421-432.
  • 7Wang P,Ng H T.A beam-search decoder for normalization of social media text with application to machine translation[C]//Proceedings of NAACL-HLT,2013:471-481.
  • 8Su J,Wu H,Wang H,et al.Translation model adaptation for statistical machine translation with monolingual topic information[C]//Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics:Long Papers-Volume 1.Association for Computational Linguistics,2012:459-468.
  • 9Huang E H,Socher R,Manning C D,et al.Improving word representations via global context and multiple word prototypes[C]//Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics:Long Papers-Volume 1.Association for Computational Linguistics,2012:873-882.
  • 10Gruber A,Weiss Y,Rosen-Zvi M.Hidden topic Markov models[C]//International Conference on Artificial Intelligence and Statistics,2007:163-170.

共引文献49

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部