期刊文献+

基于BERT_Stacked LSTM的农业病虫害问句分类方法 被引量:4

Question Classification Method of Agricultural Diseases and Pests Based on BERT_Stacked LSTM
下载PDF
导出
摘要 为解决农业病虫害问句分类过程中存在公开数据集较少、文本较短、特征稀疏、隐含语义信息较难学习等问题,以火爆农资招商网为数据源,构建了用于农业病虫害问句分类的数据集,提出了一种用于农业病虫害问句分类的深度学习模型BERT;tacked LSTM。首先,BERT部分获取各个问句的字符级语义信息,生成了包含句子级特征信息的隐藏向量。然后,使用堆叠长短期记忆网络(Stacked LSTM)学习到隐藏的复杂语义信息。实验结果表明,与其他对比模型相比,本文模型对农业病虫害问句分类更具优势,F1值达到了95.76%,并在公开通用领域数据集上进行了测试,F1值达到了98.44%,表明了模型具有较好的的泛化性。 In order to solve the thorny problems in the process of classification of agricultural diseases and insect pests questions,such as fewer public data sets,shorter texts and sparse features,and difficult to learn implicit semantic information,using the hot agricultural investment network as the data source,a data set for the classification of agricultural pests and diseases was constructed,and a deep learning model BERT_Stacked LSTM for the classification of agricultural pests and diseases was proposed.Firstly,the BERT obtained the character-level semantic information of each question,and generated a hidden vector containing sentence-level feature information.Then,stacked long short-term memory network(Stacked LSTM)structure was used to learn the hidden complex semantic information.Experimental results showed the effectiveness of the proposed model.Compared with other comparative models,the model proposed had more advantages in classifying agricultural diseases and insect pests questions.The F1 score reached 95.76%,and it was widely used in public.Tested on the domain data set,the F1 score reached 98.44%,indicating that the generalization of the model was also very good.
作者 李林 刁磊 唐詹 柏召 周晗 郭旭超 LI Lin;DIAO Lei;TANG Zhan;BAI Zhao;ZHOU Han;GUO Xuchao(College of Information and Electrical Engineering,China Agricultural University,Beijing 100083,China)
出处 《农业机械学报》 EI CAS CSCD 北大核心 2021年第S01期172-177,共6页 Transactions of the Chinese Society for Agricultural Machinery
基金 国家重点研发计划项目(2016YFD0300710)
关键词 农业病虫害 问句分类 BERT Stacked LSTM agricultural diseases and pests question classification BERT Stacked LSTM
  • 相关文献

参考文献7

二级参考文献46

  • 1王建会,王洪伟,申展,胡运发.一种实用高效的文本分类算法[J].计算机研究与发展,2005,42(1):85-93. 被引量:20
  • 2李荣陆,王建会,陈晓云,陶晓鹏,胡运发.使用最大熵模型进行中文文本分类[J].计算机研究与发展,2005,42(1):94-101. 被引量:95
  • 3苏金树,张博锋,徐昕.基于机器学习的文本分类技术研究进展[J].软件学报,2006,17(9):1848-1859. 被引量:378
  • 4Sven Martin,Hermann Ney,Jrg Zaplo. Smoothing methods in maximum entropy language modeling[C].In:IEEE International Conference on Acoustics,Speech and Signal Processing, Phoenix,AR, 1999: 545~548
  • 5Adwait Ratnaparkhi.A simple introduction to maximum entropy models for natural language processing[R].Technical Report 97-08 ,Institute for Research in Cognitive Science,University of Pennsylvania,1997
  • 6Yang Y.An evaluation of statistical approaches to text categorization[J].Information Retrieval, 1999; 1 ( 1 ) :76~88
  • 7Adwait R.Maximum entropy models for natural language ambiguity resolution[D].PhD thesis. University of Pennsylvania, 1998
  • 8Adwait Ratnaparkhi.A maximum entropy model for Part-of-Speech tagging[C].In:Proceedings of the Empirical Methods in Natural Language Processing Conference, 1996
  • 9Adam L Berger,Stephen A Della Pietra,Vincent J Della Pietra. A maximum entropy approach to natural language processing[J].Computational Linguistics,1996;22(1) :38~73
  • 10Kamal Nigam,John Lafferty,Andrew McCallum. Using maximum entropy for text classification[C].In:IJCAI-99 Workshop on Machine Learning for Information Filtering, 1999

共引文献436

同被引文献54

引证文献4

二级引证文献8

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部