期刊文献+

局部语义与上下文关系的中文短文本分类算法 被引量:7

Chinese Short Text Classification Algorithm Based on Local Semantics and Context
下载PDF
导出
摘要 短文本通常是由几个到几十个词组成,长度短、特征稀疏,导致短文本分类的准确率难以提升。为了解决此问题,提出了一种基于局部语义特征与上下文关系融合的中文短文本分类算法,称为Bi-LSTM_CNN_AT,该算法利用CNN提取文本的局部语义特征,利用Bi-LSTM提取文本的上下文语义特征,并结合注意力机制,使得Bi-LSTM_CNN_AT模型能从众多的特征中提取出和当前任务最相关的特征,更好地进行文本分类。实验结果表明,Bi-LSTM_CNN_AT模型在NLP&CC2017的新闻标题分类数据集18个类别中的分类准确率为81.31%,比单通道的CNN模型提高2.02%,比单通道的Bi-LSTM模型提高1.77%。 Short text is usually composed of several to dozens of words.Short length and sparse features make it difficult to improve the classification accuracy of short texts.In order to solve this problem,an algorithm of classification for Chinese short texts is proposed based on local semantic features and context relationships,called Bi-LSTM_CNN_AT.In this algorithm,CNN is utilized to extract the local semantic features of a text,while Bi-LSTM is used to extract the contextual semantic features of the text.Moreover,the attention mechanism is combined too.Thus,the Bi-LSTM_CNN_AT model is able to extract the most relevant features to the current task from short texts.The experimental results show that the Bi-LSTM_CNN_AT model achieves a classification accuracy of 81.31%in the 18 categories of NLP&CC2017 news headline classification dataset,which is 2.02%higher than the single-channel CNN model and 1.77%higher than the singlechannel Bi-LSTM model respectively.
作者 黄金杰 蔺江全 何勇军 何瑾洁 王雅君 HUANG Jinjie;LIN Jiangquan;HE Yongjun;HE Jinjie;WANG Yajun(School of Automation,Harbin University of Science and Technology,Harbin 150080,China;School of Computer Science and Technology,Harbin University of Science and Technology,Harbin 150080,China)
出处 《计算机工程与应用》 CSCD 北大核心 2021年第6期94-100,共7页 Computer Engineering and Applications
基金 国家自然科学基金(61305001) 黑龙江省自然科学基金(F201222)。
关键词 短文本分类 卷积神经网络 双向长短时记忆网络 注意力机制 short text classification convolutional neural network bidirectional long short-term memory network attention mechanism
  • 相关文献

参考文献4

二级参考文献17

  • 1L. Rocha, F. Mourao, H. Mota et al., "Temporal contexts: Ef- fective text classification in evolving document collections", In- formation Systems, Vol.38, No.3, pp.388-409, 2012.
  • 2M.T. Fardanesh, "Classification accuracy improvement of neu- ral network classifiers by using unlabeled data", IEEE Trans- actions on Geoscienee and Remote Sensing, Vol.36, No.3, pp.1020 1025, 1998.
  • 3T. Joachims, "Transductive inference for text classification us- ing support vector machines", Proc. of the Sixteenth In- ternational Conference on Machine Learning, Bled, Slovenia, pp.200-209, 1999.
  • 4Y. Tsuruoka, J. Tsujii, "Training a naive bayes classifier via the EM algorithm with a class distribution constraint", Proc. of the Seventh Conference on Natural Language Learning, Edmonton, Canada, pp.127-134, 2003.
  • 5R. Kothari, V. Jain, "Learning from labeled and unlabeled data using a minimal number of queries", IEEE Transaction on Neu- ral Networks, Vol.14, No.6, pp.1496 1505, 2003.
  • 6M. Efron, P. Organisciak, K. Fenlon, "Improving retrieval of short texts through document expansion", Proc. of the 35th International A CM SIGIR Conference on Research and Devel- opment in Information Retrieval, Portland, OR, United states, pp.911-920, 2012.
  • 7V. Vapnik, "The Nature of Statistical Learning Theory, Springer- Verlag, New York, 1999.
  • 8S.M. Katz, "Estimation of probabilities from sparse data for the language model component of a speech recognizer", IEEE Transactions on Acoustics, Speech, and Signal Processing, Vol.ASP-35, No.3, pp.400-401, 1987.
  • 9C.X. Zhai, "Statistical language models for information retrieval a critical review", Foundations and Trends in Information Re- trieval, Vol.2, No.3, pp.137 213, 2008.
  • 10V. Lavrenko, W.B. Croft, "Relevance based language models", Proc. of the 24th annual international A CM SIGIR conference on Research and Development in Information Retrieval, New York, USA, pp.120-127, 2001.

共引文献69

同被引文献73

引证文献7

二级引证文献28

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部