期刊文献+

Relation Classification via Sequence Features and Bi-Directional LSTMs 被引量:6

Relation Classification via Sequence Features and Bi-Directional LSTMs
原文传递
导出
摘要 Structure features need complicated pre-processing, and are probably domain-dependent. To reduce time cost of pre-processing, we propose a novel neural network architecture which is a bi-directional long-short-term-memory recurrent-neural-network(Bi-LSTM-RNN) model based on low-cost sequence features such as words and part-of-speech(POS) tags, to classify the relation of two entities. First, this model performs bi-directional recurrent computation along the tokens of sentences. Then, the sequence is divided into five parts and standard pooling functions are applied over the token representations of each part. Finally, the token representations are concatenated and fed into a softmax layer for relation classification. We evaluate our model on two standard benchmark datasets in different domains, namely Sem Eval-2010 Task 8 and Bio NLP-ST 2016 Task BB3. In Sem Eval-2010 Task 8, the performance of our model matches those of the state-of-the-art models, achieving 83.0% in F1. In Bio NLP-ST 2016 Task BB3, our model obtains F1 51.3% which is comparable with that of the best system. Moreover, we find that the context between two target entities plays an important role in relation classification and it can be a replacement of the shortest dependency path. Structure features need complicated pre-processing, and are probably domain-dependent. To reduce time cost of pre-processing, we propose a novel neural network architecture which is a bi-directional long-short-term-memory recurrent-neural-network(Bi-LSTM-RNN) model based on low-cost sequence features such as words and part-of-speech(POS) tags, to classify the relation of two entities. First, this model performs bi-directional recurrent computation along the tokens of sentences. Then, the sequence is divided into five parts and standard pooling functions are applied over the token representations of each part. Finally, the token representations are concatenated and fed into a softmax layer for relation classification. We evaluate our model on two standard benchmark datasets in different domains, namely Sem Eval-2010 Task 8 and Bio NLP-ST 2016 Task BB3. In Sem Eval-2010 Task 8, the performance of our model matches those of the state-of-the-art models, achieving 83.0% in F1. In Bio NLP-ST 2016 Task BB3, our model obtains F1 51.3% which is comparable with that of the best system. Moreover, we find that the context between two target entities plays an important role in relation classification and it can be a replacement of the shortest dependency path.
机构地区 School of Computer
出处 《Wuhan University Journal of Natural Sciences》 CAS CSCD 2017年第6期489-497,共9页 武汉大学学报(自然科学英文版)
基金 Supported by the China Postdoctoral Science Foundation(2014T70722) the Humanities and Social Science Foundation of Ministry of Education of China(16YJCZH004)
关键词 Bi-LSTM-RNN relation classification sequence features structure features Bi-LSTM-RNN relation classification sequence features structure features
  • 相关文献

参考文献2

二级参考文献33

  • 1Blake C, Keogh E, Merz C. UCI Repository of machine learning databases [ED/OL]. [2011-03-10]. http: // www.ics. uci. edu/_mlearn/MLRepository, html.
  • 2Cheng Weiyuan, Juang Chiafeng. An incremental support vector machine-trained TS-type fuzzy system for online classification problems[J]. Fuzzy Sets and Systems, 2011,163 24-44.
  • 3Yi Yang, Wu Jiansheng, Xu Wei. Incremental SVM based on reserved set for network intrusion detection[J]. Expert Systems with Applications, 2011, 38: 7698-7707.
  • 4Andrzej P, Luo Jie, Barbara C. The more you learn, the less you store: Memory-controlled incremental SVM for visual place recognition[J]. Image and Vision Computing, 2010, 28: 1080-1097.
  • 5Vapnik V. The Nature of Statistical Learning Theory[M]. New York: Springer-Verlag, 1998.
  • 6Burges C. A tutorial on support vector machines for pattern recognition[J]. Data Mining and Knowledge Discovery, 1998, 2(2): 121-167.
  • 7Cristianini N, Shawe-Taylor J. An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods[M]. Cambridge: Cambridge University Press, 2000.
  • 8Malon C, Uchida S, Suzuki M. Mathematical symbol recognition with support vector machines[J]. Pattern Recognition Letters, 2008, 29: 1326-1332.
  • 9Platt J. Fast Training Support Vector Machines Using Sequential Minimal Optimization [C]//Advances in Kernel Methods--Support Vector Learning. Cambridge: MIT Press, 1998.
  • 10Joachims T. Making large-scale support vector machine learning practical[C]//Advances in Kernel Methods--Support Vector Learning. Cambridge: MIT Press, 1998.

共引文献1

同被引文献33

引证文献6

二级引证文献40

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部