期刊文献+
共找到1篇文章
< 1 >
每页显示 20 50 100
Neural Attentional Relation Extraction with Dual Dependency Trees
1
作者 李冬 雷智磊 +2 位作者 宋宝燕 纪婉婷 寇月 《Journal of Computer Science & Technology》 SCIE EI CSCD 2022年第6期1369-1381,共13页
Relation extraction has been widely used to find semantic relations between entities from plain text.Dependency trees provide deeper semantic information for relation extraction.However,existing dependency tree based ... Relation extraction has been widely used to find semantic relations between entities from plain text.Dependency trees provide deeper semantic information for relation extraction.However,existing dependency tree based models adopt pruning strategies that are too aggressive or conservative,leading to insufficient semantic information or excessive noise in relation extraction models.To overcome this issue,we propose the Neural Attentional Relation Extraction Model with Dual Dependency Trees(called DDT-REM),which takes advantage of both the syntactic dependency tree and the semantic dependency tree to well capture syntactic features and semantic features,respectively.Specifically,we first propose novel representation learning to capture the dependency relations from both syntax and semantics.Second,for the syntactic dependency tree,we propose a local-global attention mechanism to solve semantic deficits.We design an extension of graph convolutional networks(GCNs)to perform relation extraction,which effectively improves the extraction accuracy.We conduct experimental studies based on three real-world datasets.Compared with the traditional methods,our method improves the F 1 scores by 0.3,0.1 and 1.6 on three real-world datasets,respectively. 展开更多
关键词 relation extraction graph convolutional network(GCN) syntactic dependency tree semantic dependency tree
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部