期刊文献+
共找到1篇文章
< 1 >
每页显示 20 50 100
Relation Extraction Based on Prompt Information and Feature Reuse
1
作者 Ping Feng Xin Zhang +2 位作者 Jian Zhao Yingying Wang Biao Huang 《Data Intelligence》 EI 2023年第3期824-840,共17页
To alleviate the problem of under-utilization features of sentence-level relation extraction,which leads to insufficient performance of the pre-trained language model and underutilization of the feature vector,a sente... To alleviate the problem of under-utilization features of sentence-level relation extraction,which leads to insufficient performance of the pre-trained language model and underutilization of the feature vector,a sentence-level relation extraction method based on adding prompt information and feature reuse is proposed.At first,in addition to the pair of nominals and sentence information,a piece of prompt information is added,and the overall feature information consists of sentence information,entity pair information,and prompt information,and then the features are encoded by the pre-trained language model ROBERTA.Moreover,in the pre-trained language model,BIGRU is also introduced in the composition of the neural network to extract information,and the feature information is passed through the neural network to form several sets of feature vectors.After that,these feature vectors are reused in different combinations to form multiple outputs,and the outputs are aggregated using ensemble-learning soft voting to perform relation classification.In addition to this,the sum of cross-entropy,KL divergence,and negative log-likelihood loss is used as the final loss function in this paper.In the comparison experiments,the model based on adding prompt information and feature reuse achieved higher results of the SemEval-2010 task 8 relational dataset. 展开更多
关键词 relation extraction language model prompt information feature reuse loss function
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部