期刊文献+
共找到3篇文章
< 1 >
每页显示 20 50 100
LKMT:Linguistics Knowledge-Driven Multi-Task Neural Machine Translation for Urdu and English
1
作者 Muhammad Naeem Ul Hassan Zhengtao Yu +4 位作者 Jian Wang Ying Li shengxiang gao Shuwan Yang Cunli Mao 《Computers, Materials & Continua》 SCIE EI 2024年第10期951-969,共19页
Thanks to the strong representation capability of pre-trained language models,supervised machine translation models have achieved outstanding performance.However,the performances of these models drop sharply when the ... Thanks to the strong representation capability of pre-trained language models,supervised machine translation models have achieved outstanding performance.However,the performances of these models drop sharply when the scale of the parallel training corpus is limited.Considering the pre-trained language model has a strong ability for monolingual representation,it is the key challenge for machine translation to construct the in-depth relationship between the source and target language by injecting the lexical and syntactic information into pre-trained language models.To alleviate the dependence on the parallel corpus,we propose a Linguistics Knowledge-Driven MultiTask(LKMT)approach to inject part-of-speech and syntactic knowledge into pre-trained models,thus enhancing the machine translation performance.On the one hand,we integrate part-of-speech and dependency labels into the embedding layer and exploit large-scale monolingual corpus to update all parameters of pre-trained language models,thus ensuring the updated language model contains potential lexical and syntactic information.On the other hand,we leverage an extra self-attention layer to explicitly inject linguistic knowledge into the pre-trained language model-enhanced machine translation model.Experiments on the benchmark dataset show that our proposed LKMT approach improves the Urdu-English translation accuracy by 1.97 points and the English-Urdu translation accuracy by 2.42 points,highlighting the effectiveness of our LKMT framework.Detailed ablation experiments confirm the positive impact of part-of-speech and dependency parsing on machine translation. 展开更多
关键词 Urdu NMT(neural machine translation) Urdu natural language processing Urdu Linguistic features low resources language linguistic features pretrain model
下载PDF
Review Expert Collaborative Recommendation Algorithm Based on Topic Relationship 被引量:4
2
作者 shengxiang gao Zhengtao Yu +2 位作者 Linbin Shi Xin Yan Haixia Song 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI 2015年第4期403-411,共9页
The project review information plays an important role in the recommendation of review experts. In this paper, we aim to determine review expert's rating by using the historical rating records and the final decisi... The project review information plays an important role in the recommendation of review experts. In this paper, we aim to determine review expert's rating by using the historical rating records and the final decision results on the previous projects, and by means of some rules, we construct a rating matrix for projects and experts. For the data sparseness problem of the rating matrix and the 'cold start' problem of new expert recommendation, we assume that those projects/experts with similar topics have similar feature vectors and propose a review expert collaborative recommendation algorithm based on topic relationship. Firstly, we obtain topics of projects/experts based on latent Dirichlet allocation (LDA) model, and build the topic relationship network of projects/experts. Then, through the topic relationship between projects/experts, we find a neighbor collection which shares the largest similarity with target project/expert, and integrate the collection into the collaborative filtering recommendation algorithm based on matrix factorization. Finally, by learning the rating matrix to get feature vectors of the projects and experts, we can predict the ratings that a target project will give candidate review experts, and thus achieve the review expert recommendation. Experiments on real data set show that the proposed method could predict the review expert rating more effectively, and improve the recommendation effect of review experts. © 2014 Chinese Association of Automation. 展开更多
关键词 Algorithms Collaborative filtering FACTORIZATION RATING STATISTICS
下载PDF
Semantic-aware entity alignment for low resource language knowledge graph
3
作者 Junfei TANG Ran SONG +2 位作者 Yuxin HUANG shengxiang gao Zhengtao YU 《Frontiers of Computer Science》 SCIE EI CSCD 2024年第4期97-106,共10页
Entity alignment(EA)is an important technique aiming to find the same real entity between two different source knowledge graphs(KGs).Current methods typically learn the embedding of entities for EA from the structure ... Entity alignment(EA)is an important technique aiming to find the same real entity between two different source knowledge graphs(KGs).Current methods typically learn the embedding of entities for EA from the structure of KGs for EA.Most EA models are designed for rich-resource languages,requiring sufficient resources such as a parallel corpus and pre-trained language models.However,low-resource language KGs have received less attention,and current models demonstrate poor performance on those low-resource KGs.Recently,researchers have fused relation information and attributes for entity representations to enhance the entity alignment performance,but the relation semantics are often ignored.To address these issues,we propose a novel Semantic-aware Graph Neural Network(SGNN)for entity alignment.First,we generate pseudo sentences according to the relation triples and produce representations using pre-trained models.Second,our approach explores semantic information from the connected relations by a graph neural network.Our model captures expanded feature information from KGs.Experimental results using three low-resource languages demonstrate that our proposed SGNN approach out performs better than state-of-the-art alignment methods on three proposed datasets and three public datasets. 展开更多
关键词 graph neural network knowledge graph entity alignment low-resource language
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部