期刊文献+
共找到6篇文章
< 1 >
每页显示 20 50 100
Improving Chinese Word Representation with Conceptual Semantics 被引量:1
1
作者 Tingxin Wei Weiguang Qu +3 位作者 Junsheng Zhou Yunfei Long Yanhui Gu Zhentao Xia 《Computers, Materials & Continua》 SCIE EI 2020年第9期1897-1913,共17页
The meaning of a word includes a conceptual meaning and a distributive meaning.Word embedding based on distribution suffers from insufficient conceptual semantic representation caused by data sparsity,especially for l... The meaning of a word includes a conceptual meaning and a distributive meaning.Word embedding based on distribution suffers from insufficient conceptual semantic representation caused by data sparsity,especially for low-frequency words.In knowledge bases,manually annotated semantic knowledge is stable and the essential attributes of words are accurately denoted.In this paper,we propose a Conceptual Semantics Enhanced Word Representation(CEWR)model,computing the synset embedding and hypernym embedding of Chinese words based on the Tongyici Cilin thesaurus,and aggregating it with distributed word representation to have both distributed information and the conceptual meaning encoded in the representation of words.We evaluate the CEWR model on two tasks:word similarity computation and short text classification.The Spearman correlation between model results and human judgement are improved to 64.71%,81.84%,and 85.16%on Wordsim297,MC30,and RG65,respectively.Moreover,CEWR improves the F1 score by 3%in the short text classification task.The experimental results show that CEWR can represent words in a more informative approach than distributed word embedding.This proves that conceptual semantics,especially hypernymous information,is a good complement to distributed word representation. 展开更多
关键词 Word representation conceptual semantics hypernymy similarity computation short text classification
下载PDF
Word-Representation-Based Method for Extracting Organizational Events from Online Media 被引量:1
2
作者 Jun-Qiang Zhang Xiong-Wen Deng Yu Qian 《Journal of Electronic Science and Technology》 CAS CSCD 2017年第4期407-412,共6页
Online social media exhibit massive organizational event relevant messages, and the well categorized event information can be useful in many real-world applications. In this paper, we propose a research framework to e... Online social media exhibit massive organizational event relevant messages, and the well categorized event information can be useful in many real-world applications. In this paper, we propose a research framework to extract high quality event information from massive online media data. The main contributions lie in two aspects: First, we present an event-extraction and event-categorization system for online media data; second, we present a novel approach for both discovering important event categories and classifying extracted events based on word representation and clustering model. Experimental results with real dataset show that the proposed framework is effective to extract high quality event information. 展开更多
关键词 Event detection social media text mining word representation
下载PDF
Enhancing Embedding-Based Chinese Word Similarity Evaluation with Concepts and Synonyms Knowledge
3
作者 Fulian Yin Yanyan Wang +1 位作者 Jianbo Liu Meiqi Ji 《Computer Modeling in Engineering & Sciences》 SCIE EI 2020年第8期747-764,共18页
Word similarity(WS)is a fundamental and critical task in natural language processing.Existing approaches to WS are mainly to calculate the similarity or relatedness of word pairs based on word embedding obtained by ma... Word similarity(WS)is a fundamental and critical task in natural language processing.Existing approaches to WS are mainly to calculate the similarity or relatedness of word pairs based on word embedding obtained by massive and high-quality corpus.However,it may suffer from poor performance for insufficient corpus in some specific fields,and cannot capture rich semantic and sentimental information.To address these above problems,we propose an enhancing embedding-based word similarity evaluation with character-word concepts and synonyms knowledge,namely EWS-CS model,which can provide extra semantic information to enhance word similarity evaluation.The core of our approach contains knowledge encoder and word encoder.In knowledge encoder,we incorporate the semantic knowledge extracted from knowledge resources,including character-word concepts,synonyms and sentiment lexicons,to obtain knowledge representation.Word encoder is to learn enhancing embedding-based word representation from pre-trained model and knowledge representation based on similarity task.Finally,compared with baseline models,the experiments on four similarity evaluation datasets validate the effectiveness of our EWS-CS model in WS task. 展开更多
关键词 Word representation concepts and synonyms knowledge word similarity information security
下载PDF
Application of Word Embedding to Drug Repositioning
4
作者 Duc Luu Ngo Naoki Yamamoto +5 位作者 Vu Anh Tran Ngoc Giang Nguyen Dau Phan Favorisen Rosyking Lumbanraja Mamoru Kubo Kenji Satou 《Journal of Biomedical Science and Engineering》 2016年第1期7-16,共10页
As a key technology of rapid and low-cost drug development, drug repositioning is getting popular. In this study, a text mining approach to the discovery of unknown drug-disease relation was tested. Using a word embed... As a key technology of rapid and low-cost drug development, drug repositioning is getting popular. In this study, a text mining approach to the discovery of unknown drug-disease relation was tested. Using a word embedding algorithm, senses of over 1.7 million words were well represented in sufficiently short feature vectors. Through various analysis including clustering and classification, feasibility of our approach was tested. Finally, our trained classification model achieved 87.6% accuracy in the prediction of drug-disease relation in cancer treatment and succeeded in discovering novel drug-disease relations that were actually reported in recent studies. 展开更多
关键词 Distributed representation of Word Sense Discovery of Drug-Disease Relation Word Analogy
下载PDF
Syntactic word embedding based on dependency syntax and polysemous analysis
5
作者 Zhong-lin YE Hai-xing ZHAO 《Frontiers of Information Technology & Electronic Engineering》 SCIE EI CSCD 2018年第4期524-535,共12页
Most word embedding models have the following problems:(1)In the models based on bag-of-words contexts,the structural relations of sentences are completely neglected;(2)Each word uses a single embedding,which makes th... Most word embedding models have the following problems:(1)In the models based on bag-of-words contexts,the structural relations of sentences are completely neglected;(2)Each word uses a single embedding,which makes the model indiscriminative for polysemous words;(3)Word embedding easily tends to contextual structure similarity of sentences.To solve these problems,we propose an easy-to-use representation algorithm of syntactic word embedding(SWE).The main procedures are:(1)A polysemous tagging algorithm is used for polysemous representation by the latent Dirichlet allocation(LDA)algorithm;(2)Symbols‘+’and‘-’are adopted to indicate the directions of the dependency syntax;(3)Stopwords and their dependencies are deleted;(4)Dependency skip is applied to connect indirect dependencies;(5)Dependency-based contexts are inputted to a word2vec model.Experimental results show that our model generates desirable word embedding in similarity evaluation tasks.Besides,semantic and syntactic features can be captured from dependency-based syntactic contexts,exhibiting less topical and more syntactic similarity.We conclude that SWE outperforms single embedding learning models. 展开更多
关键词 Dependency-based context Polysemous word representation representation learning Syntactic word embedding
原文传递
How Do Pronouns Affect Word Embedding
6
作者 Tonglee Chung Bin Xu +2 位作者 Yongbin Liu Juanzi Li Chunping Ouyang 《Tsinghua Science and Technology》 SCIE EI CAS CSCD 2017年第6期586-594,共9页
Word embedding has drawn a lot of attention due to its usefulness in many NLP tasks. So far a handful of neural-network based word embedding algorithms have been proposed without considering the effects of pronouns in... Word embedding has drawn a lot of attention due to its usefulness in many NLP tasks. So far a handful of neural-network based word embedding algorithms have been proposed without considering the effects of pronouns in the training corpus. In this paper, we propose using co-reference resolution to improve the word embedding by extracting better context. We evaluate four word embeddings with considerations of co-reference resolution and compare the quality of word embedding on the task of word analogy and word similarity on multiple data sets.Experiments show that by using co-reference resolution, the word embedding performance in the word analogy task can be improved by around 1.88%. We find that the words that are names of countries are affected the most,which is as expected. 展开更多
关键词 word embedding co-reference resolution representation learning
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部