期刊文献+
共找到127篇文章
< 1 2 7 >
每页显示 20 50 100
Chinese Named Entity Recognition with Character-Level BLSTM and Soft Attention Model
1
作者 Jize Yin Senlin Luo +1 位作者 Zhouting Wu Limin Pan 《Journal of Beijing Institute of Technology》 EI CAS 2020年第1期60-71,共12页
Unlike named entity recognition(NER)for English,the absence of word boundaries reduces the final accuracy for Chinese NER.To avoid accumulated error introduced by word segmentation,a deep model extracting character-le... Unlike named entity recognition(NER)for English,the absence of word boundaries reduces the final accuracy for Chinese NER.To avoid accumulated error introduced by word segmentation,a deep model extracting character-level features is carefully built and becomes a basis for a new Chinese NER method,which is proposed in this paper.This method converts the raw text to a character vector sequence,extracts global text features with a bidirectional long short-term memory and extracts local text features with a soft attention model.A linear chain conditional random field is also used to label all the characters with the help of the global and local text features.Experiments based on the Microsoft Research Asia(MSRA)dataset are designed and implemented.Results show that the proposed method has good performance compared to other methods,which proves that the global and local text features extracted have a positive influence on Chinese NER.For more variety in the test domains,a resume dataset from Sina Finance is also used to prove the effectiveness of the proposed method. 展开更多
关键词 Chinese named entity recognition(ner) character-level BIDIRECTIONAL long SHORT-TERM memory SOFT attention model
下载PDF
Positive unlabeled named entity recognition with multi-granularity linguistic information
2
作者 欧阳小叶 Chen Shudong Wang Rong 《High Technology Letters》 EI CAS 2021年第4期373-380,共8页
The research on named entity recognition for label-few domain is becoming increasingly important.In this paper,a novel algorithm,positive unlabeled named entity recognition(PUNER)with multi-granularity language inform... The research on named entity recognition for label-few domain is becoming increasingly important.In this paper,a novel algorithm,positive unlabeled named entity recognition(PUNER)with multi-granularity language information,is proposed,which combines positive unlabeled(PU)learning and deep learning to obtain the multi-granularity language information from a few labeled in-stances and many unlabeled instances to recognize named entities.First,PUNER selects reliable negative instances from unlabeled datasets,uses positive instances and a corresponding number of negative instances to train the PU learning classifier,and iterates continuously to label all unlabeled instances.Second,a neural network-based architecture to implement the PU learning classifier is used,and comprehensive text semantics through multi-granular language information are obtained,which helps the classifier correctly recognize named entities.Performance tests of the PUNER are carried out on three multilingual NER datasets,which are CoNLL2003,CoNLL 2002 and SIGHAN Bakeoff 2006.Experimental results demonstrate the effectiveness of the proposed PUNER. 展开更多
关键词 named entity recognition(ner) deep learning neural network positive-unla-beled learning label-few domain multi-granularity(PU)
下载PDF
Low Resource Chinese Geological Text Named Entity Recognition Based on Prompt Learning
3
作者 Hang He Chao Ma +6 位作者 Shan Ye Wenqiang Tang Yuxuan Zhou Zhen Yu Jiaxin Yi Li Hou Mingcai Hou 《Journal of Earth Science》 SCIE CAS CSCD 2024年第3期1035-1043,共9页
Geological reports are a significant accomplishment for geologists involved in geological investigations and scientific research as they contain rich data and textual information.With the rapid development of science ... Geological reports are a significant accomplishment for geologists involved in geological investigations and scientific research as they contain rich data and textual information.With the rapid development of science and technology,a large number of textual reports have accumulated in the field of geology.However,many non-hot topics and non-English speaking regions are neglected in mainstream geoscience databases for geological information mining,making it more challenging for some researchers to extract necessary information from these texts.Natural Language Processing(NLP)has obvious advantages in processing large amounts of textual data.The objective of this paper is to identify geological named entities from Chinese geological texts using NLP techniques.We propose the RoBERTa-Prompt-Tuning-NER method,which leverages the concept of Prompt Learning and requires only a small amount of annotated data to train superior models for recognizing geological named entities in low-resource dataset configurations.The RoBERTa layer captures context-based information and longer-distance dependencies through dynamic word vectors.Finally,we conducted experiments on the constructed Geological Named Entity Recognition(GNER)dataset.Our experimental results show that the proposed model achieves the highest F1 score of 80.64%among the four baseline algorithms,demonstrating the reliability and robustness of using the model for Named Entity Recognition of geological texts. 展开更多
关键词 Prompt Learning named entity recognition(ner) low resource geological text text information mining big data geology.
原文传递
Decoupled Two-Phase Framework for Class-Incremental Few-Shot Named Entity Recognition 被引量:1
4
作者 Yifan Chen Zhen Huang +4 位作者 Minghao Hu Dongsheng Li Changjian Wang Feng Liu Xicheng Lu 《Tsinghua Science and Technology》 SCIE EI CAS CSCD 2023年第5期976-987,共12页
Class-Incremental Few-Shot Named Entity Recognition(CIFNER)aims to identify entity categories that have appeared with only a few newly added(novel)class examples.However,existing class-incremental methods typically in... Class-Incremental Few-Shot Named Entity Recognition(CIFNER)aims to identify entity categories that have appeared with only a few newly added(novel)class examples.However,existing class-incremental methods typically introduce new parameters to adapt to new classes and treat all information equally,resulting in poor generalization.Meanwhile,few-shot methods necessitate samples for all observed classes,making them difficult to transfer into a class-incremental setting.Thus,a decoupled two-phase framework method for the CIFNER task is proposed to address the above issues.The whole task is converted to two separate tasks named Entity Span Detection(ESD)and Entity Class Discrimination(ECD)that leverage parameter-cloning and label-fusion to learn different levels of knowledge separately,such as class-generic knowledge and class-specific knowledge.Moreover,different variants,such as the Conditional Random Field-based(CRF-based),word-pair-based methods in ESD module,and add-based,Natural Language Inference-based(NLI-based)and prompt-based methods in ECD module,are investigated to demonstrate the generalizability of the decoupled framework.Extensive experiments on the three Named Entity Recognition(NER)datasets reveal that our method achieves the state-of-the-art performance in the CIFNER setting. 展开更多
关键词 named entity recognition deep learning class-incremental learning few-shot learning
原文传递
Few-shot Named Entity Recognition with Joint Token and Sentence Awareness 被引量:1
5
作者 Wen Wen Yongbin Liu +1 位作者 Qiang Lin Chunping Ouyang 《Data Intelligence》 EI 2023年第3期767-785,共19页
Few-shot learning has been proposed and rapidly emerging as a viable means for completing various tasks.Recently,few-shot models have been used for Named Entity Recognition(NER).Prototypical network shows high efficie... Few-shot learning has been proposed and rapidly emerging as a viable means for completing various tasks.Recently,few-shot models have been used for Named Entity Recognition(NER).Prototypical network shows high efficiency on few-shot NER.However,existing prototypical methods only consider the similarity of tokens in query sets and support sets and ignore the semantic similarity among the sentences which contain these entities.We present a novel model,Few-shot Named Entity Recognition with Joint Token and Sentence Awareness(JTSA),to address the issue.The sentence awareness is introduced to probe the semantic similarity among the sentences.The Token awareness is used to explore the similarity of the tokens.To further improve the robustness and results of the model,we adopt the joint learning scheme on the few-shot NER.Experimental results demonstrate that our model outperforms state-of-the-art models on two standard Fewshot NER datasets. 展开更多
关键词 few-shot Learning named entity recognition Prototypical Network
原文传递
Pointer-prototype fusion network for few-shot named entity recognition
6
作者 Zhao Haiying Guo Xuan 《The Journal of China Universities of Posts and Telecommunications》 EI CSCD 2023年第5期32-41,共10页
Few-shot named entity recognition(NER)aims to identify named entities in new domains using a limited amount of annotated data.Previous methods divided this task into entity span detection and entity classification,ach... Few-shot named entity recognition(NER)aims to identify named entities in new domains using a limited amount of annotated data.Previous methods divided this task into entity span detection and entity classification,achieving good results.However these methods are limited by the imbalance between the entity and non-entity categories due to the use of sequence labeling for entity span detection.To this end,a point-proto network(PPN)combining pointer and prototypical networks was proposed.Specifically,the pointer network generates the position of entities in sentences in the entity span detection stage.The prototypical network builds semantic prototypes of entity types and classifies entities based on their distance from these prototypes in the entity classification stage.Moreover,the low-rank adaptation(LoRA)fine-tuning method,which involves freezing the pre-trained weights and injecting a trainable decomposition matrix,reduces the parameters that need to be trained and saved.Extensive experiments on the few-shot NER Dataset(Few-NERD)and Cross-Dataset demonstrate the superiority of PPN in this domain. 展开更多
关键词 few-shot named entity recognition(ner) pointer network prototypical network low-rank adaptation
原文传递
Leveraging Continuous Prompt for Few-Shot Named Entity Recognition in Electric Power Domain with Meta-Learning
7
作者 Yang Yu Wei He +1 位作者 Yu-meng Kang You-lang Ji 《Data Intelligence》 EI 2023年第2期494-509,共16页
Conventional named entity recognition methods usually assume that the model can be trained with sufficient annotated data to obtain good recognition results.However,in Chinese named entity recognition in the electric ... Conventional named entity recognition methods usually assume that the model can be trained with sufficient annotated data to obtain good recognition results.However,in Chinese named entity recognition in the electric power domain,existing methods still face the challenges of lack of annotated data and new entities of unseen types.To address these challenges,this paper proposes a meta-learning-based continuous cue adjustment method.A generative pre-trained language model is used so that it does not change its own model structure when dealing with new entity types.To guide the pre-trained model to make full use of its own latent knowledge,a vector of learnable parameters is set as a cue to compensate for the lack of training data.In order to further improve the model's few-shot learning capability,a meta-learning strategy is used to train the model.Experimental results show that the proposed approach achieves the best results in a few-shot electric Chinese power named entity recognition dataset compared to several traditional named entity approaches. 展开更多
关键词 named entity recognition Pre-trained model Prompt-Tuning META-LEARNING few-shot
原文传递
End End-to to-End Chinese Entity Recognition Based on BERT BERT-BiLSTM BiLSTM-ATT ATT-CRF 被引量:1
8
作者 LI Daiyi TU Yaofeng +2 位作者 ZHOU Xiangsheng ZHANG Yangming MA Zongmin 《ZTE Communications》 2022年第S01期27-35,共9页
Traditional named entity recognition methods need professional domain knowl-edge and a large amount of human participation to extract features,as well as the Chinese named entity recognition method based on a neural n... Traditional named entity recognition methods need professional domain knowl-edge and a large amount of human participation to extract features,as well as the Chinese named entity recognition method based on a neural network model,which brings the prob-lem that vector representation is too singular in the process of character vector representa-tion.To solve the above problem,we propose a Chinese named entity recognition method based on the BERT-BiLSTM-ATT-CRF model.Firstly,we use the bidirectional encoder representations from transformers(BERT)pre-training language model to obtain the se-mantic vector of the word according to the context information of the word;Secondly,the word vectors trained by BERT are input into the bidirectional long-term and short-term memory network embedded with attention mechanism(BiLSTM-ATT)to capture the most important semantic information in the sentence;Finally,the conditional random field(CRF)is used to learn the dependence between adjacent tags to obtain the global optimal sentence level tag sequence.The experimental results show that the proposed model achieves state-of-the-art performance on both Microsoft Research Asia(MSRA)corpus and people’s daily corpus,with F1 values of 94.77% and 95.97% respectively. 展开更多
关键词 named entity recognition(ner) feature extraction BERT model BiLSTM at-tention mechanism CRF
下载PDF
Chinese Named Entity Recognition Augmented with Lexicon Memory
9
作者 周奕 郑骁庆 黄萱菁 《Journal of Computer Science & Technology》 SCIE EI CSCD 2023年第5期1021-1035,共15页
Inspired by the concept of content-addressable retrieval from cognitive science,we propose a novel fragment-based Chinese named entity recognition(NER)model augmented with a lexicon-based memory in which both characte... Inspired by the concept of content-addressable retrieval from cognitive science,we propose a novel fragment-based Chinese named entity recognition(NER)model augmented with a lexicon-based memory in which both character-level and word-level features are combined to generate better feature representations for possible entity names.Observing that the boundary information of entity names is particularly useful to locate and classify them into pre-defined categories,position-dependent features,such as prefix and suffix,are introduced and taken into account for NER tasks in the form of distributed representations.The lexicon-based memory is built to help generate such position-dependent features and deal with the problem of out-of-vocabulary words.Experimental results show that the proposed model,called LEMON,achieved state-of-the-art performance with an increase in the Fl-score up to 3.2%over the state-of-the-art models on four different widely-used NER datasets. 展开更多
关键词 named entity recognition(ner) lexicon-based memory content-addressable retrieval position-dependent feature neural network
原文传递
Few-shot Learning for Named Entity Recognition Based on BERT and Two-level Model Fusion
10
作者 Yuan Gong Lu Mao Changliang Li 《Data Intelligence》 EI 2021年第4期568-577,共10页
Currently,as a basic task of military document information extraction,Named Entity Recognition(NER)for military documents has received great attention.In 2020,China Conference on Knowledge Graph and Semantic Computing... Currently,as a basic task of military document information extraction,Named Entity Recognition(NER)for military documents has received great attention.In 2020,China Conference on Knowledge Graph and Semantic Computing(CCKS)and System Engineering Research Institute of Academy of Military Sciences(AMS)issued the NER task for test evaluation,which requires the recognition of four types of entities including Test Elements(TE),Performance Indicators(PI),System Components(SC)and Task Scenarios(TS).Due to the particularity and confidentiality of the military field,only 400 items of annotated data are provided by the organizer.In this paper,the task is regarded as a few-shot learning problem for NER,and a method based on BERT and two-level model fusion is proposed.Firstly,the proposed method is based on several basic models fine tuned by BERT on the training data.Then,a two-level fusion strategy applied to the prediction results of multiple basic models is proposed to alleviate the over-fitting problem.Finally,the labeling errors are eliminated by post-processing.This method achieves F1 score of 0.7203 on the test set of the evaluation task. 展开更多
关键词 few-shot learning named entity recognition BERT Two-level model fusion
原文传递
基于字形特征的血管外科命名实体识别
11
作者 张华青 夏张涛 +1 位作者 陆晓庆 童基均 《计算机工程》 CAS CSCD 北大核心 2024年第8期13-21,共9页
电子病历(EMR)作为医疗信息化建设的核心,蕴含着众多有价值的医疗实体,对电子病历进行命名实体识别有助于推进医学研究。为解决血管外科电子病历研究数据匮乏、实体复杂识别困难等问题,基于某三甲医院血管外科的真实临床数据,构建一个... 电子病历(EMR)作为医疗信息化建设的核心,蕴含着众多有价值的医疗实体,对电子病历进行命名实体识别有助于推进医学研究。为解决血管外科电子病历研究数据匮乏、实体复杂识别困难等问题,基于某三甲医院血管外科的真实临床数据,构建一个小规模的专科数据集作为实验数据集,并提出一种基于字形特征的命名实体识别模型。首先,采用掩码校正的来自Transformer的双向编码器表示(MacBERT)生成动态字向量,引入汉字四角码与汉字五笔两个维度的字形信息;然后,将文本表示传入双向门控循环单元(BiGRU)与门控空洞卷积神经网络(DGCNN)进行特征提取,并对输出结果进行拼接;最后,通过多头自注意力机制捕捉序列内部元素间的关系,利用条件随机场(CRF)进行标签解码。实验结果表明,所提模型在自建血管外科数据集上的精确率、召回率、F1值分别为96.45%、97.77%、97.10%,均优于对比模型,具有更好的实体识别性能。 展开更多
关键词 电子病历 血管外科 命名实体识别 特征融合 深度学习
下载PDF
基于知识图谱增强的领域多模态实体识别
12
作者 李华昱 张智康 +1 位作者 闫阳 岳阳 《计算机工程》 CAS CSCD 北大核心 2024年第8期31-39,共9页
针对特定领域中文命名实体识别存在的局限性,提出一种利用学科图谱和图像提高实体识别准确率的模型,旨在利用领域图谱和图像提高计算机学科领域短文本中实体识别的准确率。使用基于BERT-BiLSTMAttention的模型提取文本特征,使用ResNet15... 针对特定领域中文命名实体识别存在的局限性,提出一种利用学科图谱和图像提高实体识别准确率的模型,旨在利用领域图谱和图像提高计算机学科领域短文本中实体识别的准确率。使用基于BERT-BiLSTMAttention的模型提取文本特征,使用ResNet152提取图像特征,并使用分词工具获得句子中的名词实体。通过BERT将名词实体与图谱节点进行特征嵌入,利用余弦相似度查找句子中的分词在学科图谱中最相似的节点,保留到该节点距离为1的邻居节点,生成最佳匹配子图,作为句子的语义补充。使用多层感知机(MLP)将文本、图像和子图3种特征映射到同一空间,并通过独特的门控机制实现文本和图像的细粒度跨模态特征融合。最后,通过交叉注意力机制将多模态特征与子图特征进行融合,输入解码器进行实体标记。在Twitter2015、Twitter2017和自建计算机学科数据集上同基线模型进行实验比较,结果显示,所提方法在领域数据集上的精确率、召回率和F1值分别可达88.56%、87.47%和88.01%,与最优基线模型相比,F1值提高了1.36个百分点,表明利用领域知识图谱能有效提升实体识别效果。 展开更多
关键词 命名实体识别 多模态 领域 知识图谱 跨模态特征融合 注意力机制
下载PDF
多尺寸注意力的命名实体识别方法
13
作者 唐瑞雪 秦永彬 陈艳平 《计算机科学与探索》 CSCD 北大核心 2024年第2期506-515,共10页
命名实体识别(NER)任务的准确性将促进自然语言领域中诸多下游任务的研究。由于文本中存在大量嵌套语义,导致命名实体识别困难,成为自然语言处理中的难点。以往研究提取特征尺度单一,边界信息利用不够充分,忽略了不同尺度下的许多细节信... 命名实体识别(NER)任务的准确性将促进自然语言领域中诸多下游任务的研究。由于文本中存在大量嵌套语义,导致命名实体识别困难,成为自然语言处理中的难点。以往研究提取特征尺度单一,边界信息利用不够充分,忽略了不同尺度下的许多细节信息,从而造成实体识别错误或遗漏的情况。针对上述问题,提出一种多尺度注意力的命名实体识别方法(MSA-NER)。首先,利用BERT模型得到包含上下文信息的表示向量,并通过BiLSTM网络加强文本的上下文表示。其次,将表示向量进行枚举拼接形成跨度信息矩阵,并融合方向信息获得更丰富的交互信息。然后,利用多头注意力构建多个子空间,通过二维卷积在每个子空间下可选地聚合不同尺度的文本信息,在每个注意力层同时进行多尺度的特征融合。最后,将融合的矩阵进行跨度分类以识别命名实体。实验表明,该方法在GENIA和ACE2005英文数据集上F1分别达到81.7%和86.8%,与现有主流模型相比有更好的识别效果。 展开更多
关键词 命名实体识别(ner) 嵌套语义 多尺度注意力 卷积神经网络 子空间
下载PDF
Cybersecurity Named Entity Recognition Using Bidirectional Long Short-Term Memory with Conditional Random Fields 被引量:11
14
作者 Pingchuan Ma Bo Jiang +2 位作者 Zhigang Lu Ning Li Zhengwei Jiang 《Tsinghua Science and Technology》 SCIE EI CAS CSCD 2021年第3期259-265,共7页
Network texts have become important carriers of cybersecurity information on the Internet.These texts include the latest security events such as vulnerability exploitations,attack discoveries,advanced persistent threa... Network texts have become important carriers of cybersecurity information on the Internet.These texts include the latest security events such as vulnerability exploitations,attack discoveries,advanced persistent threats,and so on.Extracting cybersecurity entities from these unstructured texts is a critical and fundamental task in many cybersecurity applications.However,most Named Entity Recognition(NER)models are suitable only for general fields,and there has been little research focusing on cybersecurity entity extraction in the security domain.To this end,in this paper,we propose a novel cybersecurity entity identification model based on Bidirectional Long Short-Term Memory with Conditional Random Fields(Bi-LSTM with CRF)to extract security-related concepts and entities from unstructured text.This model,which we have named XBi LSTM-CRF,consists of a word-embedding layer,a bidirectional LSTM layer,and a CRF layer,and concatenates X input with bidirectional LSTM output.Via extensive experiments on an open-source dataset containing an office security bulletin,security blogs,and the Common Vulnerabilities and Exposures list,we demonstrate that XBi LSTM-CRF achieves better cybersecurity entity extraction than state-of-the-art models. 展开更多
关键词 security blogs Long Short-Term Memory(LSTM) named entity recognition(ner)
原文传递
A review on cyber security named entity recognition 被引量:4
15
作者 Chen GAO Xuan ZHANG +1 位作者 Mengting HAN Hui LIU 《Frontiers of Information Technology & Electronic Engineering》 SCIE EI CSCD 2021年第9期1153-1168,共16页
With the rapid development of Internet technology and the advent of the era of big data,more and more cyber security texts are provided on the Internet.These texts include not only security concepts,incidents,tools,gu... With the rapid development of Internet technology and the advent of the era of big data,more and more cyber security texts are provided on the Internet.These texts include not only security concepts,incidents,tools,guidelines,and policies,but also risk management approaches,best practices,assurances,technologies,and more.Through the integration of large-scale,heterogeneous,unstructured cyber security information,the identification and classification of cyber security entities can help handle cyber security issues.Due to the complexity and diversity of texts in the cyber security domain,it is difficult to identify security entities in the cyber security domain using the traditional named entity recognition(NER)methods.This paper describes various approaches and techniques for NER in this domain,including the rule-based approach,dictionary-based approach,and machine learning based approach,and discusses the problems faced by NER research in this domain,such as conjunction and disjunction,non-standardized naming convention,abbreviation,and massive nesting.Three future directions of NER in cyber security are proposed:(1)application of unsupervised or semi-supervised technology;(2)development of a more comprehensive cyber security ontology;(3)development of a more comprehensive deep learning model. 展开更多
关键词 named entity recognition(ner) Information extraction Cyber security Machine learning Deep learning
原文传递
融合多粒度语言知识与层级信息的中文命名实体识别模型
16
作者 于右任 张仰森 +1 位作者 蒋玉茹 黄改娟 《计算机应用》 CSCD 北大核心 2024年第6期1706-1712,共7页
针对当前大多数命名实体识别(NER)模型只使用字符级信息编码且缺乏对文本层次信息提取的问题,提出一种融合多粒度语言知识与层级信息的中文NER(CNER)模型(CMH)。首先,使用经过多粒度语言知识预训练的模型编码文本,使模型能够同时捕获文... 针对当前大多数命名实体识别(NER)模型只使用字符级信息编码且缺乏对文本层次信息提取的问题,提出一种融合多粒度语言知识与层级信息的中文NER(CNER)模型(CMH)。首先,使用经过多粒度语言知识预训练的模型编码文本,使模型能够同时捕获文本的细粒度和粗粒度语言信息,从而更好地表征语料;其次,使用ON-LSTM(Ordered Neurons Long Short-Term Memory network)模型提取层级信息,利用文本本身的层级结构信息增强编码间的时序关系;最后,在模型的解码端结合文本的分词信息,并将实体识别问题转化为表格填充问题,以更好地解决实体重叠问题并获得更准确的实体识别结果。同时,为解决当前模型在不同领域中的迁移能力较差的问题,提出通用实体识别的理念,通过筛选多领域的通用实体类型,构建一套提升模型在多领域中的泛化能力的通用NER数据集MDNER(Multi-Domain NER dataset)。为验证所提模型的效果,在数据集Resume、Weibo、MSRA上进行实验,与MECT(Multi-metadata Embedding based Cross-Transformer)模型相比,F1值分别提高了0.94、4.95和1.58个百分点。为了验证所提模型在多领域中的实体识别效果,在MDNER上进行实验,F1值达到了95.29%。实验结果表明,多粒度语言知识预训练、文本层级结构信息提取和高效指针解码器对模型的性能提升至关重要。 展开更多
关键词 命名实体识别 自然语言处理 知识图谱构建 高效指针 通用实体
下载PDF
基于预训练模型的医药说明书实体抽取方法研究
17
作者 陈仲永 黄雍圣 +1 位作者 张旻 姜明 《计算机科学与探索》 CSCD 北大核心 2024年第7期1911-1922,共12页
药品说明书医疗实体抽取可为用药信息智能检索及构建医疗知识图谱提供基础数据,具有重要研究意义与应用价值。针对治疗不同种类疾病的药品说明书中的医疗实体存在着较大的差异从而导致模型训练需要标注大量样本的问题,采用“大模型+小... 药品说明书医疗实体抽取可为用药信息智能检索及构建医疗知识图谱提供基础数据,具有重要研究意义与应用价值。针对治疗不同种类疾病的药品说明书中的医疗实体存在着较大的差异从而导致模型训练需要标注大量样本的问题,采用“大模型+小模型”的设计思路,提出了一种基于预训练模型的部分标签命名实体识别模型,先采用通过少量样本微调的预训练语言模型抽取药品说明书中的部分实体,再利用基于Transformer的部分标签模型进一步优化实体提取结果。部分标签模型采用平面格结构对输入文本、已识别出的部分实体及实体标签进行编码,使用Transformer提取特征表示,最后通过条件随机场(CRF)预测实体标签。为了减少训练模型的标注数据,利用标注样本实体掩盖策略,提出一种样本数据增广方法对部分标签模型进行训练。实验验证了“大模型+小模型”在医疗实体抽取的可行性,结果表明精确率(precision,P)、召回率(recall,R)和F1分数分别为85.0%、86.1%、85.6%,比其他学习方法更具优势。 展开更多
关键词 命名实体识别 预训练模型 医疗实体抽取 TRANSFORMER
下载PDF
融合先验知识和字形特征的中文命名实体识别
18
作者 董永峰 白佳明 +1 位作者 王利琴 王旭 《计算机应用》 CSCD 北大核心 2024年第3期702-708,共7页
针对命名实体识别(NER)任务中相关模型通常仅对字符及相关词汇进行建模,未充分利用汉字特有的字形结构信息和实体类型信息的问题,提出一种融合先验知识和字形特征的命名实体识别模型。首先,采用结合高斯注意力机制的Transformer对输入... 针对命名实体识别(NER)任务中相关模型通常仅对字符及相关词汇进行建模,未充分利用汉字特有的字形结构信息和实体类型信息的问题,提出一种融合先验知识和字形特征的命名实体识别模型。首先,采用结合高斯注意力机制的Transformer对输入序列进行编码,并从中文维基百科中获取实体类型的中文释义,采用双向门控循环单元(BiGRU)编码实体类型信息作为先验知识,利用注意力机制将它与字符表示进行组合;其次,采用双向长短时记忆(BiLSTM)网络编码输入序列的远距离依赖关系,通过字形编码表获得繁体的仓颉码和简体的现代五笔码,采用卷积神经网络(CNN)提取字形特征表示,并根据不同权重组合繁体与简体字形特征,利用门控机制将它与经过BiLSTM编码后的字符表示进行组合;最后,使用条件随机场(CRF)解码,得到命名实体标注序列。在偏口语化的数据集Weibo、小型数据集Boson和大型数据集PeopleDaily上的实验结果表明,与基线模型MECT(Multi-metadata Embedding based Cross-Transformer)相比,所提模型的F1值别提高了2.47、1.20和0.98个百分点,验证了模型的有效性。 展开更多
关键词 命名实体识别 注意力机制 卷积神经网络 双向长短时记忆 条件随机场
下载PDF
k-best维特比解耦合知识蒸馏的命名实体识别模型
19
作者 赵红磊 唐焕玲 +2 位作者 张玉 孙雪源 鲁明羽 《计算机科学与探索》 CSCD 北大核心 2024年第3期780-794,共15页
为提升命名实体识别(NER)模型的性能,可采用知识蒸馏方法,但是传统知识蒸馏损失函数因内部存在的耦合关系会导致蒸馏效果较差。为了解除耦合关系,有效提升输出层特征知识蒸馏的效果,提出一种结合k-best维特比解码的解耦合知识蒸馏方法(k... 为提升命名实体识别(NER)模型的性能,可采用知识蒸馏方法,但是传统知识蒸馏损失函数因内部存在的耦合关系会导致蒸馏效果较差。为了解除耦合关系,有效提升输出层特征知识蒸馏的效果,提出一种结合k-best维特比解码的解耦合知识蒸馏方法(kvDKD),该方法利用k-best维特比算法提高计算效率,能够有效提升模型性能。另外,基于深度学习的命名实体识别在数据增强时易引入噪声,因此提出了融合数据筛选和实体再平衡算法的数据增强方法,旨在减少因原数据集引入噪声和增强数据错误标注的问题,提高数据集质量,减少过度拟合。最后在上述方法的基础上,提出了一种新的命名实体识别模型NER-kvDKD。在MSRA、Resume、Weibo、CLUENER和CoNLL-2003数据集上的对比实验结果表明,该方法能够提高模型的泛化能力,同时也有效提高了学生模型性能。 展开更多
关键词 命名实体识别(ner) 知识蒸馏 k-best维特比解码 数据增强
下载PDF
基于标签语义信息感知的少样本命名实体识别方法
20
作者 张越 王长征 +4 位作者 苏雪峰 闫智超 张广军 邵文远 李茹 《北京大学学报(自然科学版)》 EI CAS CSCD 北大核心 2024年第3期413-421,共9页
在少样本命名实体识别方法中,目前广泛应用的方法是基于原型网络的两阶段模型。但是,该方法未充分利用实体标签中的语义信息,且在距离计算中过度依赖实体类型原型向量,导致模型泛化能力差。针对这些问题,提出一种基于标签语义信息感知... 在少样本命名实体识别方法中,目前广泛应用的方法是基于原型网络的两阶段模型。但是,该方法未充分利用实体标签中的语义信息,且在距离计算中过度依赖实体类型原型向量,导致模型泛化能力差。针对这些问题,提出一种基于标签语义信息感知的少样本命名实体识别方法。该方法是一种先进行实体跨度检测,再判断实体类型的两阶段方法。在构建实体类型原型向量时,将对应实体类型包含的语义信息考虑在内,通过维度转换层将其与原型向量相融合。在对新样本进行实体识别时,将实体类型的正负样本与实体类型原型向量组成实体类型三元组,依据样本到三元组的距离对其进行分类。在多个数据集上的实验结果证明,该模型的性能比以往的模型有较大的提升。 展开更多
关键词 少样本命名实体识别 标签语义信息感知 实体类型三元组 原型网络
下载PDF
上一页 1 2 7 下一页 到第
使用帮助 返回顶部