期刊文献+
共找到21篇文章
< 1 2 >
每页显示 20 50 100
TG-SMR:AText Summarization Algorithm Based on Topic and Graph Models 被引量:1
1
作者 Mohamed Ali Rakrouki Nawaf Alharbe +1 位作者 Mashael Khayyat Abeer Aljohani 《Computer Systems Science & Engineering》 SCIE EI 2023年第4期395-408,共14页
Recently,automation is considered vital in most fields since computing methods have a significant role in facilitating work such as automatic text summarization.However,most of the computing methods that are used in r... Recently,automation is considered vital in most fields since computing methods have a significant role in facilitating work such as automatic text summarization.However,most of the computing methods that are used in real systems are based on graph models,which are characterized by their simplicity and stability.Thus,this paper proposes an improved extractive text summarization algorithm based on both topic and graph models.The methodology of this work consists of two stages.First,the well-known TextRank algorithm is analyzed and its shortcomings are investigated.Then,an improved method is proposed with a new computational model of sentence weights.The experimental results were carried out on standard DUC2004 and DUC2006 datasets and compared to four text summarization methods.Finally,through experiments on the DUC2004 and DUC2006 datasets,our proposed improved graph model algorithm TG-SMR(Topic Graph-Summarizer)is compared to other text summarization systems.The experimental results prove that the proposed TG-SMR algorithm achieves higher ROUGE scores.It is foreseen that the TG-SMR algorithm will open a new horizon that concerns the performance of ROUGE evaluation indicators. 展开更多
关键词 Natural language processing text summarization graph model topic model
下载PDF
Graph Ranked Clustering Based Biomedical Text Summarization Using Top k Similarity
2
作者 Supriya Gupta Aakanksha Sharaff Naresh Kumar Nagwani 《Computer Systems Science & Engineering》 SCIE EI 2023年第6期2333-2349,共17页
Text Summarization models facilitate biomedical clinicians and researchers in acquiring informative data from enormous domain-specific literature within less time and effort.Evaluating and selecting the most informati... Text Summarization models facilitate biomedical clinicians and researchers in acquiring informative data from enormous domain-specific literature within less time and effort.Evaluating and selecting the most informative sentences from biomedical articles is always challenging.This study aims to develop a dual-mode biomedical text summarization model to achieve enhanced coverage and information.The research also includes checking the fitment of appropriate graph ranking techniques for improved performance of the summarization model.The input biomedical text is mapped as a graph where meaningful sentences are evaluated as the central node and the critical associations between them.The proposed framework utilizes the top k similarity technique in a combination of UMLS and a sampled probability-based clustering method which aids in unearthing relevant meanings of the biomedical domain-specific word vectors and finding the best possible associations between crucial sentences.The quality of the framework is assessed via different parameters like information retention,coverage,readability,cohesion,and ROUGE scores in clustering and non-clustering modes.The significant benefits of the suggested technique are capturing crucial biomedical information with increased coverage and reasonable memory consumption.The configurable settings of combined parameters reduce execution time,enhance memory utilization,and extract relevant information outperforming other biomedical baseline models.An improvement of 17%is achieved when the proposed model is checked against similar biomedical text summarizers. 展开更多
关键词 Biomedical text summarization UMLS BioBERT SDPMM clustering top K similarity PPF HITS page rank graph ranking
下载PDF
A Dual Attention Encoder-Decoder Text Summarization Model
3
作者 Nada Ali Hakami Hanan Ahmed Hosni Mahmoud 《Computers, Materials & Continua》 SCIE EI 2023年第2期3697-3710,共14页
A worthy text summarization should represent the fundamental content of the document.Recent studies on computerized text summarization tried to present solutions to this challenging problem.Attention models are employ... A worthy text summarization should represent the fundamental content of the document.Recent studies on computerized text summarization tried to present solutions to this challenging problem.Attention models are employed extensively in text summarization process.Classical attention techniques are utilized to acquire the context data in the decoding phase.Nevertheless,without real and efficient feature extraction,the produced summary may diverge from the core topic.In this article,we present an encoder-decoder attention system employing dual attention mechanism.In the dual attention mechanism,the attention algorithm gathers main data from the encoder side.In the dual attentionmodel,the system can capture and producemore rational main content.The merging of the two attention phases produces precise and rational text summaries.The enhanced attention mechanism gives high score to text repetition to increase phrase score.It also captures the relationship between phrases and the title giving them higher score.We assessed our proposed model with or without significance optimization using ablation procedure.Our model with significance optimization achieved the highest performance of 96.7%precision and the least CPU time among other models in both training and sentence extraction. 展开更多
关键词 text summarization attention model phrase significance
下载PDF
Applied Linguistics with Mixed Leader Optimizer Based English Text Summarization Model
4
作者 Hala J.Alshahrani Khaled Tarmissi +5 位作者 Ayman Yafoz Abdullah Mohamed Manar Ahmed Hamza Ishfaq Yaseen Abu Sarwar Zamani Mohammad Mahzari 《Intelligent Automation & Soft Computing》 SCIE 2023年第6期3203-3219,共17页
The term‘executed linguistics’corresponds to an interdisciplinary domain in which the solutions are identified and provided for real-time language-related problems.The exponential generation of text data on the Inte... The term‘executed linguistics’corresponds to an interdisciplinary domain in which the solutions are identified and provided for real-time language-related problems.The exponential generation of text data on the Internet must be leveraged to gain knowledgeable insights.The extraction of meaningful insights from text data is crucial since it can provide value-added solutions for business organizations and end-users.The Automatic Text Summarization(ATS)process reduces the primary size of the text without losing any basic components of the data.The current study introduces an Applied Linguistics-based English Text Summarization using a Mixed Leader-Based Optimizer with Deep Learning(ALTS-MLODL)model.The presented ALTS-MLODL technique aims to summarize the text documents in the English language.To accomplish this objective,the proposed ALTS-MLODL technique pre-processes the input documents and primarily extracts a set of features.Next,the MLO algorithm is used for the effectual selection of the extracted features.For the text summarization process,the Cascaded Recurrent Neural Network(CRNN)model is exploited whereas the Whale Optimization Algorithm(WOA)is used as a hyperparameter optimizer.The exploitation of the MLO-based feature selection and the WOA-based hyper-parameter tuning enhanced the summarization results.To validate the perfor-mance of the ALTS-MLODL technique,numerous simulation analyses were conducted.The experimental results signify the superiority of the proposed ALTS-MLODL technique over other approaches. 展开更多
关键词 text summarization deep learning hyperparameter tuning applied linguistics multi-leader optimizer
下载PDF
Abstractive Arabic Text Summarization Using Hyperparameter Tuned Denoising Deep Neural Network
5
作者 Ibrahim M.Alwayle Hala J.Alshahrani +5 位作者 Saud S.Alotaibi Khaled M.Alalayah Amira Sayed A.Aziz Khadija M.Alaidarous Ibrahim Abdulrab Ahmed Manar Ahmed Hamza 《Intelligent Automation & Soft Computing》 2023年第11期153-168,共16页
ive Arabic Text Summarization using Hyperparameter Tuned Denoising Deep Neural Network(AATS-HTDDNN)technique.The presented AATS-HTDDNN technique aims to generate summaries of Arabic text.In the presented AATS-HTDDNN t... ive Arabic Text Summarization using Hyperparameter Tuned Denoising Deep Neural Network(AATS-HTDDNN)technique.The presented AATS-HTDDNN technique aims to generate summaries of Arabic text.In the presented AATS-HTDDNN technique,the DDNN model is utilized to generate the summary.This study exploits the Chameleon Swarm Optimization(CSO)algorithm to fine-tune the hyperparameters relevant to the DDNN model since it considerably affects the summarization efficiency.This phase shows the novelty of the current study.To validate the enhanced summarization performance of the proposed AATS-HTDDNN model,a comprehensive experimental analysis was conducted.The comparison study outcomes confirmed the better performance of the AATS-HTDDNN model over other approaches. 展开更多
关键词 text summarization deep learning denoising deep neural networks hyperparameter tuning Arabic language
下载PDF
A Method of Integrating Length Constraints into Encoder-Decoder Transformer for Abstractive Text Summarization
6
作者 Ngoc-Khuong Nguyen Dac-Nhuong Le +1 位作者 Viet-Ha Nguyen Anh-Cuong Le 《Intelligent Automation & Soft Computing》 2023年第10期1-18,共18页
Text summarization aims to generate a concise version of the original text.The longer the summary text is,themore detailed it will be fromthe original text,and this depends on the intended use.Therefore,the problem of... Text summarization aims to generate a concise version of the original text.The longer the summary text is,themore detailed it will be fromthe original text,and this depends on the intended use.Therefore,the problem of generating summary texts with desired lengths is a vital task to put the research into practice.To solve this problem,in this paper,we propose a new method to integrate the desired length of the summarized text into the encoder-decoder model for the abstractive text summarization problem.This length parameter is integrated into the encoding phase at each self-attention step and the decoding process by preserving the remaining length for calculating headattention in the generation process and using it as length embeddings added to theword embeddings.We conducted experiments for the proposed model on the two data sets,Cable News Network(CNN)Daily and NEWSROOM,with different desired output lengths.The obtained results show the proposed model’s effectiveness compared with related studies. 展开更多
关键词 Length controllable abstractive text summarization length embedding
下载PDF
Automated Multi-Document Biomedical Text Summarization Using Deep Learning Model 被引量:2
7
作者 Ahmed S.Almasoud Siwar Ben Haj Hassine +5 位作者 Fahd N.Al-Wesabi Mohamed K.Nour Anwer Mustafa Hilal Mesfer Al Duhayyim Manar Ahmed Hamza Abdelwahed Motwakel 《Computers, Materials & Continua》 SCIE EI 2022年第6期5799-5815,共17页
Due to the advanced developments of the Internet and information technologies,a massive quantity of electronic data in the biomedical sector has been exponentially increased.To handle the huge amount of biomedical dat... Due to the advanced developments of the Internet and information technologies,a massive quantity of electronic data in the biomedical sector has been exponentially increased.To handle the huge amount of biomedical data,automated multi-document biomedical text summarization becomes an effective and robust approach of accessing the increased amount of technical and medical literature in the biomedical sector through the summarization of multiple source documents by retaining the significantly informative data.So,multi-document biomedical text summarization acts as a vital role to alleviate the issue of accessing precise and updated information.This paper presents a Deep Learning based Attention Long Short Term Memory(DLALSTM)Model for Multi-document Biomedical Text Summarization.The proposed DL-ALSTM model initially performs data preprocessing to convert the available medical data into a compatible format for further processing.Then,the DL-ALSTM model gets executed to summarize the contents from the multiple biomedical documents.In order to tune the summarization performance of the DL-ALSTM model,chaotic glowworm swarm optimization(CGSO)algorithm is employed.Extensive experimentation analysis is performed to ensure the betterment of the DL-ALSTM model and the results are investigated using the PubMed dataset.Comprehensive comparative result analysis is carried out to showcase the efficiency of the proposed DL-ALSTM model with the recently presented models. 展开更多
关键词 BIOMEDICAL text summarization healthcare deep learning lstm parameter tuning
下载PDF
A Novel Optimized Language-Independent Text Summarization Technique
8
作者 Hanan A.Hosni Mahmoud Alaaeldin M.Hafez 《Computers, Materials & Continua》 SCIE EI 2022年第12期5121-5136,共16页
A substantial amount of textual data is present electronically in several languages.These texts directed the gear to information redundancy.It is essential to remove this redundancy and decrease the reading time of th... A substantial amount of textual data is present electronically in several languages.These texts directed the gear to information redundancy.It is essential to remove this redundancy and decrease the reading time of these data.Therefore,we need a computerized text summarization technique to extract relevant information from group of text documents with correlated subjects.This paper proposes a language-independent extractive summarization technique.The proposed technique presents a clustering-based optimization technique.The clustering technique determines the main subjects of the text,while the proposed optimization technique minimizes redundancy,and maximizes significance.Experiments are devised and evaluated using BillSum dataset for the English language,MLSUM for German and Russian and Mawdoo3 for the Arabic language.The experiments are evaluated using ROUGE metrics.The results showed the effectiveness of the proposed technique compared to other language-dependent and languageindependent summarization techniques.Our technique achieved better ROUGE metrics for all the utilized datasets.The technique accomplished an F-measure of 41.9%for Rouge-1,18.7%for Rouge-2,39.4%for Rouge-3,and 16.8%for Rouge-4 on average for all the dataset using all three objectives.Our system also exhibited an improvement of 26.6%,35.5%,34.65%,and 31.54%w.r.t.The recent model contributed in the summarization of BillSum in terms of ROUGE metric evaluation.Our model’s performance is higher than the comparedmodels,especially in themetric results ofROUGE_2which is bi-gram matching. 展开更多
关键词 text summarization:language-independent summarization ROUGE
下载PDF
A Deep Look into Extractive Text Summarization
9
作者 Jhonathan Quillo-Espino Rosa María Romero-González Ana-Marcela Herrera-Navarro 《Journal of Computer and Communications》 2021年第6期24-37,共14页
This investigation has presented an approach to Extractive Automatic Text Summarization (EATS). A framework focused on the summary of a single document has been developed, using the Tf-ldf method (Frequency Term, Inve... This investigation has presented an approach to Extractive Automatic Text Summarization (EATS). A framework focused on the summary of a single document has been developed, using the Tf-ldf method (Frequency Term, Inverse Document Frequency) as a reference, dividing the document into a subset of documents and generating value of each of the words contained in each document, those documents that show Tf-Idf equal or higher than the threshold are those that represent greater importance, therefore;can be weighted and generate a text summary according to the user’s request. This document represents a derived model of text mining application in today’s world. We demonstrate the way of performing the summarization. Random values were used to check its performance. The experimented results show a satisfactory and understandable summary and summaries were found to be able to run efficiently and quickly, showing which are the most important text sentences according to the threshold selected by the user. 展开更多
关键词 text Mining Preprocesses text summarization Extractive text Sumarization
下载PDF
RETRACTED:Recent Approaches for Text Summarization Using Machine Learning&LSTM0
10
作者 Neeraj Kumar Sirohi Mamta Bansal S.N.Rajan 《Journal on Big Data》 2021年第1期35-47,共13页
Nowadays,data is very rapidly increasing in every domain such as social media,news,education,banking,etc.Most of the data and information is in the form of text.Most of the text contains little invaluable information ... Nowadays,data is very rapidly increasing in every domain such as social media,news,education,banking,etc.Most of the data and information is in the form of text.Most of the text contains little invaluable information and knowledge with lots of unwanted contents.To fetch this valuable information out of the huge text document,we need summarizer which is capable to extract data automatically and at the same time capable to summarize the document,particularly textual text in novel document,without losing its any vital information.The summarization could be in the form of extractive and abstractive summarization.The extractive summarization includes picking sentences of high rank from the text constructed by using sentence and word features and then putting them together to produced summary.An abstractive summarization is based on understanding the key ideas in the given text and then expressing those ideas in pure natural language.The abstractive summarization is the latest problem area for NLP(natural language processing),ML(Machine Learning)and NN(Neural Network)In this paper,the foremost techniques for automatic text summarization processes are defined.The different existing methods have been reviewed.Their effectiveness and limitations are described.Further the novel approach based on Neural Network and LSTM has been discussed.In Machine Learning approach the architecture of the underlying concept is called Encoder-Decoder. 展开更多
关键词 text summarization extractive summary abstractive summary NLP LSTM
下载PDF
A Survey of Text Summarization Approaches Based on Deep Learning 被引量:1
11
作者 Sheng-Luan Hou Xi-Kun Huang +4 位作者 Chao-Qun Fei Shu-Han Zhang Yang-Yang Li Qi-Lin Sun Chuan-Qing Wang 《Journal of Computer Science & Technology》 SCIE EI CSCD 2021年第3期633-663,共31页
Automatic text summarization(ATS)has achieved impressive performance thanks to recent advances in deep learning(DL)and the availability of large-scale corpora.The key points in ATS are to estimate the salience of info... Automatic text summarization(ATS)has achieved impressive performance thanks to recent advances in deep learning(DL)and the availability of large-scale corpora.The key points in ATS are to estimate the salience of information and to generate coherent results.Recently,a variety of DL-based approaches have been developed for better considering these two aspects.However,there is still a lack of comprehensive literature review for DL-based ATS approaches.The aim of this paper is to comprehensively review significant DL-based approaches that have been proposed in the literature with respect to the notion of generic ATS tasks and provide a walk-through of their evolution.We first give an overview of ATS and DL.The comparisons of the datasets are also given,which are commonly used for model training,validation,and evaluation.Then we summarize single-document summarization approaches.After that,an overview of multi-document summarization approaches is given.We further analyze the performance of the popular ATS models on common datasets.Various popular approaches can be employed for different ATS tasks.Finally,we propose potential research directions in this fast-growing field.We hope this exploration can provide new insights into future research of DL-based ATS. 展开更多
关键词 automatic text summarization artificial intelligence deep learning attentional encoder-decoder natural language processing
原文传递
An Efficient Long Short-Term Memory Model for Digital Cross-Language Summarization
12
作者 Y.C.A.Padmanabha Reddy Shyam Sunder Reddy Kasireddy +2 位作者 Nageswara Rao Sirisala Ramu Kuchipudi Purnachand Kollapudi 《Computers, Materials & Continua》 SCIE EI 2023年第3期6389-6409,共21页
The rise of social networking enables the development of multilingual Internet-accessible digital documents in several languages.The digital document needs to be evaluated physically through the Cross-Language Text Su... The rise of social networking enables the development of multilingual Internet-accessible digital documents in several languages.The digital document needs to be evaluated physically through the Cross-Language Text Summarization(CLTS)involved in the disparate and generation of the source documents.Cross-language document processing is involved in the generation of documents from disparate language sources toward targeted documents.The digital documents need to be processed with the contextual semantic data with the decoding scheme.This paper presented a multilingual crosslanguage processing of the documents with the abstractive and summarising of the documents.The proposed model is represented as the Hidden Markov Model LSTM Reinforcement Learning(HMMlstmRL).First,the developed model uses the Hidden Markov model for the computation of keywords in the cross-language words for the clustering.In the second stage,bi-directional long-short-term memory networks are used for key word extraction in the cross-language process.Finally,the proposed HMMlstmRL uses the voting concept in reinforcement learning for the identification and extraction of the keywords.The performance of the proposed HMMlstmRL is 2%better than that of the conventional bi-direction LSTM model. 展开更多
关键词 text summarization reinforcement learning hidden markov model CROSS-LANGUAGE MULTILINGUAL
下载PDF
Topic-Aware Abstractive Summarization Based on Heterogeneous Graph Attention Networks for Chinese Complaint Reports
13
作者 Yan Li Xiaoguang Zhang +4 位作者 Tianyu Gong Qi Dong Hailong Zhu Tianqiang Zhang Yanji Jiang 《Computers, Materials & Continua》 SCIE EI 2023年第9期3691-3705,共15页
Automatic text summarization(ATS)plays a significant role in Natural Language Processing(NLP).Abstractive summarization produces summaries by identifying and compressing the most important information in a document.Ho... Automatic text summarization(ATS)plays a significant role in Natural Language Processing(NLP).Abstractive summarization produces summaries by identifying and compressing the most important information in a document.However,there are only relatively several comprehensively evaluated abstractive summarization models that work well for specific types of reports due to their unstructured and oral language text characteristics.In particular,Chinese complaint reports,generated by urban complainers and collected by government employees,describe existing resident problems in daily life.Meanwhile,the reflected problems are required to respond speedily.Therefore,automatic summarization tasks for these reports have been developed.However,similar to traditional summarization models,the generated summaries still exist problems of informativeness and conciseness.To address these issues and generate suitably informative and less redundant summaries,a topic-based abstractive summarization method is proposed to obtain global and local features.Additionally,a heterogeneous graph of the original document is constructed using word-level and topic-level features.Experiments and analyses on public review datasets(Yelp and Amazon)and our constructed dataset(Chinese complaint reports)show that the proposed framework effectively improves the performance of the abstractive summarization model for Chinese complaint reports. 展开更多
关键词 text summarization TOPIC Chinese complaint report heterogeneous graph attention network
下载PDF
SHEL:a semantically enhanced hardware-friendly entity linking method
14
作者 亓东林 CHEN Shudong +2 位作者 DU Rong TONG Da YU Yong 《High Technology Letters》 EI CAS 2024年第1期13-22,共10页
With the help of pre-trained language models,the accuracy of the entity linking task has made great strides in recent years.However,most models with excellent performance require fine-tuning on a large amount of train... With the help of pre-trained language models,the accuracy of the entity linking task has made great strides in recent years.However,most models with excellent performance require fine-tuning on a large amount of training data using large pre-trained language models,which is a hardware threshold to accomplish this task.Some researchers have achieved competitive results with less training data through ingenious methods,such as utilizing information provided by the named entity recognition model.This paper presents a novel semantic-enhancement-based entity linking approach,named semantically enhanced hardware-friendly entity linking(SHEL),which is designed to be hardware friendly and efficient while maintaining good performance.Specifically,SHEL's semantic enhancement approach consists of three aspects:(1)semantic compression of entity descriptions using a text summarization model;(2)maximizing the capture of mention contexts using asymmetric heuristics;(3)calculating a fixed size mention representation through pooling operations.These series of semantic enhancement methods effectively improve the model's ability to capture semantic information while taking into account the hardware constraints,and significantly improve the model's convergence speed by more than 50%compared with the strong baseline model proposed in this paper.In terms of performance,SHEL is comparable to the previous method,with superior performance on six well-established datasets,even though SHEL is trained using a smaller pre-trained language model as the encoder. 展开更多
关键词 entity linking(EL) pre-trained models knowledge graph text summarization semantic enhancement
下载PDF
A Semantic Supervision Method for Abstractive Summarization 被引量:1
15
作者 Sunqiang Hu Xiaoyu Li +3 位作者 Yu Deng Yu Peng Bin Lin Shan Yang 《Computers, Materials & Continua》 SCIE EI 2021年第10期145-158,共14页
In recent years,many text summarization models based on pretraining methods have achieved very good results.However,in these text summarization models,semantic deviations are easy to occur between the original input r... In recent years,many text summarization models based on pretraining methods have achieved very good results.However,in these text summarization models,semantic deviations are easy to occur between the original input representation and the representation that passed multi-layer encoder,which may result in inconsistencies between the generated summary and the source text content.The Bidirectional Encoder Representations from Transformers(BERT)improves the performance of many tasks in Natural Language Processing(NLP).Although BERT has a strong capability to encode context,it lacks the fine-grained semantic representation.To solve these two problems,we proposed a semantic supervision method based on Capsule Network.Firstly,we extracted the fine-grained semantic representation of the input and encoded result in BERT by Capsule Network.Secondly,we used the fine-grained semantic representation of the input to supervise the fine-grained semantic representation of the encoded result.Then we evaluated our model on a popular Chinese social media dataset(LCSTS),and the result showed that our model achieved higher ROUGE scores(including R-1,R-2),and our model outperformed baseline systems.Finally,we conducted a comparative study on the stability of the model,and the experimental results showed that our model was more stable. 展开更多
关键词 text summarization semantic supervision capsule network
下载PDF
An Improved Method for Extractive Based Opinion Summarization Using Opinion Mining
16
作者 Surbhi Bhatia Mohammed AlOjail 《Computer Systems Science & Engineering》 SCIE EI 2022年第8期779-794,共16页
Opinion summarization recapitulates the opinions about a common topic automatically.The primary motive of summarization is to preserve the properties of the text and is shortened in a way with no loss in the semantics... Opinion summarization recapitulates the opinions about a common topic automatically.The primary motive of summarization is to preserve the properties of the text and is shortened in a way with no loss in the semantics of the text.The need of automatic summarization efficiently resulted in increased interest among communities of Natural Language Processing and Text Mining.This paper emphasis on building an extractive summarization system combining the features of principal component analysis for dimensionality reduction and bidirectional Recurrent Neural Networks and Long Short-Term Memory(RNN-LSTM)deep learning model for short and exact synopsis using seq2seq model.It presents a paradigm shift with regard to the way extractive summaries are generated.Novel algorithms for word extraction using assertions are proposed.The semantic framework is well-grounded in this research facilitating the correct decision making process after reviewing huge amount of online reviews,considering all its important features into account.The advantages of the proposed solution provides greater computational efficiency,better inferences from social media,data understanding,robustness and handling sparse data.Experiments on the different datasets also outperforms the previous researches and the accuracy is claimed to achieve more than the baselines,showing the efficiency and the novelty in the research paper.The comparisons are done by calculating accuracy with different baselines using Rouge tool. 展开更多
关键词 Sentiment analysis data mining text summarization opinion mining principal component analysis
下载PDF
Construction of an Automatic Bengali Text Summarizer Using Machine Learning Approaches
17
作者 Busrat Jahan Mahfuja Khatun +2 位作者 Zinat Ara Zabu Afranul Hoque Sayed Uddin Rayhan 《Journal of Data Analysis and Information Processing》 2022年第1期43-57,共15页
In our study, we chose python as the programming platform for finding an Automatic Bengali Document Summarizer. English has sufficient tools to process and receive summarized records. However, there is no specifically... In our study, we chose python as the programming platform for finding an Automatic Bengali Document Summarizer. English has sufficient tools to process and receive summarized records. However, there is no specifically applicable to Bengali since Bengali has a lot of ambiguity, it differs from English in terms of grammar. Afterward, this language holds an important place because this language is spoken by 26 core people all over the world. As a result, it has taken a new method to summarize Bengali documents. The proposed system has been designed by using the following stages: pre-processing the sample doc/input doc, word tagging, pronoun replacement, sentence ranking, as well as summary. Pronoun replacement has been used to reduce the incidence of swinging pronouns in the performance review. We ranked sentences based on sentence frequency, numerical figures, and pronoun replacement. Checking the similarity between two sentences in order to exclude one since it has less duplication. Hereby, we’ve taken 3000 data as input from newspaper and book documents and learned the words to be appropriate with syntax. In addition, to evaluate the performance of the designed summarizer, the design system looked at the different documents. According to the assessment method, the recall, precision, and F-score were 0.70, 0.82 and 0.74, respectively, representing 70%, 82% and 74% recall, precision, and F-score. It has been found that the proper pronoun replacement was 72%. 展开更多
关键词 Natural Language Processing Formatting Bangla text Summarizer Bengali Language Processing Word Tagging Pronoun Replacement Sentence Ranking
下载PDF
Document Clustering Using Graph Based Fuzzy Association Rule Generation
18
作者 P.Perumal 《Computer Systems Science & Engineering》 SCIE EI 2022年第10期203-218,共16页
With the wider growth of web-based documents,the necessity of automatic document clustering and text summarization is increased.Here,document summarization that is extracting the essential task with appropriate inform... With the wider growth of web-based documents,the necessity of automatic document clustering and text summarization is increased.Here,document summarization that is extracting the essential task with appropriate information,removal of unnecessary data and providing the data in a cohesive and coherent manner is determined to be a most confronting task.In this research,a novel intelligent model for document clustering is designed with graph model and Fuzzy based association rule generation(gFAR).Initially,the graph model is used to map the relationship among the data(multi-source)followed by the establishment of document clustering with the generation of association rule using the fuzzy concept.This method shows benefit in redundancy elimination by mapping the relevant document using graph model and reduces the time consumption and improves the accuracy using the association rule generation with fuzzy.This framework is provided in an interpretable way for document clustering.It iteratively reduces the error rate during relationship mapping among the data(clusters)with the assistance of weighted document content.Also,this model represents the significance of data features with class discrimination.It is also helpful in measuring the significance of the features during the data clustering process.The simulation is done with MATLAB 2016b environment and evaluated with the empirical standards like Relative Risk Patterns(RRP),ROUGE score,and Discrimination Information Measure(DMI)respectively.Here,DailyMail and DUC 2004 dataset is used to extract the empirical results.The proposed gFAR model gives better trade-off while compared with various prevailing approaches. 展开更多
关键词 Document clustering text summarization fuzzy model association rule generation graph model relevance mapping feature patterns
下载PDF
Recent advances of neural text generation:Core tasks,datasets,models and challenges 被引量:2
19
作者 JIN HanQi CAO Yue +2 位作者 WANG TianMing XING XinYu WAN XiaoJun 《Science China(Technological Sciences)》 SCIE EI CAS CSCD 2020年第10期1990-2010,共21页
In recent years,deep neural network has achieved great success in solving many natural language processing tasks.Particularly,substantial progress has been made on neural text generation,which takes the linguistic and... In recent years,deep neural network has achieved great success in solving many natural language processing tasks.Particularly,substantial progress has been made on neural text generation,which takes the linguistic and non-linguistic input,and generates natural language text.This survey aims to provide an up-to-date synthesis of core tasks in neural text generation and the architectures adopted to handle these tasks,and draw attention to the challenges in neural text generation.We first outline the mainstream neural text generation frameworks,and then introduce datasets,advanced models and challenges of four core text generation tasks in detail,including AMR-to-text generation,data-to-text generation,and two text-to-text generation tasks(i.e.,text summarization and paraphrase generation).Finally,we present future research directions for neural text generation.This survey can be used as a guide and reference for researchers and practitioners in this area. 展开更多
关键词 natural language generation neural text generation AMR-to-text data-to-text text summarization paraphrase generation
原文传递
The State of the Art of Natural Language Processing-A Systematic Automated Review of NLP Literature Using NLP Techniques
20
作者 Jan Sawicki Maria Ganzha Marcin Paprzycki 《Data Intelligence》 EI 2023年第3期707-749,共43页
Nowadays,natural language processing(NLP)is one of the most popular areas of,broadly understood,artificial intelligence.Therefore,every day,new research contributions are posted,for instance,to the arXiv repository.He... Nowadays,natural language processing(NLP)is one of the most popular areas of,broadly understood,artificial intelligence.Therefore,every day,new research contributions are posted,for instance,to the arXiv repository.Hence,it is rather difficult to capture the current"state of the field"and thus,to enter it.This brought the id-art NLP techniques to analyse the NLP-focused literature.As a result,(1)meta-level knowledge,concerning the current state of NLP has been captured,and(2)a guide to use of basic NLP tools is provided.It should be noted that all the tools and the dataset described in this contribution are publicly available.Furthermore,the originality of this review lies in its full automation.This allows easy reproducibility and continuation and updating of this research in the future as new researches emerge in the field of NLP. 展开更多
关键词 Natural language processing text processing Literature survey Keyword search Keyphrase search text embeddings text summarizations
原文传递
上一页 1 2 下一页 到第
使用帮助 返回顶部