期刊文献+
共找到664篇文章
< 1 2 34 >
每页显示 20 50 100
A large language model-powered literature review for high-angle annular dark field imaging
1
作者 Wenhao Yuan Cheng Peng Qian He 《Chinese Physics B》 SCIE EI CAS CSCD 2024年第9期76-81,共6页
High-angle annular dark field(HAADF)imaging in scanning transmission electron microscopy(STEM)has become an indispensable tool in materials science due to its ability to offer sub-°A resolution and provide chemic... High-angle annular dark field(HAADF)imaging in scanning transmission electron microscopy(STEM)has become an indispensable tool in materials science due to its ability to offer sub-°A resolution and provide chemical information through Z-contrast.This study leverages large language models(LLMs)to conduct a comprehensive bibliometric analysis of a large amount of HAADF-related literature(more than 41000 papers).By using LLMs,specifically ChatGPT,we were able to extract detailed information on applications,sample preparation methods,instruments used,and study conclusions.The findings highlight the capability of LLMs to provide a new perspective into HAADF imaging,underscoring its increasingly important role in materials science.Moreover,the rich information extracted from these publications can be harnessed to develop AI models that enhance the automation and intelligence of electron microscopes. 展开更多
关键词 large language models high-angle ANNULAR DARK FIELD imaging deep learning
下载PDF
Evolution and Prospects of Foundation Models: From Large Language Models to Large Multimodal Models
2
作者 Zheyi Chen Liuchang Xu +5 位作者 Hongting Zheng Luyao Chen Amr Tolba Liang Zhao Keping Yu Hailin Feng 《Computers, Materials & Continua》 SCIE EI 2024年第8期1753-1808,共56页
Since the 1950s,when the Turing Test was introduced,there has been notable progress in machine language intelligence.Language modeling,crucial for AI development,has evolved from statistical to neural models over the ... Since the 1950s,when the Turing Test was introduced,there has been notable progress in machine language intelligence.Language modeling,crucial for AI development,has evolved from statistical to neural models over the last two decades.Recently,transformer-based Pre-trained Language Models(PLM)have excelled in Natural Language Processing(NLP)tasks by leveraging large-scale training corpora.Increasing the scale of these models enhances performance significantly,introducing abilities like context learning that smaller models lack.The advancement in Large Language Models,exemplified by the development of ChatGPT,has made significant impacts both academically and industrially,capturing widespread societal interest.This survey provides an overview of the development and prospects from Large Language Models(LLM)to Large Multimodal Models(LMM).It first discusses the contributions and technological advancements of LLMs in the field of natural language processing,especially in text generation and language understanding.Then,it turns to the discussion of LMMs,which integrates various data modalities such as text,images,and sound,demonstrating advanced capabilities in understanding and generating cross-modal content,paving new pathways for the adaptability and flexibility of AI systems.Finally,the survey highlights the prospects of LMMs in terms of technological development and application potential,while also pointing out challenges in data integration,cross-modal understanding accuracy,providing a comprehensive perspective on the latest developments in this field. 展开更多
关键词 Artificial intelligence large language models large multimodal models foundation models
下载PDF
DeBERTa-GRU: Sentiment Analysis for Large Language Model
3
作者 Adel Assiri Abdu Gumaei +2 位作者 Faisal Mehmood Touqeer Abbas Sami Ullah 《Computers, Materials & Continua》 SCIE EI 2024年第6期4219-4236,共18页
Modern technological advancements have made social media an essential component of daily life.Social media allow individuals to share thoughts,emotions,and ideas.Sentiment analysis plays the function of evaluating whe... Modern technological advancements have made social media an essential component of daily life.Social media allow individuals to share thoughts,emotions,and ideas.Sentiment analysis plays the function of evaluating whether the sentiment of the text is positive,negative,neutral,or any other personal emotion to understand the sentiment context of the text.Sentiment analysis is essential in business and society because it impacts strategic decision-making.Sentiment analysis involves challenges due to lexical variation,an unlabeled dataset,and text distance correlations.The execution time increases due to the sequential processing of the sequence models.However,the calculation times for the Transformer models are reduced because of the parallel processing.This study uses a hybrid deep learning strategy to combine the strengths of the Transformer and Sequence models while ignoring their limitations.In particular,the proposed model integrates the Decoding-enhanced with Bidirectional Encoder Representations from Transformers(BERT)attention(DeBERTa)and the Gated Recurrent Unit(GRU)for sentiment analysis.Using the Decoding-enhanced BERT technique,the words are mapped into a compact,semantic word embedding space,and the Gated Recurrent Unit model can capture the distance contextual semantics correctly.The proposed hybrid model achieves F1-scores of 97%on the Twitter Large Language Model(LLM)dataset,which is much higher than the performance of new techniques. 展开更多
关键词 DeBERTa GRU Naive Bayes LSTM sentiment analysis large language model
下载PDF
Enhancing Relational Triple Extraction in Specific Domains:Semantic Enhancement and Synergy of Large Language Models and Small Pre-Trained Language Models
4
作者 Jiakai Li Jianpeng Hu Geng Zhang 《Computers, Materials & Continua》 SCIE EI 2024年第5期2481-2503,共23页
In the process of constructing domain-specific knowledge graphs,the task of relational triple extraction plays a critical role in transforming unstructured text into structured information.Existing relational triple e... In the process of constructing domain-specific knowledge graphs,the task of relational triple extraction plays a critical role in transforming unstructured text into structured information.Existing relational triple extraction models facemultiple challenges when processing domain-specific data,including insufficient utilization of semantic interaction information between entities and relations,difficulties in handling challenging samples,and the scarcity of domain-specific datasets.To address these issues,our study introduces three innovative components:Relation semantic enhancement,data augmentation,and a voting strategy,all designed to significantly improve the model’s performance in tackling domain-specific relational triple extraction tasks.We first propose an innovative attention interaction module.This method significantly enhances the semantic interaction capabilities between entities and relations by integrating semantic information fromrelation labels.Second,we propose a voting strategy that effectively combines the strengths of large languagemodels(LLMs)and fine-tuned small pre-trained language models(SLMs)to reevaluate challenging samples,thereby improving the model’s adaptability in specific domains.Additionally,we explore the use of LLMs for data augmentation,aiming to generate domain-specific datasets to alleviate the scarcity of domain data.Experiments conducted on three domain-specific datasets demonstrate that our model outperforms existing comparative models in several aspects,with F1 scores exceeding the State of the Art models by 2%,1.6%,and 0.6%,respectively,validating the effectiveness and generalizability of our approach. 展开更多
关键词 Relational triple extraction semantic interaction large language models data augmentation specific domains
下载PDF
LKPNR: Large Language Models and Knowledge Graph for Personalized News Recommendation Framework
5
作者 Hao Chen Runfeng Xie +4 位作者 Xiangyang Cui Zhou Yan Xin Wang Zhanwei Xuan Kai Zhang 《Computers, Materials & Continua》 SCIE EI 2024年第6期4283-4296,共14页
Accurately recommending candidate news to users is a basic challenge of personalized news recommendation systems.Traditional methods are usually difficult to learn and acquire complex semantic information in news text... Accurately recommending candidate news to users is a basic challenge of personalized news recommendation systems.Traditional methods are usually difficult to learn and acquire complex semantic information in news texts,resulting in unsatisfactory recommendation results.Besides,these traditional methods are more friendly to active users with rich historical behaviors.However,they can not effectively solve the long tail problem of inactive users.To address these issues,this research presents a novel general framework that combines Large Language Models(LLM)and Knowledge Graphs(KG)into traditional methods.To learn the contextual information of news text,we use LLMs’powerful text understanding ability to generate news representations with rich semantic information,and then,the generated news representations are used to enhance the news encoding in traditional methods.In addition,multi-hops relationship of news entities is mined and the structural information of news is encoded using KG,thus alleviating the challenge of long-tail distribution.Experimental results demonstrate that compared with various traditional models,on evaluation indicators such as AUC,MRR,nDCG@5 and nDCG@10,the framework significantly improves the recommendation performance.The successful integration of LLM and KG in our framework has established a feasible way for achieving more accurate personalized news recommendation.Our code is available at https://github.com/Xuan-ZW/LKPNR. 展开更多
关键词 large language models news recommendation knowledge graphs(KG)
下载PDF
Evaluating the role of large language models in inflammatory bowel disease patient information
6
作者 Eun Jeong Gong Chang Seok Bang 《World Journal of Gastroenterology》 SCIE CAS 2024年第29期3538-3540,共3页
This letter evaluates the article by Gravina et al on ChatGPT’s potential in providing medical information for inflammatory bowel disease patients.While promising,it highlights the need for advanced techniques like r... This letter evaluates the article by Gravina et al on ChatGPT’s potential in providing medical information for inflammatory bowel disease patients.While promising,it highlights the need for advanced techniques like reasoning+action and retrieval-augmented generation to improve accuracy and reliability.Emphasizing that simple question and answer testing is insufficient,it calls for more nuanced evaluation methods to truly gauge large language models’capabilities in clinical applications. 展开更多
关键词 Crohn’s disease Ulcerative colitis Inflammatory bowel disease Chat generative pre-trained transformer large language model Artificial intelligence
下载PDF
Potential use of large language models for mitigating students’problematic social media use:ChatGPT as an example
7
作者 Xin-Qiao Liu Zi-Ru Zhang 《World Journal of Psychiatry》 SCIE 2024年第3期334-341,共8页
The problematic use of social media has numerous negative impacts on individuals'daily lives,interpersonal relationships,physical and mental health,and more.Currently,there are few methods and tools to alleviate p... The problematic use of social media has numerous negative impacts on individuals'daily lives,interpersonal relationships,physical and mental health,and more.Currently,there are few methods and tools to alleviate problematic social media,and their potential is yet to be fully realized.Emerging large language models(LLMs)are becoming increasingly popular for providing information and assistance to people and are being applied in many aspects of life.In mitigating problematic social media use,LLMs such as ChatGPT can play a positive role by serving as conversational partners and outlets for users,providing personalized information and resources,monitoring and intervening in problematic social media use,and more.In this process,we should recognize both the enormous potential and endless possibilities of LLMs such as ChatGPT,leveraging their advantages to better address problematic social media use,while also acknowledging the limitations and potential pitfalls of ChatGPT technology,such as errors,limitations in issue resolution,privacy and security concerns,and potential overreliance.When we leverage the advantages of LLMs to address issues in social media usage,we must adopt a cautious and ethical approach,being vigilant of the potential adverse effects that LLMs may have in addressing problematic social media use to better harness technology to serve individuals and society. 展开更多
关键词 Problematic use of social media Social media large language models ChatGPT Chatbots
下载PDF
Evaluating Privacy Leakage and Memorization Attacks on Large Language Models (LLMs) in Generative AI Applications
8
作者 Harshvardhan Aditya Siddansh Chawla +6 位作者 Gunika Dhingra Parijat Rai Saumil Sood Tanmay Singh Zeba Mohsin Wase Arshdeep Bahga Vijay K. Madisetti 《Journal of Software Engineering and Applications》 2024年第5期421-447,共27页
The recent interest in the deployment of Generative AI applications that use large language models (LLMs) has brought to the forefront significant privacy concerns, notably the leakage of Personally Identifiable Infor... The recent interest in the deployment of Generative AI applications that use large language models (LLMs) has brought to the forefront significant privacy concerns, notably the leakage of Personally Identifiable Information (PII) and other confidential or protected information that may have been memorized during training, specifically during a fine-tuning or customization process. We describe different black-box attacks from potential adversaries and study their impact on the amount and type of information that may be recovered from commonly used and deployed LLMs. Our research investigates the relationship between PII leakage, memorization, and factors such as model size, architecture, and the nature of attacks employed. The study utilizes two broad categories of attacks: PII leakage-focused attacks (auto-completion and extraction attacks) and memorization-focused attacks (various membership inference attacks). The findings from these investigations are quantified using an array of evaluative metrics, providing a detailed understanding of LLM vulnerabilities and the effectiveness of different attacks. 展开更多
关键词 large language models PII Leakage Privacy Memorization OVERFITTING Membership Inference Attack (MIA)
下载PDF
Security Vulnerability Analyses of Large Language Models (LLMs) through Extension of the Common Vulnerability Scoring System (CVSS) Framework
9
作者 Alicia Biju Vishnupriya Ramesh Vijay K. Madisetti 《Journal of Software Engineering and Applications》 2024年第5期340-358,共19页
Large Language Models (LLMs) have revolutionized Generative Artificial Intelligence (GenAI) tasks, becoming an integral part of various applications in society, including text generation, translation, summarization, a... Large Language Models (LLMs) have revolutionized Generative Artificial Intelligence (GenAI) tasks, becoming an integral part of various applications in society, including text generation, translation, summarization, and more. However, their widespread usage emphasizes the critical need to enhance their security posture to ensure the integrity and reliability of their outputs and minimize harmful effects. Prompt injections and training data poisoning attacks are two of the most prominent vulnerabilities in LLMs, which could potentially lead to unpredictable and undesirable behaviors, such as biased outputs, misinformation propagation, and even malicious content generation. The Common Vulnerability Scoring System (CVSS) framework provides a standardized approach to capturing the principal characteristics of vulnerabilities, facilitating a deeper understanding of their severity within the security and AI communities. By extending the current CVSS framework, we generate scores for these vulnerabilities such that organizations can prioritize mitigation efforts, allocate resources effectively, and implement targeted security measures to defend against potential risks. 展开更多
关键词 Common Vulnerability Scoring System (CVSS) large language models (LLMs) DALL-E Prompt Injections Training Data Poisoning CVSS Metrics
下载PDF
Research status and application of artificial intelligence large models in the oil and gas industry
10
作者 LIU He REN Yili +6 位作者 LI Xin DENG Yue WANG Yongtao CAO Qianwen DU Jinyang LIN Zhiwei WANG Wenjie 《Petroleum Exploration and Development》 SCIE 2024年第4期1049-1065,共17页
This article elucidates the concept of large model technology,summarizes the research status of large model technology both domestically and internationally,provides an overview of the application status of large mode... This article elucidates the concept of large model technology,summarizes the research status of large model technology both domestically and internationally,provides an overview of the application status of large models in vertical industries,outlines the challenges and issues confronted in applying large models in the oil and gas sector,and offers prospects for the application of large models in the oil and gas industry.The existing large models can be briefly divided into three categories:large language models,visual large models,and multimodal large models.The application of large models in the oil and gas industry is still in its infancy.Based on open-source large language models,some oil and gas enterprises have released large language model products using methods like fine-tuning and retrieval augmented generation.Scholars have attempted to develop scenario-specific models for oil and gas operations by using visual/multimodal foundation models.A few researchers have constructed pre-trained foundation models for seismic data processing and interpretation,as well as core analysis.The application of large models in the oil and gas industry faces challenges such as current data quantity and quality being difficult to support the training of large models,high research and development costs,and poor algorithm autonomy and control.The application of large models should be guided by the needs of oil and gas business,taking the application of large models as an opportunity to improve data lifecycle management,enhance data governance capabilities,promote the construction of computing power,strengthen the construction of“artificial intelligence+energy”composite teams,and boost the autonomy and control of large model technology. 展开更多
关键词 foundation model large language mode visual large model multimodal large model large model of oil and gas industry pre-training fine-tuning
下载PDF
Smaller & Smarter: Score-Driven Network Chaining of Smaller Language Models
11
作者 Gunika Dhingra Siddansh Chawla +1 位作者 Vijay K. Madisetti Arshdeep Bahga 《Journal of Software Engineering and Applications》 2024年第1期23-42,共20页
With the continuous evolution and expanding applications of Large Language Models (LLMs), there has been a noticeable surge in the size of the emerging models. It is not solely the growth in model size, primarily meas... With the continuous evolution and expanding applications of Large Language Models (LLMs), there has been a noticeable surge in the size of the emerging models. It is not solely the growth in model size, primarily measured by the number of parameters, but also the subsequent escalation in computational demands, hardware and software prerequisites for training, all culminating in a substantial financial investment as well. In this paper, we present novel techniques like supervision, parallelization, and scoring functions to get better results out of chains of smaller language models, rather than relying solely on scaling up model size. Firstly, we propose an approach to quantify the performance of a Smaller Language Models (SLM) by introducing a corresponding supervisor model that incrementally corrects the encountered errors. Secondly, we propose an approach to utilize two smaller language models (in a network) performing the same task and retrieving the best relevant output from the two, ensuring peak performance for a specific task. Experimental evaluations establish the quantitative accuracy improvements on financial reasoning and arithmetic calculation tasks from utilizing techniques like supervisor models (in a network of model scenario), threshold scoring and parallel processing over a baseline study. 展开更多
关键词 large language models (LLMs) Smaller language models (SLMs) FINANCE NETWORKING Supervisor model Scoring Function
下载PDF
WirelessLLM:Empowering Large Language Models Towards Wireless Intelligence
12
作者 Jiawei Shao Jingwen Tong +4 位作者 Qiong Wu Wei Guo Zijian Li Zehong Lin Jun Zhang 《Journal of Communications and Information Networks》 EI CSCD 2024年第2期99-112,共14页
The rapid evolution of wireless technologies and the growing complexity of network infrastructures necessitate a paradigm shift in how communication networks are designed,configured,and managed. Recent advancements in... The rapid evolution of wireless technologies and the growing complexity of network infrastructures necessitate a paradigm shift in how communication networks are designed,configured,and managed. Recent advancements in large language models (LLMs) have sparked interest in their potential to revolutionize wireless communication systems. However, existing studies on LLMs for wireless systems are limited to a direct application for telecom language understanding. To empower LLMs with knowledge and expertise in the wireless domain, this paper proposes WirelessLLM, a comprehensive framework for adapting and enhancing LLMs to address the unique challenges and requirements of wireless communication networks. We first identify three foundational principles that underpin WirelessLLM:knowledge alignment, knowledge fusion, and knowledge evolution. Then,we investigate the enabling technologies to build WirelessLLM, including prompt engineering, retrieval augmented generation, tool usage, multi-modal pre-training, and domain-specific fine-tuning. Moreover, we present three case studies to demonstrate the practical applicability and benefits of WirelessLLM for solving typical problems in wireless networks. Finally, we conclude this paper by highlighting key challenges and outlining potential avenues for future research. 展开更多
关键词 large language models multi-modal models wireless communications power allocation spectrum sensing protocol understanding
原文传递
FAIR Enough:Develop and Assess a FAIR-Compliant Dataset for Large Language Model Training?
13
作者 Shaina Raza Shardul Ghuge +2 位作者 Chen Ding Elham Dolatabadi Deval Pandya 《Data Intelligence》 EI 2024年第2期559-585,共27页
The rapid evolution of Large Language Models(LLMs) highlights the necessity for ethical considerations and data integrity in AI development, particularly emphasizing the role of FAIR(Findable, Accessible, Interoperabl... The rapid evolution of Large Language Models(LLMs) highlights the necessity for ethical considerations and data integrity in AI development, particularly emphasizing the role of FAIR(Findable, Accessible, Interoperable, Reusable) data principles. While these principles are crucial for ethical data stewardship, their specific application in the context of LLM training data remains an under-explored area. This research gap is the focus of our study, which begins with an examination of existing literature to underline the importance of FAIR principles in managing data for LLM training. Building upon this, we propose a novel frame-work designed to integrate FAIR principles into the LLM development lifecycle. A contribution of our work is the development of a comprehensive checklist intended to guide researchers and developers in applying FAIR data principles consistently across the model development process. The utility and effectiveness of our frame-work are validated through a case study on creating a FAIR-compliant dataset aimed at detecting and mitigating biases in LLMs. We present this framework to the community as a tool to foster the creation of technologically advanced, ethically grounded, and socially responsible AI models. 展开更多
关键词 Responsible Al large language models FAIR data principles Ethical Al Biases
原文传递
LLM4CP:Adapting Large Language Models for Channel Prediction
14
作者 Boxun Liu Xuanyu Liu +2 位作者 Shijian Gao Xiang Cheng Liuqing Yang 《Journal of Communications and Information Networks》 EI CSCD 2024年第2期113-125,共13页
Channel prediction is an effective approach for reducing the feedback or estimation overhead in massive multi-input multi-output (m-MIMO) systems. However, existing channel prediction methods lack precision due to mod... Channel prediction is an effective approach for reducing the feedback or estimation overhead in massive multi-input multi-output (m-MIMO) systems. However, existing channel prediction methods lack precision due to model mismatch errors or network generalization issues. Large language models (LLMs) have demonstrated powerful modeling and generalization abilities, and have been successfully applied to cross-modal tasks, including the time series analysis. Leveraging the expressive power of LLMs, we propose a pre-trained LLM-empowered channel prediction(LLM4CP)method to predict the future downlink channel state information (CSI) sequence based on the historical uplink CSI sequence. We fine-tune the network while freezing most of the parameters of the pre-trained LLM for better cross-modality knowledge transfer. To bridge the gap between the channel data and the feature space of the LLM,preprocessor, embedding, and output modules are specifically tailored by taking into account unique channel characteristics. Simulations validate that the proposed method achieves state-of-the-art (SOTA) prediction performance on full-sample, few-shot, and generalization tests with low training and inference costs. 展开更多
关键词 channel prediction massive multi-input multi-output(m-MIMO) large language models(LLMs) fine-tuning time-series
原文传递
RecBERT:Semantic recommendation engine with large language model enhanced query segmentation for k-nearest neighbors ranking retrieval
15
作者 Richard Wu 《Intelligent and Converged Networks》 EI 2024年第1期42-52,共11页
The increasing amount of user traffic on Internet discussion forums has led to a huge amount of unstructured natural language data in the form of user comments.Most modern recommendation systems rely on manual tagging... The increasing amount of user traffic on Internet discussion forums has led to a huge amount of unstructured natural language data in the form of user comments.Most modern recommendation systems rely on manual tagging,relying on administrators to label the features of a class,or story,which a user comment corresponds to.Another common approach is to use pre-trained word embeddings to compare class descriptions for textual similarity,then use a distance metric such as cosine similarity or Euclidean distance to find top k neighbors.However,neither approach is able to fully utilize this user-generated unstructured natural language data,reducing the scope of these recommendation systems.This paper studies the application of domain adaptation on a transformer for the set of user comments to be indexed,and the use of simple contrastive learning for the sentence transformer fine-tuning process to generate meaningful semantic embeddings for the various user comments that apply to each class.In order to match a query containing content from multiple user comments belonging to the same class,the construction of a subquery channel for computing class-level similarity is proposed.This channel uses query segmentation of the aggregate query into subqueries,performing k-nearest neighbors(KNN)search on each individual subquery.RecBERT achieves state-of-the-art performance,outperforming other state-of-the-art models in accuracy,precision,recall,and F1 score for classifying comments between four and eight classes,respectively.RecBERT outperforms the most precise state-of-the-art model(distilRoBERTa)in precision by 6.97%for matching comments between eight classes. 展开更多
关键词 sentence transformer simple contrastive learning large language models query segmentation k-nearest neighbors
原文传递
Harnessing AI-human synergy for deep learning research analysis in ophthalmology with large language models assisting humans
16
作者 罗明杰 张玮星 +5 位作者 张哲铭 庞健宇 林桢哲 赵兰琴 林铎儒 林浩添 《Eye Science》 2024年第1期7-25,共19页
Background:Research innovations inocular disease screening,diagnosis,and management have been boosted by deep learning(DL)in the last decade.To assess historical research trends and current advances,we conducted an ar... Background:Research innovations inocular disease screening,diagnosis,and management have been boosted by deep learning(DL)in the last decade.To assess historical research trends and current advances,we conducted an artificial intelligence(AI)-human hybrid analysis of publications on DL in ophthalmology.Methods:All DL-related articles in ophthalmology,which were published between 2012 and 2022 from Web of Science,were included.500 high-impact articles annotated with key research information were used to fine-tune a large language models(LLM)for reviewing medical literature and extracting information.After verifying the LLM's accuracy in extracting diseases and imaging modalities,we analyzed trend of DL in ophthalmology with 2535 articles.Results:Researchers using LLM for literature analysis were 70%(P=0.0001)faster than those who did not,while achieving comparable accuracy(97%versus 98%,P=0.7681).The field of DL in ophthalmology has grown 116%annually,paralleling trends of the broader DL domain.The publications focused mainly on diabetic retinopathy(P=0.0003),glaucoma(P=0.0011),and age-related macular diseases(P=0.0001)using retinal fundus photographs(FP,P=0.0015)and optical coherence tomography(OCT,P=0.0001).DL studies utilizing multimodal images have been growing,with FP and OCT combined being the most frequent.Among the 500 high-impact articles,laboratory studies constituted the majority at 65.3%.Notably,a discernible decline in model accuracy was observed when categorizing by study design,notwithstanding its statistical insignificance.Furthermore,43 publicly available ocular image datasets were summarized.Conclusion:This study has characterized the landscape of publications on DL in ophthalmology,by identifying the trends and breakthroughs among research topics and the fast-growing areas.This study provides an efficient framework for combined AI-human analysis to comprehensively assess the current status and future trends in the field. 展开更多
关键词 large language model AI-human collaboration research trends OPHTHALMOLOGY model performance
原文传递
ChatModeler:基于大语言模型的人机协作迭代式需求获取和建模方法 被引量:1
17
作者 靳东明 金芝 +1 位作者 陈小红 王春晖 《计算机研究与发展》 EI CSCD 北大核心 2024年第2期338-350,共13页
需求获取和建模是需求工程中的关键步骤,影响后续系统设计与实现.传统的需求获取和建模方法通常由需求提供者、需求分析师等多类干系人共同协作、反复迭代完成,需要耗费大量的人力.如何减轻需求提供者与需求分析师的负担、提高获取和建... 需求获取和建模是需求工程中的关键步骤,影响后续系统设计与实现.传统的需求获取和建模方法通常由需求提供者、需求分析师等多类干系人共同协作、反复迭代完成,需要耗费大量的人力.如何减轻需求提供者与需求分析师的负担、提高获取和建模的效率有着重要意义.现有工作中有的使用知识库来提供更多知识,以辅助获取或者建模,有的利用自然语言处理等技术对获取或者建模过程进行自动化,但是它们并没有减轻需求提供者的负担.利用大语言模型(large language models,LLMs)的生成能力,提供了一种人机协作的迭代式需求获取和建模框架ChatModeler.具体来说,根据真实世界中需求团队的分工及协作关系,将部分需求提供者、需求分析师等角色的工作由大语言模型承担,而需求提供者只需要进行确认.为大语言模型扮演的各种角色进行了提示词设计,该提示词会随需求的元模型而变化.ChatModeler在7个需求案例上与3种需求模型的自动建模方法进行了14组对比实验,证明了ChatModeler在降低需求提供者的负担和生成高质量需求模型2个方面上的优越性. 展开更多
关键词 需求工程 需求诱导 大语言模型 人机协作 提示工程
下载PDF
Opportunities for large language models and discourse in engineering
18
作者 Jan Göpfert Jann M.Weinand +1 位作者 Patrick Kuckertz Detlef Stolten 《Energy and AI》 EI 2024年第3期136-142,共7页
In recent years, large language models have achieved breakthroughs on a wide range of benchmarks in natural language processing and continue to increase in performance. Recently, the advances of large language models ... In recent years, large language models have achieved breakthroughs on a wide range of benchmarks in natural language processing and continue to increase in performance. Recently, the advances of large language models have raised interest outside the natural language processing community and could have a large impact on daily life. In this paper, we pose the question: How will large language models and other foundation models shape the future product development process? We provide the reader with an overview of the subject by summarizing both recent advances in natural language processing and the use of information technology in the engineering design process. We argue that discourse should be regarded as the core of engineering design processes, and therefore should be represented in a digital artifact. On this basis, we describe how foundation models such as large language models could contribute to the design discourse by automating parts thereof that involve creativity and reasoning, and were previously reserved for humans. We describe how simulations, experiments, topology optimizations, and other process steps can be integrated into a machine-actionable, discourse-centric design process. As an example, we present a design discourse on the optimization of wind turbine blades. Finally, we outline the future research that will be necessary for the implementation of the conceptualized framework. 展开更多
关键词 Product development process Conceptual design Design methodology Design generation Natural language processing Foundation models multi-modal models
原文传递
A survey on large language model(LLM)security and privacy:The Good,The Bad,and The Ugly
19
作者 Yifan Yao Jinhao Duan +3 位作者 Kaidi Xu Yuanfang Cai Zhibo Sun Yue Zhang 《High-Confidence Computing》 EI 2024年第2期12-32,共21页
Large Language Models(LLMs),such as ChatGPT and Bard,have revolutionized natural language understanding and generation.They possess deep language comprehension,human-like text generation capabilities,contextual awaren... Large Language Models(LLMs),such as ChatGPT and Bard,have revolutionized natural language understanding and generation.They possess deep language comprehension,human-like text generation capabilities,contextual awareness,and robust problem-solving skills,making them invaluable in various domains(e.g.,search engines,customer support,translation).In the meantime,LLMs have also gained traction in the security community,revealing security vulnerabilities and showcasing their potential in security-related tasks.This paper explores the intersection of LLMs with security and privacy.Specifically,we investigate how LLMs positively impact security and privacy,potential risks and threats associated with their use,and inherent vulnerabilities within LLMs.Through a comprehensive literature review,the paper categorizes the papers into‘‘The Good’’(beneficial LLM applications),‘‘The Bad’’(offensive applications),and‘‘The Ugly’’(vulnerabilities of LLMs and their defenses).We have some interesting findings.For example,LLMs have proven to enhance code security(code vulnerability detection)and data privacy(data confidentiality protection),outperforming traditional methods.However,they can also be harnessed for various attacks(particularly user-level attacks)due to their human-like reasoning abilities.We have identified areas that require further research efforts.For example,Research on model and parameter extraction attacks is limited and often theoretical,hindered by LLM parameter scale and confidentiality.Safe instruction tuning,a recent development,requires more exploration.We hope that our work can shed light on the LLMs’potential to both bolster and jeopardize cybersecurity. 展开更多
关键词 large language model(LLM) LLM security LLM privacy ChatGPT LLM attacks LLM vulnerabilities
原文传递
Impact of Large Language Models on Scholarly Publication Titles and Abstracts:A Comparative Analysis
20
作者 Phoey Lee Teh Chukwudi Festus Uwasomba 《Journal of Social Computing》 EI 2024年第2期105-121,共17页
Artificial Intelligence(AI)tools become essential across industries,distinguishing AI-generated from human-authored text is increasingly challenging.This study assesses the coherence of AI-generated titles and corresp... Artificial Intelligence(AI)tools become essential across industries,distinguishing AI-generated from human-authored text is increasingly challenging.This study assesses the coherence of AI-generated titles and corresponding abstracts in anticipation of rising AI-assisted document production.Our main goal is to examine the correlation between original and AI-generated titles,emphasizing semantic depth and similarity measures,particularly in the context of Large Language Models(LLMs).We argue that LLMs have transformed research focus,dissemination,and citation patterns across five selected knowledge areas:Business Administration and Management(BAM),Computer Science and Information Technology(CS),Engineering and Material Science(EMS),Medicine and Healthcare(MH),and Psychology and Behavioral Sciences(PBS).We collected 15000 titles and abstracts,narrowing the selection to 2000 through a rigorous multi-stage screening process adhering to our study’s criteria.Result shows that there is insufficient evidence to suggest that LLM outperforms human authors in article title generation or articles from the LLM era demonstrates a marked difference in semantic richness and readability compared to those from the pre-LLM.Instead,it asserts that LLM is a valuable tool and can assist researchers in generating titles.With LLM’s assistance,the researcher ensures that the content is reflective of the finalized abstract and core research themes,potentially increasing the impact and accessibility and readability of the academic work. 展开更多
关键词 semantic alignment cosine similarity patterns large language models technology impact
原文传递
上一页 1 2 34 下一页 到第
使用帮助 返回顶部