This paper retrieves and classifies reports on Jiangxi from 2010 to June 2024 in the NOW corpus,analyzing the status quo of Jiangxi’s international image portrayed by English news media through word frequency and col...This paper retrieves and classifies reports on Jiangxi from 2010 to June 2024 in the NOW corpus,analyzing the status quo of Jiangxi’s international image portrayed by English news media through word frequency and collocates.It is found that international attention towards Jiangxi is closely related to the holding of the World Conference on VR Industry in Nanchang,and there are apparent regional disparities.With a severe labeling of its mining industry,Jiangxi’s international image lacks independence and distinctiveness,resulting in a relatively monotonous overall image.Based on this,relevant suggestions for further enhancing Jiangxi’s international image are proposed.展开更多
Amid the increasingly severe global natural eco-environment,it is necessary to build a natural ecological civilization by constructing an ecological civilization discourse.Against this background,this study compiles a...Amid the increasingly severe global natural eco-environment,it is necessary to build a natural ecological civilization by constructing an ecological civilization discourse.Against this background,this study compiles a corpus of natural ecological discourse in the English translation of Xi Jinping:The Governance of China(short for Xi).By using Wordsmith and AntConc,this study explores the linguistic features of the ecological discourse in English translation of Xi in the following dimensions,including high-frequency words,keywords,word collocations,concordance lines.This study aims to analyze the concepts and attitudes towards natural ecology,so as to provide certain valuable insights for the construction of China’s discourse on natural ecological civilization.The study found that the natural ecology discourse involving in the English translation of Xi turned out to be ecologically beneficial.展开更多
Based on the Corpus of Contemporary American English(COCA),this study examines acupuncture from the perspective of mass media.Acupuncture has been circulating throughout the Chinese Cultural Circle since the Qin and H...Based on the Corpus of Contemporary American English(COCA),this study examines acupuncture from the perspective of mass media.Acupuncture has been circulating throughout the Chinese Cultural Circle since the Qin and Han Dynasties,and has since spread directly or indirectly to the rest of the world.The United States boasts the world’s second-largest acupuncture market,with its laws,regulations,industry growth,research,and education all positively influencing the development of acupuncture in other nations.This study uses COCA to analyze the form and content of acupuncture’s dissemination in eight different types of media.The findings show that acupuncture appears in COCA a total of 1,788 times,with the highest frequency in magazines,followed by blogs,and the lowest frequency in fiction.These findings reveal the popularity of acupuncture in mass media in the United States and provide empirical data and insights for the future dissemination and development of acupuncture in the United States.展开更多
This paper explores the application and effect evaluation of corpus linguistics in English teaching.It first introduces the basic concepts and development history of corpus linguistics,then analyzes its connection wit...This paper explores the application and effect evaluation of corpus linguistics in English teaching.It first introduces the basic concepts and development history of corpus linguistics,then analyzes its connection with language teaching theories,discussing the advantages and challenges of using corpora in language teaching.Next,it delves into the methods and technologies for building and utilizing corpora,as well as their role in developing teaching resources.Lastly,within the framework and methods of teaching effect evaluation,specific application cases of corpus linguistics in teaching design are analyzed,and their effects are assessed.Recommendations for teaching improvement and future development directions are also proposed.展开更多
Using Corpus of Contemporary American English as the source data,this paper carries out a corpus-based behavioral profile study to investigate four near-synonymous adjectives(serious,severe,grave,and grievous),focusin...Using Corpus of Contemporary American English as the source data,this paper carries out a corpus-based behavioral profile study to investigate four near-synonymous adjectives(serious,severe,grave,and grievous),focusing on their register and the types of nouns they each modify.Although sharing core meaning,these adjectives exhibit variations in formality levels and usage patterns.The identification of fine-grained usage differences complements the current inadequacies in describing these adjectives.Furthermore,the study reaffirms the effectiveness of the corpus-based behavioral profile approach in examining synonym differences.展开更多
Medical English,as a specialized form of English,has attracted considerable attention due to its unique linguistic features.Grammaticalization and lexicalization represent critical mechanisms in the evolution of langu...Medical English,as a specialized form of English,has attracted considerable attention due to its unique linguistic features.Grammaticalization and lexicalization represent critical mechanisms in the evolution of language and are essential for understanding and mastering medical English.This study,drawing on the MEDLINE corpus,investigates the characteristics of grammaticalization and lexicalization in medical English from the perspective of cognitive linguistics,aiming to unveil its underlying cognitive foundations.The findings suggest that the processes of grammaticalization and lexicalization in medical English reflect cognitive mechanisms such as categorization,inferential processing,and pragmatic strategies.These insights provide novel approaches and methodologies for the teaching of medical English.展开更多
In Content and Language Integrated Learning, one key issue is the differentiation between content area language (discipline-specific language) and everyday language and how the linguistic and competence gap can be bri...In Content and Language Integrated Learning, one key issue is the differentiation between content area language (discipline-specific language) and everyday language and how the linguistic and competence gap can be bridged. Language is the conveyor of meanings and it functions differently when the context that it is embedded in changes. In everyday language, frequently-encountered objects and concepts are present and the understanding of an idea is supported with abundant contextual clues. However, academic language is, in a sense, de-contextualized, in other words, context-reduced (Cummins, 2000). Notable features of academic language are high density of meanings, high level of formality, different syntactic structures from familiar and simpler everyday language, and discipline-specific terminology also poses great challenges to the comprehension. This paper will examine the differences between everyday language and the language of particular disciplines and the nature of discipline-specific language features. Drawn from the analysis, pedagogical suggestions are presented with reference to the teaching strategies and ways to bridge the gap. In addition, educational implications for CLIL will also be briefly discussed in the final section.展开更多
Nowadays,the usage of socialmedia platforms is rapidly increasing,and rumours or false information are also rising,especially among Arab nations.This false information is harmful to society and individuals.Blocking an...Nowadays,the usage of socialmedia platforms is rapidly increasing,and rumours or false information are also rising,especially among Arab nations.This false information is harmful to society and individuals.Blocking and detecting the spread of fake news in Arabic becomes critical.Several artificial intelligence(AI)methods,including contemporary transformer techniques,BERT,were used to detect fake news.Thus,fake news in Arabic is identified by utilizing AI approaches.This article develops a new hunterprey optimization with hybrid deep learning-based fake news detection(HPOHDL-FND)model on the Arabic corpus.The HPOHDL-FND technique undergoes extensive data pre-processing steps to transform the input data into a useful format.Besides,the HPOHDL-FND technique utilizes long-term memory with a recurrent neural network(LSTM-RNN)model for fake news detection and classification.Finally,hunter prey optimization(HPO)algorithm is exploited for optimal modification of the hyperparameters related to the LSTM-RNN model.The performance validation of the HPOHDL-FND technique is tested using two Arabic datasets.The outcomes exemplified better performance over the other existing techniques with maximum accuracy of 96.57%and 93.53%on Covid19Fakes and satirical datasets,respectively.展开更多
An obviously challenging problem in named entity recognition is the construction of the kind data set of entities.Although some research has been conducted on entity database construction,the majority of them are dire...An obviously challenging problem in named entity recognition is the construction of the kind data set of entities.Although some research has been conducted on entity database construction,the majority of them are directed at Wikipedia or the minority at structured entities such as people,locations and organizational nouns in the news.This paper focuses on the identification of scientific entities in carbonate platforms in English literature,using the example of carbonate platforms in sedimentology.Firstly,based on the fact that the reasons for writing literature in key disciplines are likely to be provided by multidisciplinary experts,this paper designs a literature content extraction method that allows dealing with complex text structures.Secondly,based on the literature extraction content,we formalize the entity extraction task(lexicon and lexical-based entity extraction)for entity extraction.Furthermore,for testing the accuracy of entity extraction,three currently popular recognition methods are chosen to perform entity detection in this paper.Experiments show that the entity data set provided by the lexicon and lexical-based entity extraction method is of significant assistance for the named entity recognition task.This study presents a pilot study of entity extraction,which involves the use of a complex structure and specialized literature on carbonate platforms in English.展开更多
Natural Language Processing(NLP)for the Arabic language has gained much significance in recent years.The most commonly-utilized NLP task is the‘Text Classification’process.Its main intention is to apply the Machine ...Natural Language Processing(NLP)for the Arabic language has gained much significance in recent years.The most commonly-utilized NLP task is the‘Text Classification’process.Its main intention is to apply the Machine Learning(ML)approaches for automatically classifying the textual files into one or more pre-defined categories.In ML approaches,the first and foremost crucial step is identifying an appropriate large dataset to test and train the method.One of the trending ML techniques,i.e.,Deep Learning(DL)technique needs huge volumes of different types of datasets for training to yield the best outcomes.The current study designs a new Dice Optimization with a Deep Hybrid Boltzmann Machinebased Arabic Corpus Classification(DODHBM-ACC)model in this background.The presented DODHBM-ACC model primarily relies upon different stages of pre-processing and the word2vec word embedding process.For Arabic text classification,the DHBM technique is utilized.This technique is a hybrid version of the Deep Boltzmann Machine(DBM)and Deep Belief Network(DBN).It has the advantage of learning the decisive intention of the classification process.To adjust the hyperparameters of the DHBM technique,the Dice Optimization Algorithm(DOA)is exploited in this study.The experimental analysis was conducted to establish the superior performance of the proposed DODHBM-ACC model.The outcomes inferred the better performance of the proposed DODHBM-ACC model over other recent approaches.展开更多
Sentiment analysis(SA)of the Arabic language becomes important despite scarce annotated corpora and confined sources.Arabic affect Analysis has become an active research zone nowadays.But still,the Arabic language lag...Sentiment analysis(SA)of the Arabic language becomes important despite scarce annotated corpora and confined sources.Arabic affect Analysis has become an active research zone nowadays.But still,the Arabic language lags behind adequate language sources for enabling the SA tasks.Thus,Arabic still faces challenges in natural language processing(NLP)tasks because of its structure complexities,history,and distinct cultures.It has gained lesser effort than the other languages.This paper developed a Multi-versus Optimization with Deep Reinforcement Learning Enabled Affect Analysis(MVODRL-AA)on Arabic Corpus.The presented MVODRL-AAmodelmajorly concentrates on identifying and classifying effects or emotions that occurred in the Arabic corpus.Firstly,the MVODRL-AA model follows data pre-processing and word embedding.Next,an n-gram model is utilized to generate word embeddings.A deep Q-learning network(DQLN)model is then exploited to identify and classify the effect on the Arabic corpus.At last,the MVO algorithm is used as a hyperparameter tuning approach to adjust the hyperparameters related to the DQLN model,showing the novelty of the work.A series of simulations were carried out to exhibit the promising performance of the MVODRL-AA model.The simulation outcomes illustrate the betterment of the MVODRL-AA method over the other approaches with an accuracy of 99.27%.展开更多
Background:The relationship between microRNA(miRNA)expression patterns and tumor mutation burden(TMB)in uterine corpus endometrial carcinoma(UCEC)was investigated in this study.Methods:The UCEC dataset from The Cancer...Background:The relationship between microRNA(miRNA)expression patterns and tumor mutation burden(TMB)in uterine corpus endometrial carcinoma(UCEC)was investigated in this study.Methods:The UCEC dataset from The Cancer Genome Atlas(TCGA)database was used to identify the miRNAs that differ in expression between high TMB and low TMB sample sets.The total sample sets were divided into a training set and a test set.TMB levels were predicted using miRNA-based signature classifiers developed by Lasso Cox regression.Test sets were used to validate the classifier.This study investigated the relationship between a miRNA-based signature classifier and three immune checkpoint molecules(programmed cell death protein 1[PD-1],programmed cell death ligand 1[PD-L1],cytotoxic T lymphocyte-associated antigen 4[CTLA-4]).For the miRNA-based signature classifier,functional enrichment analysis was performed on the miRNAs.An analysis of the relationship between PD-1,PD-L1,and CTLA-4 immune checkpoint genes was carried out using the miRNA-based signature classifier.Results:We identified 27 differentially expressed miRNAs in miRNA-base signature.For predicting the TMB level,27-miRNA-based signature classifiers had accuracies of 0.8689 in the training cohort,0.8276 in the test cohort,and 0.8524 in the total cohort.The correlation between the miRNA-based signature classifier and PD-1 was negative,while the correlation between PD-L1 and CTLA4 was positive.Based on the miRNA profiling described above,we validated the expression levels of 9 miRNAs in clinical samples by quantitative reverse transcription PCR(qRT-PCR).Four of them were highly expressed and many cancer-related and immune-associated biological processes were linked to these 27 miRNAs.Thus,the developed miRNA-based signature classifier was correlated with TMB levels that could also predict TMB levels in UCEC samples.Conclusion:In this study,we investigated the relationship between a miRNAbased signature classifier and TMB levels in Uterine Corpus Endometrial Carcinoma.Further,this is the first study to confirm their relationship in clinical samples,which may provide more evidence support for immunotherapy of endometrial cancer.展开更多
The term‘corpus’refers to a huge volume of structured datasets containing machine-readable texts.Such texts are generated in a natural communicative setting.The explosion of social media permitted individuals to spr...The term‘corpus’refers to a huge volume of structured datasets containing machine-readable texts.Such texts are generated in a natural communicative setting.The explosion of social media permitted individuals to spread data with minimal examination and filters freely.Due to this,the old problem of fake news has resurfaced.It has become an important concern due to its negative impact on the community.To manage the spread of fake news,automatic recognition approaches have been investigated earlier using Artificial Intelligence(AI)and Machine Learning(ML)techniques.To perform the medicinal text classification tasks,the ML approaches were applied,and they performed quite effectively.Still,a huge effort is required from the human side to generate the labelled training data.The recent progress of the Deep Learning(DL)methods seems to be a promising solution to tackle difficult types of Natural Language Processing(NLP)tasks,especially fake news detection.To unlock social media data,an automatic text classifier is highly helpful in the domain of NLP.The current research article focuses on the design of the Optimal Quad ChannelHybrid Long Short-Term Memory-based Fake News Classification(QCLSTM-FNC)approach.The presented QCLSTM-FNC approach aims to identify and differentiate fake news from actual news.To attain this,the proposed QCLSTM-FNC approach follows two methods such as the pre-processing data method and the Glovebased word embedding process.Besides,the QCLSTM model is utilized for classification.To boost the classification results of the QCLSTM model,a Quasi-Oppositional Sandpiper Optimization(QOSPO)algorithm is utilized to fine-tune the hyperparameters.The proposed QCLSTM-FNC approach was experimentally validated against a benchmark dataset.The QCLSTMFNC approach successfully outperformed all other existing DL models under different measures.展开更多
The difficulty of learning English is not only related to interest,but also related to the correctness of learning methods.Especially in English teaching,a comprehensive and in-depth mastery of vocabulary can improve ...The difficulty of learning English is not only related to interest,but also related to the correctness of learning methods.Especially in English teaching,a comprehensive and in-depth mastery of vocabulary can improve the level of English language,learn English knowledge better,and improve the level of cross-cultural communication.The application of corpus in English classroom vocabulary teaching can provide more educational space for vocabulary teaching,enrich teaching methods,and at the same time,facilitate students to learn vocabulary and lay a foundation for learning English language.To this end,this article first describes the important role of corpus application in vocabulary learning in English classroom teaching.Secondly,it discusses the difficulties of vocabulary learning and the factors that affect the quality of learning.Finally,in order to enhance the learning effect of students and improve the teaching level,several learning strategies have been formulated to continuously highlight the practicality of the corpus.展开更多
Background: Cytotoxic lesions of the corpus callosum (CLOCCs) represent a collection of disparate conditions that can cause a signal change in the corpus callosum, usually involving the splenium. CLOCCs is present in ...Background: Cytotoxic lesions of the corpus callosum (CLOCCs) represent a collection of disparate conditions that can cause a signal change in the corpus callosum, usually involving the splenium. CLOCCs is present in a variety of disorders, such as cerebral infarction, bleeding, multiple sclerosis, acute disseminated encephalomyelitis, glioblastoma, lymphoma, metabolic diseases, and infections. Since 2020, World Health Organization (W.H.O) defined Coronavirus disease 2019 (COVID-19), caused by SARS-CoV-2, as a pandemic. Numerous CLOCCs cases have been reported in adults in particular in Japan, in China, and recently in children in Turkey associated with SARS-CoV-2. We report the first case of CLOCCs diagnosed in West Africa (Côte d’Ivoire) in an adult associated with SARS-CoV-2. Case Report: A 60 year-old-woman with a medical history of high blood pressure and diabetes, presented to the emergency department with confusion without fever. Neurological examination was normal apart from temporospatial disorientation. Brain magnetic resonance imaging (MRI) showed abnormal signals in the splenium of the corpus callosum (SCC). Forty-eight hours (48 h) after admission, the patient experienced a fever (temperature: 385˚C), several episodes of hypoglycemia (capillary blood glycemia levels below 0.5 g/l) and a dry cough. Lung CT imaging showed typical features with ground-glass opacities. Oropharyngeal swab was positive for SARS-CoV-2 on reverse-transcriptase–polymerase-chain-reaction (RT-PCR) assay. The clinical course was favorable. One month after disease onset, a follow-up Brain MRI showed considerable regression of SCC abnormal signal. The multiple episodes of hypoglycemia and SARS-COV 2 infection were incriminated as the causal factors. Conclusion: The improvement of the technical platform in our context of work gives us the possibility to identify the etiological factors of this rare clinico-radiological entity.展开更多
The True Story of Ah-Q written by Lu Xun has been translated into forty different languages,making great contributions to the external dissemination of Chinese culture.This paper adopts the corpus stylistics method to...The True Story of Ah-Q written by Lu Xun has been translated into forty different languages,making great contributions to the external dissemination of Chinese culture.This paper adopts the corpus stylistics method to analyze the translators’styles presented by William A.Lyell and Julia Lovell’s translations of Ah-Q.Based on the self-built comparable corpus of the two translations,the paper investigates and analyzes them from the lexical and sentence level.In terms of vocabulary,the translations are statistically analyzed from the standardized type-token ratio,word frequency,word length and vocabulary density;at the sentence level,the article studies the average sentence lengths and sentence complexity.The first major finding is that Lovell’s vocabulary is richer than that of Lyell,and Lovell’s average word length is longer.Also,the average sentence length of both translations exceeds the original text,and Lovell has the highest average sentence length.Secondly,Lovell prefers to use simple sentences while Lyell prefers complex sentences.Nevertheless,both of them strive to make Lu Xun’s works understandable to foreign readers,thus understanding Chinese culture better.展开更多
Da Sheng Bian,a significant work on obstetrics and gynecology that emerged in the early Qing Dynasty,was initially published as“A Treatise on Midwifery”in 1842 by William Lockhart,a British missionary to China.In 18...Da Sheng Bian,a significant work on obstetrics and gynecology that emerged in the early Qing Dynasty,was initially published as“A Treatise on Midwifery”in 1842 by William Lockhart,a British missionary to China.In 1894,John G.Kerr,an American missionary,translated the text as“The Tat Shang Pin”.This paper conducts a comparative study of the two English translations using a self-constructed English-Chinese parallel corpus of Da Sheng Bian.The study explores the translation styles of the two translators by examining the token-types ratio and frequency at the lexical level,mean sentence length at the syntactic level,and the use of conjunctions at the discourse level.The observed differences in translation styles between the two translations are analyzed in relation to the translators’backgrounds and translation strategies.展开更多
Objective: To explore the related factors of surgical treatment of patients with corpus luteum rupture and establish a risk prediction model of surgical treatment of corpus luteum rupture. Methods: 222 patients with c...Objective: To explore the related factors of surgical treatment of patients with corpus luteum rupture and establish a risk prediction model of surgical treatment of corpus luteum rupture. Methods: 222 patients with corpus luteum rupture treated in Jingzhou First People’s Hospital from January 2015 to March 2022 were analyzed retrospectively, including 45 cases of surgery and 177 cases of conservative treatment. The training set and validation set were randomly assigned according to 7:3. We collected the basic information, laboratory and ultrasonic examination data of 222 patients. Logistic regression analysis was used to determine the independent risk factors and combined predictors of surgical treatment of corpus luteum rupture. The risk prediction model was established and the nomogram was drawn. The discrimination and calibration of the prediction model were verified and evaluated by receiver operating characteristic (ROC) curve, calibration curve and Hosmer-Lemeshow goodness of fit test;Decision curve analysis (DCA) was used to evaluate the clinical effectiveness of the prediction model. Results: Univariate logistic regression showed that whole abdominal pain (OR: 2.314, 95% CI: 1.090 - 4.912), abdominal muscle tension (OR: 2.379, 95% CI: 1.112 - 5.089), adnexal mass ≥ 4 cm (OR: 3.926, 95% CI: 1.771 - 8.266), hemoglobin Conclusion: The nomogram prediction model containing three predictive variables (hemoglobin, depth of pelvic effusion under ultrasound and cervical lifting pain) can be used to predict the risk of surgical treatment in patients with corpus luteum rupture.展开更多
This essay accesses to the approach of corpus applied to translation teaching, in order to improve the teaching methods, lay the foundation for the translation teaching reform, cultivate students research ability, and...This essay accesses to the approach of corpus applied to translation teaching, in order to improve the teaching methods, lay the foundation for the translation teaching reform, cultivate students research ability, and finally to establish a new type of translation teaching design-- "ability-development-oriented design". Also, this paper takes the word "good" for example, looking for the general rules to translate it and its common collocation, in order to design a translation class. Corpus-based learning and teaching provides us a new feasible way of translation class.展开更多
基金supported by Project of Improving the Basic Ability of Scientific Research of Young and Middle-Aged Teachers in Guangxi Universities (2021KY0277).
文摘This paper retrieves and classifies reports on Jiangxi from 2010 to June 2024 in the NOW corpus,analyzing the status quo of Jiangxi’s international image portrayed by English news media through word frequency and collocates.It is found that international attention towards Jiangxi is closely related to the holding of the World Conference on VR Industry in Nanchang,and there are apparent regional disparities.With a severe labeling of its mining industry,Jiangxi’s international image lacks independence and distinctiveness,resulting in a relatively monotonous overall image.Based on this,relevant suggestions for further enhancing Jiangxi’s international image are proposed.
文摘Amid the increasingly severe global natural eco-environment,it is necessary to build a natural ecological civilization by constructing an ecological civilization discourse.Against this background,this study compiles a corpus of natural ecological discourse in the English translation of Xi Jinping:The Governance of China(short for Xi).By using Wordsmith and AntConc,this study explores the linguistic features of the ecological discourse in English translation of Xi in the following dimensions,including high-frequency words,keywords,word collocations,concordance lines.This study aims to analyze the concepts and attitudes towards natural ecology,so as to provide certain valuable insights for the construction of China’s discourse on natural ecological civilization.The study found that the natural ecology discourse involving in the English translation of Xi turned out to be ecologically beneficial.
文摘Based on the Corpus of Contemporary American English(COCA),this study examines acupuncture from the perspective of mass media.Acupuncture has been circulating throughout the Chinese Cultural Circle since the Qin and Han Dynasties,and has since spread directly or indirectly to the rest of the world.The United States boasts the world’s second-largest acupuncture market,with its laws,regulations,industry growth,research,and education all positively influencing the development of acupuncture in other nations.This study uses COCA to analyze the form and content of acupuncture’s dissemination in eight different types of media.The findings show that acupuncture appears in COCA a total of 1,788 times,with the highest frequency in magazines,followed by blogs,and the lowest frequency in fiction.These findings reveal the popularity of acupuncture in mass media in the United States and provide empirical data and insights for the future dissemination and development of acupuncture in the United States.
文摘This paper explores the application and effect evaluation of corpus linguistics in English teaching.It first introduces the basic concepts and development history of corpus linguistics,then analyzes its connection with language teaching theories,discussing the advantages and challenges of using corpora in language teaching.Next,it delves into the methods and technologies for building and utilizing corpora,as well as their role in developing teaching resources.Lastly,within the framework and methods of teaching effect evaluation,specific application cases of corpus linguistics in teaching design are analyzed,and their effects are assessed.Recommendations for teaching improvement and future development directions are also proposed.
文摘Using Corpus of Contemporary American English as the source data,this paper carries out a corpus-based behavioral profile study to investigate four near-synonymous adjectives(serious,severe,grave,and grievous),focusing on their register and the types of nouns they each modify.Although sharing core meaning,these adjectives exhibit variations in formality levels and usage patterns.The identification of fine-grained usage differences complements the current inadequacies in describing these adjectives.Furthermore,the study reaffirms the effectiveness of the corpus-based behavioral profile approach in examining synonym differences.
基金Project of Social Science Achievement Evaluation Committee of Hunan Province of China(XSP2023WXC037)Project of Teaching Reform Research of Hunan Province of China(202401001822)。
文摘Medical English,as a specialized form of English,has attracted considerable attention due to its unique linguistic features.Grammaticalization and lexicalization represent critical mechanisms in the evolution of language and are essential for understanding and mastering medical English.This study,drawing on the MEDLINE corpus,investigates the characteristics of grammaticalization and lexicalization in medical English from the perspective of cognitive linguistics,aiming to unveil its underlying cognitive foundations.The findings suggest that the processes of grammaticalization and lexicalization in medical English reflect cognitive mechanisms such as categorization,inferential processing,and pragmatic strategies.These insights provide novel approaches and methodologies for the teaching of medical English.
文摘In Content and Language Integrated Learning, one key issue is the differentiation between content area language (discipline-specific language) and everyday language and how the linguistic and competence gap can be bridged. Language is the conveyor of meanings and it functions differently when the context that it is embedded in changes. In everyday language, frequently-encountered objects and concepts are present and the understanding of an idea is supported with abundant contextual clues. However, academic language is, in a sense, de-contextualized, in other words, context-reduced (Cummins, 2000). Notable features of academic language are high density of meanings, high level of formality, different syntactic structures from familiar and simpler everyday language, and discipline-specific terminology also poses great challenges to the comprehension. This paper will examine the differences between everyday language and the language of particular disciplines and the nature of discipline-specific language features. Drawn from the analysis, pedagogical suggestions are presented with reference to the teaching strategies and ways to bridge the gap. In addition, educational implications for CLIL will also be briefly discussed in the final section.
基金The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through Small Groups Project under Grant Number(120/43)Princess Nourah bint Abdulrahman University Researchers Supporting Project Number(PNURSP2022R281)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code:(22UQU4331004DSR32).
文摘Nowadays,the usage of socialmedia platforms is rapidly increasing,and rumours or false information are also rising,especially among Arab nations.This false information is harmful to society and individuals.Blocking and detecting the spread of fake news in Arabic becomes critical.Several artificial intelligence(AI)methods,including contemporary transformer techniques,BERT,were used to detect fake news.Thus,fake news in Arabic is identified by utilizing AI approaches.This article develops a new hunterprey optimization with hybrid deep learning-based fake news detection(HPOHDL-FND)model on the Arabic corpus.The HPOHDL-FND technique undergoes extensive data pre-processing steps to transform the input data into a useful format.Besides,the HPOHDL-FND technique utilizes long-term memory with a recurrent neural network(LSTM-RNN)model for fake news detection and classification.Finally,hunter prey optimization(HPO)algorithm is exploited for optimal modification of the hyperparameters related to the LSTM-RNN model.The performance validation of the HPOHDL-FND technique is tested using two Arabic datasets.The outcomes exemplified better performance over the other existing techniques with maximum accuracy of 96.57%and 93.53%on Covid19Fakes and satirical datasets,respectively.
基金supported by the National Natural Science Foundation of China under Grant No.42050102the National Science Foundation of China(Grant No.62001236)the Natural Science Foundation of the Jiangsu Higher Education Institutions of China(Grant No.20KJA520003).
文摘An obviously challenging problem in named entity recognition is the construction of the kind data set of entities.Although some research has been conducted on entity database construction,the majority of them are directed at Wikipedia or the minority at structured entities such as people,locations and organizational nouns in the news.This paper focuses on the identification of scientific entities in carbonate platforms in English literature,using the example of carbonate platforms in sedimentology.Firstly,based on the fact that the reasons for writing literature in key disciplines are likely to be provided by multidisciplinary experts,this paper designs a literature content extraction method that allows dealing with complex text structures.Secondly,based on the literature extraction content,we formalize the entity extraction task(lexicon and lexical-based entity extraction)for entity extraction.Furthermore,for testing the accuracy of entity extraction,three currently popular recognition methods are chosen to perform entity detection in this paper.Experiments show that the entity data set provided by the lexicon and lexical-based entity extraction method is of significant assistance for the named entity recognition task.This study presents a pilot study of entity extraction,which involves the use of a complex structure and specialized literature on carbonate platforms in English.
基金Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2022R263)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code:(22UQU4310373DSR53).
文摘Natural Language Processing(NLP)for the Arabic language has gained much significance in recent years.The most commonly-utilized NLP task is the‘Text Classification’process.Its main intention is to apply the Machine Learning(ML)approaches for automatically classifying the textual files into one or more pre-defined categories.In ML approaches,the first and foremost crucial step is identifying an appropriate large dataset to test and train the method.One of the trending ML techniques,i.e.,Deep Learning(DL)technique needs huge volumes of different types of datasets for training to yield the best outcomes.The current study designs a new Dice Optimization with a Deep Hybrid Boltzmann Machinebased Arabic Corpus Classification(DODHBM-ACC)model in this background.The presented DODHBM-ACC model primarily relies upon different stages of pre-processing and the word2vec word embedding process.For Arabic text classification,the DHBM technique is utilized.This technique is a hybrid version of the Deep Boltzmann Machine(DBM)and Deep Belief Network(DBN).It has the advantage of learning the decisive intention of the classification process.To adjust the hyperparameters of the DHBM technique,the Dice Optimization Algorithm(DOA)is exploited in this study.The experimental analysis was conducted to establish the superior performance of the proposed DODHBM-ACC model.The outcomes inferred the better performance of the proposed DODHBM-ACC model over other recent approaches.
基金Princess Nourah bint Abdulrahman University Researchers Supporting Project Number (PNURSP2022R263)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Ara-bia.The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code:22UQU4340237DSR38.
文摘Sentiment analysis(SA)of the Arabic language becomes important despite scarce annotated corpora and confined sources.Arabic affect Analysis has become an active research zone nowadays.But still,the Arabic language lags behind adequate language sources for enabling the SA tasks.Thus,Arabic still faces challenges in natural language processing(NLP)tasks because of its structure complexities,history,and distinct cultures.It has gained lesser effort than the other languages.This paper developed a Multi-versus Optimization with Deep Reinforcement Learning Enabled Affect Analysis(MVODRL-AA)on Arabic Corpus.The presented MVODRL-AAmodelmajorly concentrates on identifying and classifying effects or emotions that occurred in the Arabic corpus.Firstly,the MVODRL-AA model follows data pre-processing and word embedding.Next,an n-gram model is utilized to generate word embeddings.A deep Q-learning network(DQLN)model is then exploited to identify and classify the effect on the Arabic corpus.At last,the MVO algorithm is used as a hyperparameter tuning approach to adjust the hyperparameters related to the DQLN model,showing the novelty of the work.A series of simulations were carried out to exhibit the promising performance of the MVODRL-AA model.The simulation outcomes illustrate the betterment of the MVODRL-AA method over the other approaches with an accuracy of 99.27%.
基金the National Natural Science Foundation(81803877,82104705)the Natural Science Foundation of Guangdong Province of China(2017A030310178)+5 种基金the Guangdong Sci-Tech Commissioner(20211800500322)the China Postdoctoral Science Foundation(2020M682817)Guangdong Basic and Applied Basic Research Foundation(2020A1515110651,2020B1515120063)Guangdong Medical Science and Technology Research Foundation(A2021476)Traditional Chinese Medicine Research Project of Guangdong Province Traditional Chinese Medicine Bureau(20221256)the Dongguan Social Technology Development Fund(202050715001207).
文摘Background:The relationship between microRNA(miRNA)expression patterns and tumor mutation burden(TMB)in uterine corpus endometrial carcinoma(UCEC)was investigated in this study.Methods:The UCEC dataset from The Cancer Genome Atlas(TCGA)database was used to identify the miRNAs that differ in expression between high TMB and low TMB sample sets.The total sample sets were divided into a training set and a test set.TMB levels were predicted using miRNA-based signature classifiers developed by Lasso Cox regression.Test sets were used to validate the classifier.This study investigated the relationship between a miRNA-based signature classifier and three immune checkpoint molecules(programmed cell death protein 1[PD-1],programmed cell death ligand 1[PD-L1],cytotoxic T lymphocyte-associated antigen 4[CTLA-4]).For the miRNA-based signature classifier,functional enrichment analysis was performed on the miRNAs.An analysis of the relationship between PD-1,PD-L1,and CTLA-4 immune checkpoint genes was carried out using the miRNA-based signature classifier.Results:We identified 27 differentially expressed miRNAs in miRNA-base signature.For predicting the TMB level,27-miRNA-based signature classifiers had accuracies of 0.8689 in the training cohort,0.8276 in the test cohort,and 0.8524 in the total cohort.The correlation between the miRNA-based signature classifier and PD-1 was negative,while the correlation between PD-L1 and CTLA4 was positive.Based on the miRNA profiling described above,we validated the expression levels of 9 miRNAs in clinical samples by quantitative reverse transcription PCR(qRT-PCR).Four of them were highly expressed and many cancer-related and immune-associated biological processes were linked to these 27 miRNAs.Thus,the developed miRNA-based signature classifier was correlated with TMB levels that could also predict TMB levels in UCEC samples.Conclusion:In this study,we investigated the relationship between a miRNAbased signature classifier and TMB levels in Uterine Corpus Endometrial Carcinoma.Further,this is the first study to confirm their relationship in clinical samples,which may provide more evidence support for immunotherapy of endometrial cancer.
基金Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2022R281)PrincessNourah bint Abdulrahman University,Riyadh,Saudi Arabia.The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code:(22UQU4331004DSR41).
文摘The term‘corpus’refers to a huge volume of structured datasets containing machine-readable texts.Such texts are generated in a natural communicative setting.The explosion of social media permitted individuals to spread data with minimal examination and filters freely.Due to this,the old problem of fake news has resurfaced.It has become an important concern due to its negative impact on the community.To manage the spread of fake news,automatic recognition approaches have been investigated earlier using Artificial Intelligence(AI)and Machine Learning(ML)techniques.To perform the medicinal text classification tasks,the ML approaches were applied,and they performed quite effectively.Still,a huge effort is required from the human side to generate the labelled training data.The recent progress of the Deep Learning(DL)methods seems to be a promising solution to tackle difficult types of Natural Language Processing(NLP)tasks,especially fake news detection.To unlock social media data,an automatic text classifier is highly helpful in the domain of NLP.The current research article focuses on the design of the Optimal Quad ChannelHybrid Long Short-Term Memory-based Fake News Classification(QCLSTM-FNC)approach.The presented QCLSTM-FNC approach aims to identify and differentiate fake news from actual news.To attain this,the proposed QCLSTM-FNC approach follows two methods such as the pre-processing data method and the Glovebased word embedding process.Besides,the QCLSTM model is utilized for classification.To boost the classification results of the QCLSTM model,a Quasi-Oppositional Sandpiper Optimization(QOSPO)algorithm is utilized to fine-tune the hyperparameters.The proposed QCLSTM-FNC approach was experimentally validated against a benchmark dataset.The QCLSTMFNC approach successfully outperformed all other existing DL models under different measures.
基金This paper is the research result of the school level scientific research project of Chengdu College of Arts and Sciences,“Research on the Interpretation of Chinese Cultural Images and Translation Teaching Strategies in Accordance With Internationalization Strategy”(Fund Project No.:WLYB2022066)the research result of China Private Education Association’s 2022 annual planning topic“Research on the Application of Corpus in English Teaching in Private Colleges and Universities”(Fund Project No.:CANFZG22201).
文摘The difficulty of learning English is not only related to interest,but also related to the correctness of learning methods.Especially in English teaching,a comprehensive and in-depth mastery of vocabulary can improve the level of English language,learn English knowledge better,and improve the level of cross-cultural communication.The application of corpus in English classroom vocabulary teaching can provide more educational space for vocabulary teaching,enrich teaching methods,and at the same time,facilitate students to learn vocabulary and lay a foundation for learning English language.To this end,this article first describes the important role of corpus application in vocabulary learning in English classroom teaching.Secondly,it discusses the difficulties of vocabulary learning and the factors that affect the quality of learning.Finally,in order to enhance the learning effect of students and improve the teaching level,several learning strategies have been formulated to continuously highlight the practicality of the corpus.
文摘Background: Cytotoxic lesions of the corpus callosum (CLOCCs) represent a collection of disparate conditions that can cause a signal change in the corpus callosum, usually involving the splenium. CLOCCs is present in a variety of disorders, such as cerebral infarction, bleeding, multiple sclerosis, acute disseminated encephalomyelitis, glioblastoma, lymphoma, metabolic diseases, and infections. Since 2020, World Health Organization (W.H.O) defined Coronavirus disease 2019 (COVID-19), caused by SARS-CoV-2, as a pandemic. Numerous CLOCCs cases have been reported in adults in particular in Japan, in China, and recently in children in Turkey associated with SARS-CoV-2. We report the first case of CLOCCs diagnosed in West Africa (Côte d’Ivoire) in an adult associated with SARS-CoV-2. Case Report: A 60 year-old-woman with a medical history of high blood pressure and diabetes, presented to the emergency department with confusion without fever. Neurological examination was normal apart from temporospatial disorientation. Brain magnetic resonance imaging (MRI) showed abnormal signals in the splenium of the corpus callosum (SCC). Forty-eight hours (48 h) after admission, the patient experienced a fever (temperature: 385˚C), several episodes of hypoglycemia (capillary blood glycemia levels below 0.5 g/l) and a dry cough. Lung CT imaging showed typical features with ground-glass opacities. Oropharyngeal swab was positive for SARS-CoV-2 on reverse-transcriptase–polymerase-chain-reaction (RT-PCR) assay. The clinical course was favorable. One month after disease onset, a follow-up Brain MRI showed considerable regression of SCC abnormal signal. The multiple episodes of hypoglycemia and SARS-COV 2 infection were incriminated as the causal factors. Conclusion: The improvement of the technical platform in our context of work gives us the possibility to identify the etiological factors of this rare clinico-radiological entity.
文摘The True Story of Ah-Q written by Lu Xun has been translated into forty different languages,making great contributions to the external dissemination of Chinese culture.This paper adopts the corpus stylistics method to analyze the translators’styles presented by William A.Lyell and Julia Lovell’s translations of Ah-Q.Based on the self-built comparable corpus of the two translations,the paper investigates and analyzes them from the lexical and sentence level.In terms of vocabulary,the translations are statistically analyzed from the standardized type-token ratio,word frequency,word length and vocabulary density;at the sentence level,the article studies the average sentence lengths and sentence complexity.The first major finding is that Lovell’s vocabulary is richer than that of Lyell,and Lovell’s average word length is longer.Also,the average sentence length of both translations exceeds the original text,and Lovell has the highest average sentence length.Secondly,Lovell prefers to use simple sentences while Lyell prefers complex sentences.Nevertheless,both of them strive to make Lu Xun’s works understandable to foreign readers,thus understanding Chinese culture better.
基金Phased research results of“The Research and Practice of International Communication Ability Training of English Students”(Approval No.HNJG-2022-0678)。
文摘Da Sheng Bian,a significant work on obstetrics and gynecology that emerged in the early Qing Dynasty,was initially published as“A Treatise on Midwifery”in 1842 by William Lockhart,a British missionary to China.In 1894,John G.Kerr,an American missionary,translated the text as“The Tat Shang Pin”.This paper conducts a comparative study of the two English translations using a self-constructed English-Chinese parallel corpus of Da Sheng Bian.The study explores the translation styles of the two translators by examining the token-types ratio and frequency at the lexical level,mean sentence length at the syntactic level,and the use of conjunctions at the discourse level.The observed differences in translation styles between the two translations are analyzed in relation to the translators’backgrounds and translation strategies.
文摘Objective: To explore the related factors of surgical treatment of patients with corpus luteum rupture and establish a risk prediction model of surgical treatment of corpus luteum rupture. Methods: 222 patients with corpus luteum rupture treated in Jingzhou First People’s Hospital from January 2015 to March 2022 were analyzed retrospectively, including 45 cases of surgery and 177 cases of conservative treatment. The training set and validation set were randomly assigned according to 7:3. We collected the basic information, laboratory and ultrasonic examination data of 222 patients. Logistic regression analysis was used to determine the independent risk factors and combined predictors of surgical treatment of corpus luteum rupture. The risk prediction model was established and the nomogram was drawn. The discrimination and calibration of the prediction model were verified and evaluated by receiver operating characteristic (ROC) curve, calibration curve and Hosmer-Lemeshow goodness of fit test;Decision curve analysis (DCA) was used to evaluate the clinical effectiveness of the prediction model. Results: Univariate logistic regression showed that whole abdominal pain (OR: 2.314, 95% CI: 1.090 - 4.912), abdominal muscle tension (OR: 2.379, 95% CI: 1.112 - 5.089), adnexal mass ≥ 4 cm (OR: 3.926, 95% CI: 1.771 - 8.266), hemoglobin Conclusion: The nomogram prediction model containing three predictive variables (hemoglobin, depth of pelvic effusion under ultrasound and cervical lifting pain) can be used to predict the risk of surgical treatment in patients with corpus luteum rupture.
文摘This essay accesses to the approach of corpus applied to translation teaching, in order to improve the teaching methods, lay the foundation for the translation teaching reform, cultivate students research ability, and finally to establish a new type of translation teaching design-- "ability-development-oriented design". Also, this paper takes the word "good" for example, looking for the general rules to translate it and its common collocation, in order to design a translation class. Corpus-based learning and teaching provides us a new feasible way of translation class.