In order to convey complete meanings,there is a phenomenon in Chinese of using multiple running sentences.Xu Jingning(2023,p.66)states,“In communication,a complete expression of meaning often requires more than one c...In order to convey complete meanings,there is a phenomenon in Chinese of using multiple running sentences.Xu Jingning(2023,p.66)states,“In communication,a complete expression of meaning often requires more than one clause,which is common in human languages.”Domestic research on running sentences includes discussions on defining the concept and structural features of running sentences,sentence properties,sentence pattern classifications and their criteria,as well as issues related to translating running sentences into English.This article primarily focuses on scholarly research into the English translation of running sentences in China,highlighting recent achievements and identifying existing issues in the study of running sentence translation.However,by reviewing literature on the translation of running sentences,it is found that current research in the academic community on non-core running sentences is limited.Therefore,this paper proposes relevant strategies to address this issue.展开更多
Sentence classification is the process of categorizing a sentence based on the context of the sentence.Sentence categorization requires more semantic highlights than other tasks,such as dependence parsing,which requir...Sentence classification is the process of categorizing a sentence based on the context of the sentence.Sentence categorization requires more semantic highlights than other tasks,such as dependence parsing,which requires more syntactic elements.Most existing strategies focus on the general semantics of a conversation without involving the context of the sentence,recognizing the progress and comparing impacts.An ensemble pre-trained language model was taken up here to classify the conversation sentences from the conversation corpus.The conversational sentences are classified into four categories:information,question,directive,and commission.These classification label sequences are for analyzing the conversation progress and predicting the pecking order of the conversation.Ensemble of Bidirectional Encoder for Representation of Transformer(BERT),Robustly Optimized BERT pretraining Approach(RoBERTa),Generative Pre-Trained Transformer(GPT),DistilBERT and Generalized Autoregressive Pretraining for Language Understanding(XLNet)models are trained on conversation corpus with hyperparameters.Hyperparameter tuning approach is carried out for better performance on sentence classification.This Ensemble of Pre-trained Language Models with a Hyperparameter Tuning(EPLM-HT)system is trained on an annotated conversation dataset.The proposed approach outperformed compared to the base BERT,GPT,DistilBERT and XLNet transformer models.The proposed ensemble model with the fine-tuned parameters achieved an F1_score of 0.88.展开更多
针对火电机组SO_(2)排放质量浓度的影响因素众多,难以准确预测的问题,提出一种改进向量加权平均(weighted mean of vectors,INFO)算法与双向长短期记忆(bi-directional long short term memory,Bi-LSTM)神经网络相结合的预测模型(改进IN...针对火电机组SO_(2)排放质量浓度的影响因素众多,难以准确预测的问题,提出一种改进向量加权平均(weighted mean of vectors,INFO)算法与双向长短期记忆(bi-directional long short term memory,Bi-LSTM)神经网络相结合的预测模型(改进INFO-Bi-LSTM模型)。采用Circle混沌映射和反向学习产生高质量初始化种群,引入自适应t分布提升INFO算法跳出局部最优解和全局搜索的能力。选取改进INFO-Bi-LSTM模型和多种预测模型对炉内外联合脱硫过程中4种典型工况下的SO_(2)排放质量浓度进行预测,将预测结果进行验证对比。结果表明:改进INFO算法的寻优能力得到提升,并且改进INFO-Bi-LSTM模型精度更高,更加适用于SO_(2)排放质量浓度的预测,可为变工况下的脱硫控制提供控制理论支撑。展开更多
文摘In order to convey complete meanings,there is a phenomenon in Chinese of using multiple running sentences.Xu Jingning(2023,p.66)states,“In communication,a complete expression of meaning often requires more than one clause,which is common in human languages.”Domestic research on running sentences includes discussions on defining the concept and structural features of running sentences,sentence properties,sentence pattern classifications and their criteria,as well as issues related to translating running sentences into English.This article primarily focuses on scholarly research into the English translation of running sentences in China,highlighting recent achievements and identifying existing issues in the study of running sentence translation.However,by reviewing literature on the translation of running sentences,it is found that current research in the academic community on non-core running sentences is limited.Therefore,this paper proposes relevant strategies to address this issue.
文摘Sentence classification is the process of categorizing a sentence based on the context of the sentence.Sentence categorization requires more semantic highlights than other tasks,such as dependence parsing,which requires more syntactic elements.Most existing strategies focus on the general semantics of a conversation without involving the context of the sentence,recognizing the progress and comparing impacts.An ensemble pre-trained language model was taken up here to classify the conversation sentences from the conversation corpus.The conversational sentences are classified into four categories:information,question,directive,and commission.These classification label sequences are for analyzing the conversation progress and predicting the pecking order of the conversation.Ensemble of Bidirectional Encoder for Representation of Transformer(BERT),Robustly Optimized BERT pretraining Approach(RoBERTa),Generative Pre-Trained Transformer(GPT),DistilBERT and Generalized Autoregressive Pretraining for Language Understanding(XLNet)models are trained on conversation corpus with hyperparameters.Hyperparameter tuning approach is carried out for better performance on sentence classification.This Ensemble of Pre-trained Language Models with a Hyperparameter Tuning(EPLM-HT)system is trained on an annotated conversation dataset.The proposed approach outperformed compared to the base BERT,GPT,DistilBERT and XLNet transformer models.The proposed ensemble model with the fine-tuned parameters achieved an F1_score of 0.88.
文摘针对火电机组SO_(2)排放质量浓度的影响因素众多,难以准确预测的问题,提出一种改进向量加权平均(weighted mean of vectors,INFO)算法与双向长短期记忆(bi-directional long short term memory,Bi-LSTM)神经网络相结合的预测模型(改进INFO-Bi-LSTM模型)。采用Circle混沌映射和反向学习产生高质量初始化种群,引入自适应t分布提升INFO算法跳出局部最优解和全局搜索的能力。选取改进INFO-Bi-LSTM模型和多种预测模型对炉内外联合脱硫过程中4种典型工况下的SO_(2)排放质量浓度进行预测,将预测结果进行验证对比。结果表明:改进INFO算法的寻优能力得到提升,并且改进INFO-Bi-LSTM模型精度更高,更加适用于SO_(2)排放质量浓度的预测,可为变工况下的脱硫控制提供控制理论支撑。