Recognizing translatability instead of untranslatability is of significance, since a wide-spread recognition of untranslatability may daunt the efforts of translators. The paper approaches the question of untranslatab...Recognizing translatability instead of untranslatability is of significance, since a wide-spread recognition of untranslatability may daunt the efforts of translators. The paper approaches the question of untranslatability thorough categorizing untranslatability into different groups and then examining untranslatability in each group by analyzing some typical examples from David Hawkes' translation of Hong Lou Meng. Through the analysis of how David Hawkes translated the untranslatables, the paper argues that real untranslatability is rare, while translatability rules.展开更多
The culture-loaded word is a symbol of national culture,since national culture with distinctive features is directly or indirectly reflected in its vocabulary.This paper is driven both by qualitative research and by q...The culture-loaded word is a symbol of national culture,since national culture with distinctive features is directly or indirectly reflected in its vocabulary.This paper is driven both by qualitative research and by quantitative method through the comparison and analysis of translation strategies involved in the five types of material culture-loaded-words,namely,apparel,diet,equipment,architecture and medicine,in The Story of Stone by David Hawkes.With the final quantitative statistics of the proportion of each translation method,it aims to scrutinize either the broad spectrum or the specific characteristics of those translation strategies so as to provide a perspective for the study of cultural translations.展开更多
In this study,our aim is to address the problem of gene selection by proposing a hybrid bio-inspired evolutionary algorithm that combines Grey Wolf Optimization(GWO)with Harris Hawks Optimization(HHO)for feature selec...In this study,our aim is to address the problem of gene selection by proposing a hybrid bio-inspired evolutionary algorithm that combines Grey Wolf Optimization(GWO)with Harris Hawks Optimization(HHO)for feature selection.Themotivation for utilizingGWOandHHOstems fromtheir bio-inspired nature and their demonstrated success in optimization problems.We aimto leverage the strengths of these algorithms to enhance the effectiveness of feature selection in microarray-based cancer classification.We selected leave-one-out cross-validation(LOOCV)to evaluate the performance of both two widely used classifiers,k-nearest neighbors(KNN)and support vector machine(SVM),on high-dimensional cancer microarray data.The proposed method is extensively tested on six publicly available cancer microarray datasets,and a comprehensive comparison with recently published methods is conducted.Our hybrid algorithm demonstrates its effectiveness in improving classification performance,Surpassing alternative approaches in terms of precision.The outcomes confirm the capability of our method to substantially improve both the precision and efficiency of cancer classification,thereby advancing the development ofmore efficient treatment strategies.The proposed hybridmethod offers a promising solution to the gene selection problem in microarray-based cancer classification.It improves the accuracy and efficiency of cancer diagnosis and treatment,and its superior performance compared to other methods highlights its potential applicability in realworld cancer classification tasks.By harnessing the complementary search mechanisms of GWO and HHO,we leverage their bio-inspired behavior to identify informative genes relevant to cancer diagnosis and treatment.展开更多
AIM:To develop a classifier for traditional Chinese medicine(TCM)syndrome differentiation of diabetic retinopathy(DR),using optimized machine learning algorithms,which can provide the basis for TCM objective and intel...AIM:To develop a classifier for traditional Chinese medicine(TCM)syndrome differentiation of diabetic retinopathy(DR),using optimized machine learning algorithms,which can provide the basis for TCM objective and intelligent syndrome differentiation.METHODS:Collated data on real-world DR cases were collected.A variety of machine learning methods were used to construct TCM syndrome classification model,and the best performance was selected as the basic model.Genetic Algorithm(GA)was used for feature selection to obtain the optimal feature combination.Harris Hawk Optimization(HHO)was used for parameter optimization,and a classification model based on feature selection and parameter optimization was constructed.The performance of the model was compared with other optimization algorithms.The models were evaluated with accuracy,precision,recall,and F1 score as indicators.RESULTS:Data on 970 cases that met screening requirements were collected.Support Vector Machine(SVM)was the best basic classification model.The accuracy rate of the model was 82.05%,the precision rate was 82.34%,the recall rate was 81.81%,and the F1 value was 81.76%.After GA screening,the optimal feature combination contained 37 feature values,which was consistent with TCM clinical practice.The model based on optimal combination and SVM(GA_SVM)had an accuracy improvement of 1.92%compared to the basic classifier.SVM model based on HHO and GA optimization(HHO_GA_SVM)had the best performance and convergence speed compared with other optimization algorithms.Compared with the basic classification model,the accuracy was improved by 3.51%.CONCLUSION:HHO and GA optimization can improve the model performance of SVM in TCM syndrome differentiation of DR.It provides a new method and research idea for TCM intelligent assisted syndrome differentiation.展开更多
Flexible job shop scheduling problem(FJSP)is the core decision-making problem of intelligent manufacturing production management.The Harris hawk optimization(HHO)algorithm,as a typical metaheuristic algorithm,has been...Flexible job shop scheduling problem(FJSP)is the core decision-making problem of intelligent manufacturing production management.The Harris hawk optimization(HHO)algorithm,as a typical metaheuristic algorithm,has been widely employed to solve scheduling problems.However,HHO suffers from premature convergence when solving NP-hard problems.Therefore,this paper proposes an improved HHO algorithm(GNHHO)to solve the FJSP.GNHHO introduces an elitism strategy,a chaotic mechanism,a nonlinear escaping energy update strategy,and a Gaussian random walk strategy to prevent premature convergence.A flexible job shop scheduling model is constructed,and the static and dynamic FJSP is investigated to minimize the makespan.This paper chooses a two-segment encoding mode based on the job and the machine of the FJSP.To verify the effectiveness of GNHHO,this study tests it in 23 benchmark functions,10 standard job shop scheduling problems(JSPs),and 5 standard FJSPs.Besides,this study collects data from an agricultural company and uses the GNHHO algorithm to optimize the company’s FJSP.The optimized scheduling scheme demonstrates significant improvements in makespan,with an advancement of 28.16%for static scheduling and 35.63%for dynamic scheduling.Moreover,it achieves an average increase of 21.50%in the on-time order delivery rate.The results demonstrate that the performance of the GNHHO algorithm in solving FJSP is superior to some existing algorithms.展开更多
In the study of Terrestrial Gamma-ray Flashes (TGFs) and Sonoluminescence, we observe parallels with larger cosmic events. Specifically, sonoluminescence involves the rapid collapse of bubbles, which closely resembles...In the study of Terrestrial Gamma-ray Flashes (TGFs) and Sonoluminescence, we observe parallels with larger cosmic events. Specifically, sonoluminescence involves the rapid collapse of bubbles, which closely resembles gravitational collapse in space. This observation suggests the potential formation of low-density quantum black holes. These entities, which might be related to dark matter, are thought to experience a kind of transient evaporation similar to Hawking radiation seen in cosmic black holes. Consequently, sonoluminescence could be a valuable tool for investigating phenomena typically linked to cosmic scale events. Furthermore, the role of the Higgs boson is considered in this context, possibly connecting it to both TGFs and sonoluminescence. This research could enhance our understanding of the quantum mechanics of black holes and their relation to dark matter on Earth.展开更多
Using a rigorous mathematical approach, we demonstrate how the Cosmic Microwave Background (CMB) temperature could simply be a form of geometric mean temperature between the minimum time-dependent Hawking Hubble tempe...Using a rigorous mathematical approach, we demonstrate how the Cosmic Microwave Background (CMB) temperature could simply be a form of geometric mean temperature between the minimum time-dependent Hawking Hubble temperature and the maximum Planck temperature of the expanding universe over the course of cosmic time. This mathematical discovery suggests a re-consideration of Rh=ctcosmological models, including black hole cosmological models, even if it possibly could also be consistent with the Λ-CDM model. Most importantly, this paper contributes to the growing literature in the past year asserting a tightly constrained mathematical relationship between the CMB temperature, the Hubble constant, and other global parameters of the Hubble sphere. Our approach suggests a solid theoretical framework for predicting and understanding the CMB temperature rather than solely observing it.1.展开更多
In real-world applications, datasets frequently contain outliers, which can hinder the generalization ability of machine learning models. Bayesian classifiers, a popular supervised learning method, rely on accurate pr...In real-world applications, datasets frequently contain outliers, which can hinder the generalization ability of machine learning models. Bayesian classifiers, a popular supervised learning method, rely on accurate probability density estimation for classifying continuous datasets. However, achieving precise density estimation with datasets containing outliers poses a significant challenge. This paper introduces a Bayesian classifier that utilizes optimized robust kernel density estimation to address this issue. Our proposed method enhances the accuracy of probability density distribution estimation by mitigating the impact of outliers on the training sample’s estimated distribution. Unlike the conventional kernel density estimator, our robust estimator can be seen as a weighted kernel mapping summary for each sample. This kernel mapping performs the inner product in the Hilbert space, allowing the kernel density estimation to be considered the average of the samples’ mapping in the Hilbert space using a reproducing kernel. M-estimation techniques are used to obtain accurate mean values and solve the weights. Meanwhile, complete cross-validation is used as the objective function to search for the optimal bandwidth, which impacts the estimator. The Harris Hawks Optimisation optimizes the objective function to improve the estimation accuracy. The experimental results show that it outperforms other optimization algorithms regarding convergence speed and objective function value during the bandwidth search. The optimal robust kernel density estimator achieves better fitness performance than the traditional kernel density estimator when the training data contains outliers. The Naïve Bayesian with optimal robust kernel density estimation improves the generalization in the classification with outliers.展开更多
Aiming at the problems that the original Harris Hawk optimization algorithm is easy to fall into local optimum and slow in finding the optimum,this paper proposes an improved Harris Hawk optimization algorithm(GHHO).F...Aiming at the problems that the original Harris Hawk optimization algorithm is easy to fall into local optimum and slow in finding the optimum,this paper proposes an improved Harris Hawk optimization algorithm(GHHO).Firstly,we used a Gaussian chaotic mapping strategy to initialize the positions of individuals in the population,which enriches the initial individual species characteristics.Secondly,by optimizing the energy parameter and introducing the cosine strategy,the algorithm's ability to jump out of the local optimum is enhanced,which improves the performance of the algorithm.Finally,comparison experiments with other intelligent algorithms were conducted on 13 classical test function sets.The results show that GHHO has better performance in all aspects compared to other optimization algorithms.The improved algorithm is more suitable for generalization to real optimization problems.展开更多
文摘Recognizing translatability instead of untranslatability is of significance, since a wide-spread recognition of untranslatability may daunt the efforts of translators. The paper approaches the question of untranslatability thorough categorizing untranslatability into different groups and then examining untranslatability in each group by analyzing some typical examples from David Hawkes' translation of Hong Lou Meng. Through the analysis of how David Hawkes translated the untranslatables, the paper argues that real untranslatability is rare, while translatability rules.
文摘The culture-loaded word is a symbol of national culture,since national culture with distinctive features is directly or indirectly reflected in its vocabulary.This paper is driven both by qualitative research and by quantitative method through the comparison and analysis of translation strategies involved in the five types of material culture-loaded-words,namely,apparel,diet,equipment,architecture and medicine,in The Story of Stone by David Hawkes.With the final quantitative statistics of the proportion of each translation method,it aims to scrutinize either the broad spectrum or the specific characteristics of those translation strategies so as to provide a perspective for the study of cultural translations.
基金the Deputyship for Research and Innovation,“Ministry of Education”in Saudi Arabia for funding this research(IFKSUOR3-014-3).
文摘In this study,our aim is to address the problem of gene selection by proposing a hybrid bio-inspired evolutionary algorithm that combines Grey Wolf Optimization(GWO)with Harris Hawks Optimization(HHO)for feature selection.Themotivation for utilizingGWOandHHOstems fromtheir bio-inspired nature and their demonstrated success in optimization problems.We aimto leverage the strengths of these algorithms to enhance the effectiveness of feature selection in microarray-based cancer classification.We selected leave-one-out cross-validation(LOOCV)to evaluate the performance of both two widely used classifiers,k-nearest neighbors(KNN)and support vector machine(SVM),on high-dimensional cancer microarray data.The proposed method is extensively tested on six publicly available cancer microarray datasets,and a comprehensive comparison with recently published methods is conducted.Our hybrid algorithm demonstrates its effectiveness in improving classification performance,Surpassing alternative approaches in terms of precision.The outcomes confirm the capability of our method to substantially improve both the precision and efficiency of cancer classification,thereby advancing the development ofmore efficient treatment strategies.The proposed hybridmethod offers a promising solution to the gene selection problem in microarray-based cancer classification.It improves the accuracy and efficiency of cancer diagnosis and treatment,and its superior performance compared to other methods highlights its potential applicability in realworld cancer classification tasks.By harnessing the complementary search mechanisms of GWO and HHO,we leverage their bio-inspired behavior to identify informative genes relevant to cancer diagnosis and treatment.
基金Supported by Hunan Province Traditional Chinese Medicine Research Project(No.B2023043)Hunan Provincial Department of Education Scientific Research Project(No.22B0386)Hunan University of Traditional Chinese Medicine Campus level Research Fund Project(No.2022XJZKC004).
文摘AIM:To develop a classifier for traditional Chinese medicine(TCM)syndrome differentiation of diabetic retinopathy(DR),using optimized machine learning algorithms,which can provide the basis for TCM objective and intelligent syndrome differentiation.METHODS:Collated data on real-world DR cases were collected.A variety of machine learning methods were used to construct TCM syndrome classification model,and the best performance was selected as the basic model.Genetic Algorithm(GA)was used for feature selection to obtain the optimal feature combination.Harris Hawk Optimization(HHO)was used for parameter optimization,and a classification model based on feature selection and parameter optimization was constructed.The performance of the model was compared with other optimization algorithms.The models were evaluated with accuracy,precision,recall,and F1 score as indicators.RESULTS:Data on 970 cases that met screening requirements were collected.Support Vector Machine(SVM)was the best basic classification model.The accuracy rate of the model was 82.05%,the precision rate was 82.34%,the recall rate was 81.81%,and the F1 value was 81.76%.After GA screening,the optimal feature combination contained 37 feature values,which was consistent with TCM clinical practice.The model based on optimal combination and SVM(GA_SVM)had an accuracy improvement of 1.92%compared to the basic classifier.SVM model based on HHO and GA optimization(HHO_GA_SVM)had the best performance and convergence speed compared with other optimization algorithms.Compared with the basic classification model,the accuracy was improved by 3.51%.CONCLUSION:HHO and GA optimization can improve the model performance of SVM in TCM syndrome differentiation of DR.It provides a new method and research idea for TCM intelligent assisted syndrome differentiation.
文摘Flexible job shop scheduling problem(FJSP)is the core decision-making problem of intelligent manufacturing production management.The Harris hawk optimization(HHO)algorithm,as a typical metaheuristic algorithm,has been widely employed to solve scheduling problems.However,HHO suffers from premature convergence when solving NP-hard problems.Therefore,this paper proposes an improved HHO algorithm(GNHHO)to solve the FJSP.GNHHO introduces an elitism strategy,a chaotic mechanism,a nonlinear escaping energy update strategy,and a Gaussian random walk strategy to prevent premature convergence.A flexible job shop scheduling model is constructed,and the static and dynamic FJSP is investigated to minimize the makespan.This paper chooses a two-segment encoding mode based on the job and the machine of the FJSP.To verify the effectiveness of GNHHO,this study tests it in 23 benchmark functions,10 standard job shop scheduling problems(JSPs),and 5 standard FJSPs.Besides,this study collects data from an agricultural company and uses the GNHHO algorithm to optimize the company’s FJSP.The optimized scheduling scheme demonstrates significant improvements in makespan,with an advancement of 28.16%for static scheduling and 35.63%for dynamic scheduling.Moreover,it achieves an average increase of 21.50%in the on-time order delivery rate.The results demonstrate that the performance of the GNHHO algorithm in solving FJSP is superior to some existing algorithms.
文摘In the study of Terrestrial Gamma-ray Flashes (TGFs) and Sonoluminescence, we observe parallels with larger cosmic events. Specifically, sonoluminescence involves the rapid collapse of bubbles, which closely resembles gravitational collapse in space. This observation suggests the potential formation of low-density quantum black holes. These entities, which might be related to dark matter, are thought to experience a kind of transient evaporation similar to Hawking radiation seen in cosmic black holes. Consequently, sonoluminescence could be a valuable tool for investigating phenomena typically linked to cosmic scale events. Furthermore, the role of the Higgs boson is considered in this context, possibly connecting it to both TGFs and sonoluminescence. This research could enhance our understanding of the quantum mechanics of black holes and their relation to dark matter on Earth.
文摘Using a rigorous mathematical approach, we demonstrate how the Cosmic Microwave Background (CMB) temperature could simply be a form of geometric mean temperature between the minimum time-dependent Hawking Hubble temperature and the maximum Planck temperature of the expanding universe over the course of cosmic time. This mathematical discovery suggests a re-consideration of Rh=ctcosmological models, including black hole cosmological models, even if it possibly could also be consistent with the Λ-CDM model. Most importantly, this paper contributes to the growing literature in the past year asserting a tightly constrained mathematical relationship between the CMB temperature, the Hubble constant, and other global parameters of the Hubble sphere. Our approach suggests a solid theoretical framework for predicting and understanding the CMB temperature rather than solely observing it.1.
文摘In real-world applications, datasets frequently contain outliers, which can hinder the generalization ability of machine learning models. Bayesian classifiers, a popular supervised learning method, rely on accurate probability density estimation for classifying continuous datasets. However, achieving precise density estimation with datasets containing outliers poses a significant challenge. This paper introduces a Bayesian classifier that utilizes optimized robust kernel density estimation to address this issue. Our proposed method enhances the accuracy of probability density distribution estimation by mitigating the impact of outliers on the training sample’s estimated distribution. Unlike the conventional kernel density estimator, our robust estimator can be seen as a weighted kernel mapping summary for each sample. This kernel mapping performs the inner product in the Hilbert space, allowing the kernel density estimation to be considered the average of the samples’ mapping in the Hilbert space using a reproducing kernel. M-estimation techniques are used to obtain accurate mean values and solve the weights. Meanwhile, complete cross-validation is used as the objective function to search for the optimal bandwidth, which impacts the estimator. The Harris Hawks Optimisation optimizes the objective function to improve the estimation accuracy. The experimental results show that it outperforms other optimization algorithms regarding convergence speed and objective function value during the bandwidth search. The optimal robust kernel density estimator achieves better fitness performance than the traditional kernel density estimator when the training data contains outliers. The Naïve Bayesian with optimal robust kernel density estimation improves the generalization in the classification with outliers.
文摘Aiming at the problems that the original Harris Hawk optimization algorithm is easy to fall into local optimum and slow in finding the optimum,this paper proposes an improved Harris Hawk optimization algorithm(GHHO).Firstly,we used a Gaussian chaotic mapping strategy to initialize the positions of individuals in the population,which enriches the initial individual species characteristics.Secondly,by optimizing the energy parameter and introducing the cosine strategy,the algorithm's ability to jump out of the local optimum is enhanced,which improves the performance of the algorithm.Finally,comparison experiments with other intelligent algorithms were conducted on 13 classical test function sets.The results show that GHHO has better performance in all aspects compared to other optimization algorithms.The improved algorithm is more suitable for generalization to real optimization problems.