Predictive analytics have been widely used in the literature with respect to laparoscopic surgery and risk stratification.However,most predictive analytics in this field exploit generalized linearmodels for predictive...Predictive analytics have been widely used in the literature with respect to laparoscopic surgery and risk stratification.However,most predictive analytics in this field exploit generalized linearmodels for predictive purposes,which are limited by model assumptionsdincluding linearity between response variables and additive interactions between variables.In many instances,such assumptions may not hold true,and the complex relationship between predictors and response variables is usually unknown.To address this limitation,machine-learning algorithms can be employed to model the underlying data.The advantage of machine learning algorithms is that they usually do not require strict assumptions regarding data structure,and they are able to learn complex functional forms using a nonparametric approach.Furthermore,two or more machine learning algorithms can be synthesized to further improve predictive accuracy.Such a process is referred to as ensemble modeling,and it has been used broadly in various industries.However,this approach has not been widely reported in the laparoscopic surgical literature due to its complexity in both model training and interpretation.With this technical note,we provide a comprehensive overview of the ensemble-modeling technique and a step-by-step tutorial on how to implement ensemble modeling.展开更多
Strong mechanical vibration and acoustical signals of grinding process contain useful information related to load parameters in ball mills. It is a challenge to extract latent features and construct soft sensor model ...Strong mechanical vibration and acoustical signals of grinding process contain useful information related to load parameters in ball mills. It is a challenge to extract latent features and construct soft sensor model with high dimensional frequency spectra of these signals. This paper aims to develop a selective ensemble modeling approach based on nonlinear latent frequency spectral feature extraction for accurate measurement of material to ball volume ratio. Latent features are first extracted from different vibrations and acoustic spectral segments by kernel partial least squares. Algorithms of bootstrap and least squares support vector machines are employed to produce candidate sub-models using these latent features as inputs. Ensemble sub-models are selected based on genetic algorithm optimization toolbox. Partial least squares regression is used to combine these sub-models to eliminate collinearity among their prediction outputs. Results indicate that the proposed modeling approach has better prediction performance than previous ones.展开更多
Machine learning(ML)and data mining are used in various fields such as data analysis,prediction,image processing and especially in healthcare.Researchers in the past decade have focused on applying ML and data mining ...Machine learning(ML)and data mining are used in various fields such as data analysis,prediction,image processing and especially in healthcare.Researchers in the past decade have focused on applying ML and data mining to generate conclusions from historical data in order to improve healthcare systems by making predictions about the results.Using ML algorithms,researchers have developed applications for decision support,analyzed clinical aspects,extracted informative information from historical data,predicted the outcomes and categorized diseases which help physicians make better decisions.It is observed that there is a huge difference between women depending on the region and their social lives.Due to these differences,scholars have been encouraged to conduct studies at a local level in order to better understand those factors that affect maternal health and the expected child.In this study,the ensemble modeling technique is applied to classify birth outcomes based on either cesarean section(C-Section)or normal delivery.A voting ensemble model for the classification of a birth dataset was made by using a Random Forest(RF),Gradient Boosting Classifier,Extra Trees Classifier and Bagging Classifier as base learners.It is observed that the voting ensemble modal of proposed classifiers provides the best accuracy,i.e.,94.78%,as compared to the individual classifiers.ML algorithms are more accurate due to ensemble models,which reduce variance and classification errors.It is reported that when a suitable classification model has been developed for birth classification,decision support systems can be created to enable clinicians to gain in-depth insights into the patterns in the datasets.Developing such a system will not only allow health organizations to improve maternal health assessment processes,but also open doors for interdisciplinary research in two different fields in the region.展开更多
Stomatopods are better known as mantis shrimp with considerable ecological importance in wide coastal waters globally. Some stomatopod species are exploited commercially, including Oratosquilla oratoria in the Northwe...Stomatopods are better known as mantis shrimp with considerable ecological importance in wide coastal waters globally. Some stomatopod species are exploited commercially, including Oratosquilla oratoria in the Northwest Pacific. Yet, few studies have published to promote accurate habitat identification of stomatopods, obstructing scientific management and conservation of these valuable organisms. This study provides an ensemble modeling framework for habitat suitability modeling of stomatopods, utilizing the O. oratoria stock in the Bohai Sea as an example. Two modeling techniques(i.e., generalized additive model(GAM) and geographical weighted regression(GWR)) were applied to select environmental predictors(especially the selection between two types of sediment metrics) that better characterize O. oratoria distribution and build separate habitat suitability models(HSM). The performance of the individual HSMs were compared on interpolation accuracy and transferability.Then, they were integrated to check whether the ensemble model outperforms either individual model, according to fishers’ knowledge and scientific survey data. As a result, grain-size metrics of sediment outperformed sediment content metrics in modeling O. oratoria habitat, possibly because grain-size metrics not only reflect the effect of substrates on burrow development, but also link to sediment heat capacity which influences individual thermoregulation. Moreover, the GWR-based HSM outperformed the GAM-based HSM in interpolation accuracy,while the latter one displayed better transferability. On balance, the ensemble HSM appeared to improve the predictive performance overall, as it could avoid dependence on a single model type and successfully identified fisher-recognized and survey-indicated suitable habitats in either sparsely sampled or well investigated areas.展开更多
Sentence classification is the process of categorizing a sentence based on the context of the sentence.Sentence categorization requires more semantic highlights than other tasks,such as dependence parsing,which requir...Sentence classification is the process of categorizing a sentence based on the context of the sentence.Sentence categorization requires more semantic highlights than other tasks,such as dependence parsing,which requires more syntactic elements.Most existing strategies focus on the general semantics of a conversation without involving the context of the sentence,recognizing the progress and comparing impacts.An ensemble pre-trained language model was taken up here to classify the conversation sentences from the conversation corpus.The conversational sentences are classified into four categories:information,question,directive,and commission.These classification label sequences are for analyzing the conversation progress and predicting the pecking order of the conversation.Ensemble of Bidirectional Encoder for Representation of Transformer(BERT),Robustly Optimized BERT pretraining Approach(RoBERTa),Generative Pre-Trained Transformer(GPT),DistilBERT and Generalized Autoregressive Pretraining for Language Understanding(XLNet)models are trained on conversation corpus with hyperparameters.Hyperparameter tuning approach is carried out for better performance on sentence classification.This Ensemble of Pre-trained Language Models with a Hyperparameter Tuning(EPLM-HT)system is trained on an annotated conversation dataset.The proposed approach outperformed compared to the base BERT,GPT,DistilBERT and XLNet transformer models.The proposed ensemble model with the fine-tuned parameters achieved an F1_score of 0.88.展开更多
Landslide is a serious natural disaster next only to earthquake and flood,which will cause a great threat to people’s lives and property safety.The traditional research of landslide disaster based on experience-drive...Landslide is a serious natural disaster next only to earthquake and flood,which will cause a great threat to people’s lives and property safety.The traditional research of landslide disaster based on experience-driven or statistical model and its assessment results are subjective,difficult to quantify,and no pertinence.As a new research method for landslide susceptibility assessment,machine learning can greatly improve the landslide susceptibility model’s accuracy by constructing statistical models.Taking Western Henan for example,the study selected 16 landslide influencing factors such as topography,geological environment,hydrological conditions,and human activities,and 11 landslide factors with the most significant influence on the landslide were selected by the recursive feature elimination(RFE)method.Five machine learning methods[Support Vector Machines(SVM),Logistic Regression(LR),Random Forest(RF),Extreme Gradient Boosting(XGBoost),and Linear Discriminant Analysis(LDA)]were used to construct the spatial distribution model of landslide susceptibility.The models were evaluated by the receiver operating characteristic curve and statistical index.After analysis and comparison,the XGBoost model(AUC 0.8759)performed the best and was suitable for dealing with regression problems.The model had a high adaptability to landslide data.According to the landslide susceptibility map of the five models,the overall distribution can be observed.The extremely high and high susceptibility areas are distributed in the Funiu Mountain range in the southwest,the Xiaoshan Mountain range in the west,and the Yellow River Basin in the north.These areas have large terrain fluctuations,complicated geological structural environments and frequent human engineering activities.The extremely high and highly prone areas were 12043.3 km^(2)and 3087.45 km^(2),accounting for 47.61%and 12.20%of the total area of the study area,respectively.Our study reflects the distribution of landslide susceptibility in western Henan Province,which provides a scientific basis for regional disaster warning,prediction,and resource protection.The study has important practical significance for subsequent landslide disaster management.展开更多
Contactless verification is possible with iris biometric identification,which helps prevent infections like COVID-19 from spreading.Biometric systems have grown unsteady and dangerous as a result of spoofing assaults ...Contactless verification is possible with iris biometric identification,which helps prevent infections like COVID-19 from spreading.Biometric systems have grown unsteady and dangerous as a result of spoofing assaults employing contact lenses,replayed the video,and print attacks.The work demonstrates an iris liveness detection approach by utilizing fragmental coefficients of Haar transformed Iris images as signatures to prevent spoofing attacks for the very first time in the identification of iris liveness.Seven assorted feature creation ways are studied in the presented solutions,and these created features are explored for the training of eight distinct machine learning classifiers and ensembles.The predicted iris liveness identification variants are evaluated using recall,F-measure,precision,accuracy,APCER,BPCER,and ACER.Three standard datasets were used in the investigation.The main contribution of our study is achieving a good accuracy of 99.18%with a smaller feature vector.The fragmental coefficients of Haar transformed iris image of size 8∗8 utilizing random forest algorithm showed superior iris liveness detection with reduced featured vector size(64 features).Random forest gave 99.18%accuracy.Additionally,conduct an extensive experiment on cross datasets for detailed analysis.The results of our experiments showthat the iris biometric template is decreased in size tomake the proposed framework suitable for algorithmic verification in real-time environments and settings.展开更多
Numerous factors affect the increased temperature of a machine tool, including prolonged and high-intensity usage,tool-workpiece interaction, mechanical friction, and elevated ambient temperatures, among others. Conse...Numerous factors affect the increased temperature of a machine tool, including prolonged and high-intensity usage,tool-workpiece interaction, mechanical friction, and elevated ambient temperatures, among others. Consequently,spindle thermal displacement occurs, and machining precision suffers. To prevent the errors caused by thetemperature rise of the Spindle fromaffecting the accuracy during themachining process, typically, the factory willwarm up themachine before themanufacturing process.However, if there is noway to understand the tool spindle’sthermal deformation, the machining quality will be greatly affected. In order to solve the above problem, thisstudy aims to predict the thermal displacement of the machine tool by using intelligent algorithms. In the practicalapplication, only a few temperature sensors are used to input the information into the prediction model for realtimethermal displacement prediction. This approach has greatly improved the quality of tool processing.However,each algorithm has different performances in different environments. In this study, an ensemble model is used tointegrate Long Short-TermMemory (LSTM) with Support VectorMachine (SVM). The experimental results showthat the prediction performance of LSTM-SVM is higher than that of other machine learning algorithms.展开更多
Breast cancer is one of the leading cancers among women.It has the second-highest mortality rate in women after lung cancer.Timely detection,especially in the early stages,can help increase survival rates.However,manu...Breast cancer is one of the leading cancers among women.It has the second-highest mortality rate in women after lung cancer.Timely detection,especially in the early stages,can help increase survival rates.However,manual diagnosis of breast cancer is a tedious and time-consuming process,and the accuracy of detection is reliant on the quality of the images and the radiologist’s experience.However,computer-aided medical diagnosis has recently shown promising results,leading to the need to develop an efficient system that can aid radiologists in diagnosing breast cancer in its early stages.The research presented in this paper is focused on the multi-class classification of breast cancer.The deep transfer learning approach has been utilized to train the deep learning models,and a pre-processing technique has been used to improve the quality of the ultrasound dataset.The proposed technique utilizes two deep learning models,Mobile-NetV2 and DenseNet201,for the composition of the deep ensemble model.Deep learning models are fine-tuned along with hyperparameter tuning to achieve better results.Subsequently,entropy-based feature selection is used.Breast cancer identification using the proposed classification approach was found to attain an accuracy of 97.04%,while the sensitivity and F1 score were 96.87%and 96.76%,respectively.The performance of the proposed model is very effective and outperforms other state-of-the-art techniques presented in the literature.展开更多
Stroke is a life-threatening disease usually due to blockage of blood or insufficient blood flow to the brain.It has a tremendous impact on every aspect of life since it is the leading global factor of disability and ...Stroke is a life-threatening disease usually due to blockage of blood or insufficient blood flow to the brain.It has a tremendous impact on every aspect of life since it is the leading global factor of disability and morbidity.Strokes can range from minor to severe(extensive).Thus,early stroke assessment and treatment can enhance survival rates.Manual prediction is extremely time and resource intensive.Automated prediction methods such as Modern Information and Communication Technologies(ICTs),particularly those inMachine Learning(ML)area,are crucial for the early diagnosis and prognosis of stroke.Therefore,this research proposed an ensemble voting model based on three Machine Learning(ML)algorithms:Random Forest(RF),Extreme Gradient Boosting(XGBoost),and Light Gradient Boosting Machine(LGBM).We apply data preprocessing to manage the outliers and useless instances in the dataset.Furthermore,to address the problem of imbalanced data,we enhance the minority class’s representation using the Synthetic Minority Over-Sampling Technique(SMOTE),allowing it to engage in the learning process actively.Results reveal that the suggested model outperforms existing studies and other classifiers with 0.96%accuracy,0.97%precision,0.97%recall,and 0.96%F1-score.The experiment demonstrates that the proposed ensemble voting model outperforms state-of-the-art and other traditional approaches.展开更多
In forecasting real time environmental factors,large data is needed to analyse the pattern behind the data values.Air pollution is a major threat towards developing countries and it is proliferating every year.Many me...In forecasting real time environmental factors,large data is needed to analyse the pattern behind the data values.Air pollution is a major threat towards developing countries and it is proliferating every year.Many methods in time ser-ies prediction and deep learning models to estimate the severity of air pollution.Each independent variable contributing towards pollution is necessary to analyse the trend behind the air pollution in that particular locality.This approach selects multivariate time series and coalesce a real time updatable autoregressive model to forecast Particulate matter(PM)PM2.5.To perform experimental analysis the data from the Central Pollution Control Board(CPCB)is used.Prediction is car-ried out for Chennai with seven locations and estimated PM’s using the weighted ensemble method.Proposed method for air pollution prediction unveiled effective and moored performance in long term prediction.Dynamic budge with high weighted k-models are used simultaneously and devising an ensemble helps to achieve stable forecasting.Computational time of ensemble decreases with paral-lel processing in each sub model.Weighted ensemble model shows high perfor-mance in long term prediction when compared to the traditional time series models like Vector Auto-Regression(VAR),Autoregressive Integrated with Mov-ing Average(ARIMA),Autoregressive Moving Average with Extended terms(ARMEX).Evaluation metrics like Root Mean Square Error(RMSE),Mean Absolute Error(MAE)and the time to achieve the time series are compared.展开更多
Covid-19 is a deadly virus that is rapidly spread around the world towards the end of the 2020.The consequences of this virus are quite frightening,especially when accompanied by an underlying disease.The novelty of t...Covid-19 is a deadly virus that is rapidly spread around the world towards the end of the 2020.The consequences of this virus are quite frightening,especially when accompanied by an underlying disease.The novelty of the virus,the constant emergence of different variants and its rapid spread have a negative impact on the control and treatment process.Although the new test kits provide almost certain results,chest X-rays are extremely important to detect the progression and degree of the disease.In addition to the Covid-19 virus,pneumonia and harmless opacity of the lungs also complicate the diagnosis.Considering the negative results caused by the virus and the treatment costs,the importance of fast and accurate diagnosis is clearly seen.In this context,deep learning methods appear as an extremely popular approach.In this study,a hybrid model design with superior properties of convolutional neural networks is presented to correctly classify the Covid-19 disease.In addition,in order to contribute to the literature,a suitable dataset with balanced case numbers that can be used in all artificial intelligence classification studies is presented.With this ensemble model design,quite remarkable results are obtained for the diagnosis of three and four-class Covid-19.The proposed model can classify normal,pneumonia,and Covid-19 with 92.6%accuracy and 82.6%for normal,pneumonia,Covid-19,and lung opacity.展开更多
Using the semiclassical ensemble model,the dependence of relative amplitude for the recollision dynamics in nonsequential double ionization(NSDI)of neon atom driven by the orthogonally polarized two-color field(OTC)la...Using the semiclassical ensemble model,the dependence of relative amplitude for the recollision dynamics in nonsequential double ionization(NSDI)of neon atom driven by the orthogonally polarized two-color field(OTC)laser field is theoretically studied.And the dynamics in two typical collision pathways,recollision-impact-ionization(RII)and recollisionexcitation with subsequent ionization(RESI),is systematically explored.Our results reveal that the V-shaped structure in the correlated momentum distribution is mainly caused by the RII mechanism when the relative amplitude of the OTC laser field is zero,and the first ionized electrons will quickly skim through the nucleus and share few energy with the second electron.As the relative amplitude increases,the V-shaped structure gradually disappears and electrons are concentrated on the diagonal in the electron correlation spectrum,indicating that the energy sharing after electrons collision is symmetric for OTC laser fields with large relative amplitudes.Our studies show that changing the relative amplitude of the OTC laser field can efficiently control the electron–electron collisions and energy exchange efficiency in the NSDI process.展开更多
Changing climate will jeopardize biodiversity,particularly the geographic distribution of endemic species.One such species is the Javan Hawk-Eagle(JHE,Nisaetus bartelsi),a charismatic raptor found only on Java Island,...Changing climate will jeopardize biodiversity,particularly the geographic distribution of endemic species.One such species is the Javan Hawk-Eagle(JHE,Nisaetus bartelsi),a charismatic raptor found only on Java Island,Indonesia.Thus,it is crucial to develop an appropriate conservation strategy to preserve the species.Ecological niche modeling is considered a valuable tool for designing conservation plans for the JHE.We provide an ecological niche modeling approach and transfer its model to future climate scenarios for the JHE.We utilize various machine learning algorithms under sustainability and business-as-usual(BAU)scenarios for 2050.Additionally,we investigate the conservation vulnerability of the JHE,capturing multifaceted pressures on the species from climate dissimilarities and human disturbance variables.Our study reveals that the ensemble model performs exceptionally well,with temperature emerging as the most critical factor affecting the JHE distribution.This finding indicates that climate change will have a significant impact on the JHE species.Our results suggest that the JHE distribution will likely decrease by 28.41%and 40.16%from the current JHE distribution under sustainability and BAU scenarios,respectively.Furthermore,our study reveals high-potential refugia for future JHE,covering 7,596 km^(2)(61%)under the sustainability scenario and only 4,403 km^(2)(35%)under the BAU scenario.Therefore,effective management and planning,including habitat restoration,refugia preservation,habitat connectivity,and local community inclusivity,should be well-managed to achieve JHE conservation targets.展开更多
Cold surges(CSs)often occur in the mid-latitude regions of the Northern Hemisphere and have enormous effects on socioeconomic development.We report that the occurrences of CSs and persistent CSs(PCSs)have rebounded si...Cold surges(CSs)often occur in the mid-latitude regions of the Northern Hemisphere and have enormous effects on socioeconomic development.We report that the occurrences of CSs and persistent CSs(PCSs)have rebounded since the 1990s,but the trends related to the frequencies of strong CSs(SCSs)and extreme CSs(ECSs)changed from increasing to decreasing after 2000.The highest-ranked model ensemble approach was used to project the occurrences of various CSs under the SSP1-2.6,SSP2-4.5,and SSP5-8.5 scenarios.The frequencies of the total CSs show overall decreasing trends.However,under the SSP1-2.6 scenario,slight increasing trends are noted for SCSs and ECSs in China.Atmospheric circulations that are characterized by an anomalous anticyclonic circulation with a significantly positive 500-hPa geopotential height(Z500)anomaly at high latitudes along with significant negative anomalies in China were favorable for cold air intrusions into China.In addition,the frequencies of all CS types under the SPP5-8.5 scenario greatly decreased in the long term(2071-2100),a finding which is thought to be related to negative SST anomalies in the central and western North Pacific,differences in sea level pressure(SLP)between high-and mid-latitude regions,and a weaker East Asian trough.In terms of ECSs,the decreasing trends observed during the historical period were maintained until 2024 under the SSP1-2.6 scenario.Compared to the SSP1-2.6 scenario,the Z500 pattern showed a trend of strengthened ridges over the Ural region and northern East Asia and weakened troughs over Siberia(60°-90°E)under the SSP2-4.5 and SSP5-8.5 scenarios,contributing to the shift to increasing trends of ECSs after 2014.展开更多
BACKGROUND There is a lack of literature discussing the utilization of the stacking ensemble algorithm for predicting depression in patients with heart failure(HF).AIM To create a stacking model for predicting depress...BACKGROUND There is a lack of literature discussing the utilization of the stacking ensemble algorithm for predicting depression in patients with heart failure(HF).AIM To create a stacking model for predicting depression in patients with HF.METHODS This study analyzed data on 1084 HF patients from the National Health and Nutrition Examination Survey database spanning from 2005 to 2018.Through univariate analysis and the use of an artificial neural network algorithm,predictors significantly linked to depression were identified.These predictors were utilized to create a stacking model employing tree-based learners.The performances of both the individual models and the stacking model were assessed by using the test dataset.Furthermore,the SHapley additive exPlanations(SHAP)model was applied to interpret the stacking model.RESULTS The models included five predictors.Among these models,the stacking model demonstrated the highest performance,achieving an area under the curve of 0.77(95%CI:0.71-0.84),a sensitivity of 0.71,and a specificity of 0.68.The calibration curve supported the reliability of the models,and decision curve analysis confirmed their clinical value.The SHAP plot demonstrated that age had the most significant impact on the stacking model's output.CONCLUSION The stacking model demonstrated strong predictive performance.Clinicians can utilize this model to identify highrisk depression patients with HF,thus enabling early provision of psychological interventions.展开更多
When a customer uses the software, then it is possible to occur defects that can be removed in the updated versions of the software. Hence, in the present work, a robust examination of cross-project software defect pr...When a customer uses the software, then it is possible to occur defects that can be removed in the updated versions of the software. Hence, in the present work, a robust examination of cross-project software defect prediction is elaborated through an innovative hybrid machine learning framework. The proposed technique combines an advanced deep neural network architecture with ensemble models such as Support Vector Machine (SVM), Random Forest (RF), and XGBoost. The study evaluates the performance by considering multiple software projects like CM1, JM1, KC1, and PC1 using datasets from the PROMISE Software Engineering Repository. The three hybrid models that are compared are Hybrid Model-1 (SVM, RandomForest, XGBoost, Neural Network), Hybrid Model-2 (GradientBoosting, DecisionTree, LogisticRegression, Neural Network), and Hybrid Model-3 (KNeighbors, GaussianNB, Support Vector Classification (SVC), Neural Network), and the Hybrid Model 3 surpasses the others in terms of recall, F1-score, accuracy, ROC AUC, and precision. The presented work offers valuable insights into the effectiveness of hybrid techniques for cross-project defect prediction, providing a comparative perspective on early defect identification and mitigation strategies. .展开更多
This study examines the performance of coupling the deterministic four-dimensional variational assimilation system (4DVAR) with an ensemble Kalman filter (EnKF) to produce a superior hybrid approach for data assim...This study examines the performance of coupling the deterministic four-dimensional variational assimilation system (4DVAR) with an ensemble Kalman filter (EnKF) to produce a superior hybrid approach for data assimilation. The coupled assimilation scheme (E4DVAR) benefits from using the state-dependent uncertainty provided by EnKF while taking advantage of 4DVAR in preventing filter divergence: the 4DVAR analysis produces posterior maximum likelihood solutions through minimization of a cost function about which the ensemble perturbations are transformed, and the resulting ensemble analysis can be propagated forward both for the next assimilation cycle and as a basis for ensemble forecasting. The feasibility and effectiveness of this coupled approach are demonstrated in an idealized model with simulated observations. It is found that the E4DVAR is capable of outperforming both 4DVAR and the EnKF under both perfect- and imperfect-model scenarios. The performance of the coupled scheme is also less sensitive to either the ensemble size or the assimilation window length than those for standard EnKF or 4DVAR implementations.展开更多
This study implements a hybrid ensemble machine learning method for forecasting the rate of penetration(ROP) of tunnel boring machine(TBM),which is becoming a prerequisite for reliable cost assessment and project sche...This study implements a hybrid ensemble machine learning method for forecasting the rate of penetration(ROP) of tunnel boring machine(TBM),which is becoming a prerequisite for reliable cost assessment and project scheduling in tunnelling and underground projects in a rock environment.For this purpose,a sum of 185 datasets was collected from the literature and used to predict the ROP of TBM.Initially,the main dataset was utilised to construct and validate four conventional soft computing(CSC)models,i.e.minimax probability machine regression,relevance vector machine,extreme learning machine,and functional network.Consequently,the estimated outputs of CSC models were united and trained using an artificial neural network(ANN) to construct a hybrid ensemble model(HENSM).The outcomes of the proposed HENSM are superior to other CSC models employed in this study.Based on the experimental results(training RMSE=0.0283 and testing RMSE=0.0418),the newly proposed HENSM is potential to assist engineers in predicting ROP of TBM in the design phase of tunnelling and underground projects.展开更多
Metamaterial Antenna is a special class of antennas that uses metamaterial to enhance their performance.Antenna size affects the quality factor and the radiation loss of the antenna.Metamaterial antennas can overcome ...Metamaterial Antenna is a special class of antennas that uses metamaterial to enhance their performance.Antenna size affects the quality factor and the radiation loss of the antenna.Metamaterial antennas can overcome the limitation of bandwidth for small antennas.Machine learning(ML)model is recently applied to predict antenna parameters.ML can be used as an alternative approach to the trial-and-error process of finding proper parameters of the simulated antenna.The accuracy of the prediction depends mainly on the selected model.Ensemble models combine two or more base models to produce a better-enhanced model.In this paper,a weighted average ensemble model is proposed to predict the bandwidth of the Metamaterial Antenna.Two base models are used namely:Multilayer Perceptron(MLP)and Support Vector Machines(SVM).To calculate the weights for each model,an optimization algorithm is used to find the optimal weights of the ensemble.Dynamic Group-Based Cooperative Optimizer(DGCO)is employed to search for optimal weight for the base models.The proposed model is compared with three based models and the average ensemble model.The results show that the proposed model is better than other models and can predict antenna bandwidth efficiently.展开更多
基金funding from RUIYI emergency medical research fund(202013)Open Foundation of Artificial Intelligence Key Laboratory of Sichuan Province(2020RYY03)+1 种基金Research project of Health and Family Planning Commission of Sichuan Province(17PJ136)funding from Key Research&Development project of Zhejiang Province(2021C03071).
文摘Predictive analytics have been widely used in the literature with respect to laparoscopic surgery and risk stratification.However,most predictive analytics in this field exploit generalized linearmodels for predictive purposes,which are limited by model assumptionsdincluding linearity between response variables and additive interactions between variables.In many instances,such assumptions may not hold true,and the complex relationship between predictors and response variables is usually unknown.To address this limitation,machine-learning algorithms can be employed to model the underlying data.The advantage of machine learning algorithms is that they usually do not require strict assumptions regarding data structure,and they are able to learn complex functional forms using a nonparametric approach.Furthermore,two or more machine learning algorithms can be synthesized to further improve predictive accuracy.Such a process is referred to as ensemble modeling,and it has been used broadly in various industries.However,this approach has not been widely reported in the laparoscopic surgical literature due to its complexity in both model training and interpretation.With this technical note,we provide a comprehensive overview of the ensemble-modeling technique and a step-by-step tutorial on how to implement ensemble modeling.
基金Supported partially by the Post Doctoral Natural Science Foundation of China(2013M532118,2015T81082)the National Natural Science Foundation of China(61573364,61273177,61503066)+2 种基金the State Key Laboratory of Synthetical Automation for Process Industriesthe National High Technology Research and Development Program of China(2015AA043802)the Scientific Research Fund of Liaoning Provincial Education Department(L2013272)
文摘Strong mechanical vibration and acoustical signals of grinding process contain useful information related to load parameters in ball mills. It is a challenge to extract latent features and construct soft sensor model with high dimensional frequency spectra of these signals. This paper aims to develop a selective ensemble modeling approach based on nonlinear latent frequency spectral feature extraction for accurate measurement of material to ball volume ratio. Latent features are first extracted from different vibrations and acoustic spectral segments by kernel partial least squares. Algorithms of bootstrap and least squares support vector machines are employed to produce candidate sub-models using these latent features as inputs. Ensemble sub-models are selected based on genetic algorithm optimization toolbox. Partial least squares regression is used to combine these sub-models to eliminate collinearity among their prediction outputs. Results indicate that the proposed modeling approach has better prediction performance than previous ones.
基金Natural Sciences and Engineering Research Council of Canada(NSERC)and New Brunswick Innovation Foundation(NBIF)for the financial support of the global project.These granting agencies did not contribute in the design of the study and collection,analysis,and interpretation of data。
文摘Machine learning(ML)and data mining are used in various fields such as data analysis,prediction,image processing and especially in healthcare.Researchers in the past decade have focused on applying ML and data mining to generate conclusions from historical data in order to improve healthcare systems by making predictions about the results.Using ML algorithms,researchers have developed applications for decision support,analyzed clinical aspects,extracted informative information from historical data,predicted the outcomes and categorized diseases which help physicians make better decisions.It is observed that there is a huge difference between women depending on the region and their social lives.Due to these differences,scholars have been encouraged to conduct studies at a local level in order to better understand those factors that affect maternal health and the expected child.In this study,the ensemble modeling technique is applied to classify birth outcomes based on either cesarean section(C-Section)or normal delivery.A voting ensemble model for the classification of a birth dataset was made by using a Random Forest(RF),Gradient Boosting Classifier,Extra Trees Classifier and Bagging Classifier as base learners.It is observed that the voting ensemble modal of proposed classifiers provides the best accuracy,i.e.,94.78%,as compared to the individual classifiers.ML algorithms are more accurate due to ensemble models,which reduce variance and classification errors.It is reported that when a suitable classification model has been developed for birth classification,decision support systems can be created to enable clinicians to gain in-depth insights into the patterns in the datasets.Developing such a system will not only allow health organizations to improve maternal health assessment processes,but also open doors for interdisciplinary research in two different fields in the region.
基金The National Natural Science Foundation of China under contract No.31902375the David and Lucile Packard Foundation+1 种基金the Innovation Team of Fishery Resources and Ecology in the Yellow Sea and Bohai Sea under contract No.2020TD01the Special Funds for Taishan Scholars Project of Shandong Province。
文摘Stomatopods are better known as mantis shrimp with considerable ecological importance in wide coastal waters globally. Some stomatopod species are exploited commercially, including Oratosquilla oratoria in the Northwest Pacific. Yet, few studies have published to promote accurate habitat identification of stomatopods, obstructing scientific management and conservation of these valuable organisms. This study provides an ensemble modeling framework for habitat suitability modeling of stomatopods, utilizing the O. oratoria stock in the Bohai Sea as an example. Two modeling techniques(i.e., generalized additive model(GAM) and geographical weighted regression(GWR)) were applied to select environmental predictors(especially the selection between two types of sediment metrics) that better characterize O. oratoria distribution and build separate habitat suitability models(HSM). The performance of the individual HSMs were compared on interpolation accuracy and transferability.Then, they were integrated to check whether the ensemble model outperforms either individual model, according to fishers’ knowledge and scientific survey data. As a result, grain-size metrics of sediment outperformed sediment content metrics in modeling O. oratoria habitat, possibly because grain-size metrics not only reflect the effect of substrates on burrow development, but also link to sediment heat capacity which influences individual thermoregulation. Moreover, the GWR-based HSM outperformed the GAM-based HSM in interpolation accuracy,while the latter one displayed better transferability. On balance, the ensemble HSM appeared to improve the predictive performance overall, as it could avoid dependence on a single model type and successfully identified fisher-recognized and survey-indicated suitable habitats in either sparsely sampled or well investigated areas.
文摘Sentence classification is the process of categorizing a sentence based on the context of the sentence.Sentence categorization requires more semantic highlights than other tasks,such as dependence parsing,which requires more syntactic elements.Most existing strategies focus on the general semantics of a conversation without involving the context of the sentence,recognizing the progress and comparing impacts.An ensemble pre-trained language model was taken up here to classify the conversation sentences from the conversation corpus.The conversational sentences are classified into four categories:information,question,directive,and commission.These classification label sequences are for analyzing the conversation progress and predicting the pecking order of the conversation.Ensemble of Bidirectional Encoder for Representation of Transformer(BERT),Robustly Optimized BERT pretraining Approach(RoBERTa),Generative Pre-Trained Transformer(GPT),DistilBERT and Generalized Autoregressive Pretraining for Language Understanding(XLNet)models are trained on conversation corpus with hyperparameters.Hyperparameter tuning approach is carried out for better performance on sentence classification.This Ensemble of Pre-trained Language Models with a Hyperparameter Tuning(EPLM-HT)system is trained on an annotated conversation dataset.The proposed approach outperformed compared to the base BERT,GPT,DistilBERT and XLNet transformer models.The proposed ensemble model with the fine-tuned parameters achieved an F1_score of 0.88.
基金This work was financially supported by National Natural Science Foundation of China(41972262)Hebei Natural Science Foundation for Excellent Young Scholars(D2020504032)+1 种基金Central Plains Science and technology innovation leader Project(214200510030)Key research and development Project of Henan province(221111321500).
文摘Landslide is a serious natural disaster next only to earthquake and flood,which will cause a great threat to people’s lives and property safety.The traditional research of landslide disaster based on experience-driven or statistical model and its assessment results are subjective,difficult to quantify,and no pertinence.As a new research method for landslide susceptibility assessment,machine learning can greatly improve the landslide susceptibility model’s accuracy by constructing statistical models.Taking Western Henan for example,the study selected 16 landslide influencing factors such as topography,geological environment,hydrological conditions,and human activities,and 11 landslide factors with the most significant influence on the landslide were selected by the recursive feature elimination(RFE)method.Five machine learning methods[Support Vector Machines(SVM),Logistic Regression(LR),Random Forest(RF),Extreme Gradient Boosting(XGBoost),and Linear Discriminant Analysis(LDA)]were used to construct the spatial distribution model of landslide susceptibility.The models were evaluated by the receiver operating characteristic curve and statistical index.After analysis and comparison,the XGBoost model(AUC 0.8759)performed the best and was suitable for dealing with regression problems.The model had a high adaptability to landslide data.According to the landslide susceptibility map of the five models,the overall distribution can be observed.The extremely high and high susceptibility areas are distributed in the Funiu Mountain range in the southwest,the Xiaoshan Mountain range in the west,and the Yellow River Basin in the north.These areas have large terrain fluctuations,complicated geological structural environments and frequent human engineering activities.The extremely high and highly prone areas were 12043.3 km^(2)and 3087.45 km^(2),accounting for 47.61%and 12.20%of the total area of the study area,respectively.Our study reflects the distribution of landslide susceptibility in western Henan Province,which provides a scientific basis for regional disaster warning,prediction,and resource protection.The study has important practical significance for subsequent landslide disaster management.
基金supported by theResearchers Supporting Project No.RSP-2021/14,King Saud University,Riyadh,Saudi Arabia.
文摘Contactless verification is possible with iris biometric identification,which helps prevent infections like COVID-19 from spreading.Biometric systems have grown unsteady and dangerous as a result of spoofing assaults employing contact lenses,replayed the video,and print attacks.The work demonstrates an iris liveness detection approach by utilizing fragmental coefficients of Haar transformed Iris images as signatures to prevent spoofing attacks for the very first time in the identification of iris liveness.Seven assorted feature creation ways are studied in the presented solutions,and these created features are explored for the training of eight distinct machine learning classifiers and ensembles.The predicted iris liveness identification variants are evaluated using recall,F-measure,precision,accuracy,APCER,BPCER,and ACER.Three standard datasets were used in the investigation.The main contribution of our study is achieving a good accuracy of 99.18%with a smaller feature vector.The fragmental coefficients of Haar transformed iris image of size 8∗8 utilizing random forest algorithm showed superior iris liveness detection with reduced featured vector size(64 features).Random forest gave 99.18%accuracy.Additionally,conduct an extensive experiment on cross datasets for detailed analysis.The results of our experiments showthat the iris biometric template is decreased in size tomake the proposed framework suitable for algorithmic verification in real-time environments and settings.
基金supported by the Ministry of Science and Technology,Taiwan,under Grant MOST 110-2218-E-194-010。
文摘Numerous factors affect the increased temperature of a machine tool, including prolonged and high-intensity usage,tool-workpiece interaction, mechanical friction, and elevated ambient temperatures, among others. Consequently,spindle thermal displacement occurs, and machining precision suffers. To prevent the errors caused by thetemperature rise of the Spindle fromaffecting the accuracy during themachining process, typically, the factory willwarm up themachine before themanufacturing process.However, if there is noway to understand the tool spindle’sthermal deformation, the machining quality will be greatly affected. In order to solve the above problem, thisstudy aims to predict the thermal displacement of the machine tool by using intelligent algorithms. In the practicalapplication, only a few temperature sensors are used to input the information into the prediction model for realtimethermal displacement prediction. This approach has greatly improved the quality of tool processing.However,each algorithm has different performances in different environments. In this study, an ensemble model is used tointegrate Long Short-TermMemory (LSTM) with Support VectorMachine (SVM). The experimental results showthat the prediction performance of LSTM-SVM is higher than that of other machine learning algorithms.
基金This research work was funded by Institutional Fund Projects under Grant No.(IFPIP:1614-611-1442)from the Ministry of Education and King Abdulaziz University,DSR,Jeddah,Saudi Arabia.
文摘Breast cancer is one of the leading cancers among women.It has the second-highest mortality rate in women after lung cancer.Timely detection,especially in the early stages,can help increase survival rates.However,manual diagnosis of breast cancer is a tedious and time-consuming process,and the accuracy of detection is reliant on the quality of the images and the radiologist’s experience.However,computer-aided medical diagnosis has recently shown promising results,leading to the need to develop an efficient system that can aid radiologists in diagnosing breast cancer in its early stages.The research presented in this paper is focused on the multi-class classification of breast cancer.The deep transfer learning approach has been utilized to train the deep learning models,and a pre-processing technique has been used to improve the quality of the ultrasound dataset.The proposed technique utilizes two deep learning models,Mobile-NetV2 and DenseNet201,for the composition of the deep ensemble model.Deep learning models are fine-tuned along with hyperparameter tuning to achieve better results.Subsequently,entropy-based feature selection is used.Breast cancer identification using the proposed classification approach was found to attain an accuracy of 97.04%,while the sensitivity and F1 score were 96.87%and 96.76%,respectively.The performance of the proposed model is very effective and outperforms other state-of-the-art techniques presented in the literature.
文摘Stroke is a life-threatening disease usually due to blockage of blood or insufficient blood flow to the brain.It has a tremendous impact on every aspect of life since it is the leading global factor of disability and morbidity.Strokes can range from minor to severe(extensive).Thus,early stroke assessment and treatment can enhance survival rates.Manual prediction is extremely time and resource intensive.Automated prediction methods such as Modern Information and Communication Technologies(ICTs),particularly those inMachine Learning(ML)area,are crucial for the early diagnosis and prognosis of stroke.Therefore,this research proposed an ensemble voting model based on three Machine Learning(ML)algorithms:Random Forest(RF),Extreme Gradient Boosting(XGBoost),and Light Gradient Boosting Machine(LGBM).We apply data preprocessing to manage the outliers and useless instances in the dataset.Furthermore,to address the problem of imbalanced data,we enhance the minority class’s representation using the Synthetic Minority Over-Sampling Technique(SMOTE),allowing it to engage in the learning process actively.Results reveal that the suggested model outperforms existing studies and other classifiers with 0.96%accuracy,0.97%precision,0.97%recall,and 0.96%F1-score.The experiment demonstrates that the proposed ensemble voting model outperforms state-of-the-art and other traditional approaches.
文摘In forecasting real time environmental factors,large data is needed to analyse the pattern behind the data values.Air pollution is a major threat towards developing countries and it is proliferating every year.Many methods in time ser-ies prediction and deep learning models to estimate the severity of air pollution.Each independent variable contributing towards pollution is necessary to analyse the trend behind the air pollution in that particular locality.This approach selects multivariate time series and coalesce a real time updatable autoregressive model to forecast Particulate matter(PM)PM2.5.To perform experimental analysis the data from the Central Pollution Control Board(CPCB)is used.Prediction is car-ried out for Chennai with seven locations and estimated PM’s using the weighted ensemble method.Proposed method for air pollution prediction unveiled effective and moored performance in long term prediction.Dynamic budge with high weighted k-models are used simultaneously and devising an ensemble helps to achieve stable forecasting.Computational time of ensemble decreases with paral-lel processing in each sub model.Weighted ensemble model shows high perfor-mance in long term prediction when compared to the traditional time series models like Vector Auto-Regression(VAR),Autoregressive Integrated with Mov-ing Average(ARIMA),Autoregressive Moving Average with Extended terms(ARMEX).Evaluation metrics like Root Mean Square Error(RMSE),Mean Absolute Error(MAE)and the time to achieve the time series are compared.
文摘Covid-19 is a deadly virus that is rapidly spread around the world towards the end of the 2020.The consequences of this virus are quite frightening,especially when accompanied by an underlying disease.The novelty of the virus,the constant emergence of different variants and its rapid spread have a negative impact on the control and treatment process.Although the new test kits provide almost certain results,chest X-rays are extremely important to detect the progression and degree of the disease.In addition to the Covid-19 virus,pneumonia and harmless opacity of the lungs also complicate the diagnosis.Considering the negative results caused by the virus and the treatment costs,the importance of fast and accurate diagnosis is clearly seen.In this context,deep learning methods appear as an extremely popular approach.In this study,a hybrid model design with superior properties of convolutional neural networks is presented to correctly classify the Covid-19 disease.In addition,in order to contribute to the literature,a suitable dataset with balanced case numbers that can be used in all artificial intelligence classification studies is presented.With this ensemble model design,quite remarkable results are obtained for the diagnosis of three and four-class Covid-19.The proposed model can classify normal,pneumonia,and Covid-19 with 92.6%accuracy and 82.6%for normal,pneumonia,Covid-19,and lung opacity.
基金Project supported by the National Natural Science Foundation of China (Grant Nos.12204132 and 12304376)Excellent Youth Science Foundation of Shandong Province (Overseas) (Grant No.2022HWYQ-073)+1 种基金the Fundamental Research Funds for the Central Universities (Grant No.HIT.OCEF.2022042)Natural Science Foundation of Shandong Province (Grant No.ZR2023QA075)。
文摘Using the semiclassical ensemble model,the dependence of relative amplitude for the recollision dynamics in nonsequential double ionization(NSDI)of neon atom driven by the orthogonally polarized two-color field(OTC)laser field is theoretically studied.And the dynamics in two typical collision pathways,recollision-impact-ionization(RII)and recollisionexcitation with subsequent ionization(RESI),is systematically explored.Our results reveal that the V-shaped structure in the correlated momentum distribution is mainly caused by the RII mechanism when the relative amplitude of the OTC laser field is zero,and the first ionized electrons will quickly skim through the nucleus and share few energy with the second electron.As the relative amplitude increases,the V-shaped structure gradually disappears and electrons are concentrated on the diagonal in the electron correlation spectrum,indicating that the energy sharing after electrons collision is symmetric for OTC laser fields with large relative amplitudes.Our studies show that changing the relative amplitude of the OTC laser field can efficiently control the electron–electron collisions and energy exchange efficiency in the NSDI process.
文摘Changing climate will jeopardize biodiversity,particularly the geographic distribution of endemic species.One such species is the Javan Hawk-Eagle(JHE,Nisaetus bartelsi),a charismatic raptor found only on Java Island,Indonesia.Thus,it is crucial to develop an appropriate conservation strategy to preserve the species.Ecological niche modeling is considered a valuable tool for designing conservation plans for the JHE.We provide an ecological niche modeling approach and transfer its model to future climate scenarios for the JHE.We utilize various machine learning algorithms under sustainability and business-as-usual(BAU)scenarios for 2050.Additionally,we investigate the conservation vulnerability of the JHE,capturing multifaceted pressures on the species from climate dissimilarities and human disturbance variables.Our study reveals that the ensemble model performs exceptionally well,with temperature emerging as the most critical factor affecting the JHE distribution.This finding indicates that climate change will have a significant impact on the JHE species.Our results suggest that the JHE distribution will likely decrease by 28.41%and 40.16%from the current JHE distribution under sustainability and BAU scenarios,respectively.Furthermore,our study reveals high-potential refugia for future JHE,covering 7,596 km^(2)(61%)under the sustainability scenario and only 4,403 km^(2)(35%)under the BAU scenario.Therefore,effective management and planning,including habitat restoration,refugia preservation,habitat connectivity,and local community inclusivity,should be well-managed to achieve JHE conservation targets.
基金supported by the National Key R&D Program of China(Grant No.2022YFF0801703)the State Key Laboratory of Earth Surface Processes and Resource Ecology(Grant No.2022-GS-01)in China.
文摘Cold surges(CSs)often occur in the mid-latitude regions of the Northern Hemisphere and have enormous effects on socioeconomic development.We report that the occurrences of CSs and persistent CSs(PCSs)have rebounded since the 1990s,but the trends related to the frequencies of strong CSs(SCSs)and extreme CSs(ECSs)changed from increasing to decreasing after 2000.The highest-ranked model ensemble approach was used to project the occurrences of various CSs under the SSP1-2.6,SSP2-4.5,and SSP5-8.5 scenarios.The frequencies of the total CSs show overall decreasing trends.However,under the SSP1-2.6 scenario,slight increasing trends are noted for SCSs and ECSs in China.Atmospheric circulations that are characterized by an anomalous anticyclonic circulation with a significantly positive 500-hPa geopotential height(Z500)anomaly at high latitudes along with significant negative anomalies in China were favorable for cold air intrusions into China.In addition,the frequencies of all CS types under the SPP5-8.5 scenario greatly decreased in the long term(2071-2100),a finding which is thought to be related to negative SST anomalies in the central and western North Pacific,differences in sea level pressure(SLP)between high-and mid-latitude regions,and a weaker East Asian trough.In terms of ECSs,the decreasing trends observed during the historical period were maintained until 2024 under the SSP1-2.6 scenario.Compared to the SSP1-2.6 scenario,the Z500 pattern showed a trend of strengthened ridges over the Ural region and northern East Asia and weakened troughs over Siberia(60°-90°E)under the SSP2-4.5 and SSP5-8.5 scenarios,contributing to the shift to increasing trends of ECSs after 2014.
文摘BACKGROUND There is a lack of literature discussing the utilization of the stacking ensemble algorithm for predicting depression in patients with heart failure(HF).AIM To create a stacking model for predicting depression in patients with HF.METHODS This study analyzed data on 1084 HF patients from the National Health and Nutrition Examination Survey database spanning from 2005 to 2018.Through univariate analysis and the use of an artificial neural network algorithm,predictors significantly linked to depression were identified.These predictors were utilized to create a stacking model employing tree-based learners.The performances of both the individual models and the stacking model were assessed by using the test dataset.Furthermore,the SHapley additive exPlanations(SHAP)model was applied to interpret the stacking model.RESULTS The models included five predictors.Among these models,the stacking model demonstrated the highest performance,achieving an area under the curve of 0.77(95%CI:0.71-0.84),a sensitivity of 0.71,and a specificity of 0.68.The calibration curve supported the reliability of the models,and decision curve analysis confirmed their clinical value.The SHAP plot demonstrated that age had the most significant impact on the stacking model's output.CONCLUSION The stacking model demonstrated strong predictive performance.Clinicians can utilize this model to identify highrisk depression patients with HF,thus enabling early provision of psychological interventions.
文摘When a customer uses the software, then it is possible to occur defects that can be removed in the updated versions of the software. Hence, in the present work, a robust examination of cross-project software defect prediction is elaborated through an innovative hybrid machine learning framework. The proposed technique combines an advanced deep neural network architecture with ensemble models such as Support Vector Machine (SVM), Random Forest (RF), and XGBoost. The study evaluates the performance by considering multiple software projects like CM1, JM1, KC1, and PC1 using datasets from the PROMISE Software Engineering Repository. The three hybrid models that are compared are Hybrid Model-1 (SVM, RandomForest, XGBoost, Neural Network), Hybrid Model-2 (GradientBoosting, DecisionTree, LogisticRegression, Neural Network), and Hybrid Model-3 (KNeighbors, GaussianNB, Support Vector Classification (SVC), Neural Network), and the Hybrid Model 3 surpasses the others in terms of recall, F1-score, accuracy, ROC AUC, and precision. The presented work offers valuable insights into the effectiveness of hybrid techniques for cross-project defect prediction, providing a comparative perspective on early defect identification and mitigation strategies. .
基金sponsored by the U.S. National Science Foundation (Grant No.ATM0205599)the U.S. Offce of Navy Research under Grant N000140410471Dr. James A. Hansen was partially supported by US Offce of Naval Research (Grant No. N00014-06-1-0500)
文摘This study examines the performance of coupling the deterministic four-dimensional variational assimilation system (4DVAR) with an ensemble Kalman filter (EnKF) to produce a superior hybrid approach for data assimilation. The coupled assimilation scheme (E4DVAR) benefits from using the state-dependent uncertainty provided by EnKF while taking advantage of 4DVAR in preventing filter divergence: the 4DVAR analysis produces posterior maximum likelihood solutions through minimization of a cost function about which the ensemble perturbations are transformed, and the resulting ensemble analysis can be propagated forward both for the next assimilation cycle and as a basis for ensemble forecasting. The feasibility and effectiveness of this coupled approach are demonstrated in an idealized model with simulated observations. It is found that the E4DVAR is capable of outperforming both 4DVAR and the EnKF under both perfect- and imperfect-model scenarios. The performance of the coupled scheme is also less sensitive to either the ensemble size or the assimilation window length than those for standard EnKF or 4DVAR implementations.
文摘This study implements a hybrid ensemble machine learning method for forecasting the rate of penetration(ROP) of tunnel boring machine(TBM),which is becoming a prerequisite for reliable cost assessment and project scheduling in tunnelling and underground projects in a rock environment.For this purpose,a sum of 185 datasets was collected from the literature and used to predict the ROP of TBM.Initially,the main dataset was utilised to construct and validate four conventional soft computing(CSC)models,i.e.minimax probability machine regression,relevance vector machine,extreme learning machine,and functional network.Consequently,the estimated outputs of CSC models were united and trained using an artificial neural network(ANN) to construct a hybrid ensemble model(HENSM).The outcomes of the proposed HENSM are superior to other CSC models employed in this study.Based on the experimental results(training RMSE=0.0283 and testing RMSE=0.0418),the newly proposed HENSM is potential to assist engineers in predicting ROP of TBM in the design phase of tunnelling and underground projects.
文摘Metamaterial Antenna is a special class of antennas that uses metamaterial to enhance their performance.Antenna size affects the quality factor and the radiation loss of the antenna.Metamaterial antennas can overcome the limitation of bandwidth for small antennas.Machine learning(ML)model is recently applied to predict antenna parameters.ML can be used as an alternative approach to the trial-and-error process of finding proper parameters of the simulated antenna.The accuracy of the prediction depends mainly on the selected model.Ensemble models combine two or more base models to produce a better-enhanced model.In this paper,a weighted average ensemble model is proposed to predict the bandwidth of the Metamaterial Antenna.Two base models are used namely:Multilayer Perceptron(MLP)and Support Vector Machines(SVM).To calculate the weights for each model,an optimization algorithm is used to find the optimal weights of the ensemble.Dynamic Group-Based Cooperative Optimizer(DGCO)is employed to search for optimal weight for the base models.The proposed model is compared with three based models and the average ensemble model.The results show that the proposed model is better than other models and can predict antenna bandwidth efficiently.