The Indian Himalayan region is frequently experiencing climate change-induced landslides.Thus,landslide susceptibility assessment assumes greater significance for lessening the impact of a landslide hazard.This paper ...The Indian Himalayan region is frequently experiencing climate change-induced landslides.Thus,landslide susceptibility assessment assumes greater significance for lessening the impact of a landslide hazard.This paper makes an attempt to assess landslide susceptibility in Shimla district of the northwest Indian Himalayan region.It examined the effectiveness of random forest(RF),multilayer perceptron(MLP),sequential minimal optimization regression(SMOreg)and bagging ensemble(B-RF,BSMOreg,B-MLP)models.A landslide inventory map comprising 1052 locations of past landslide occurrences was classified into training(70%)and testing(30%)datasets.The site-specific influencing factors were selected by employing a multicollinearity test.The relationship between past landslide occurrences and influencing factors was established using the frequency ratio method.The effectiveness of machine learning models was verified through performance assessors.The landslide susceptibility maps were validated by the area under the receiver operating characteristic curves(ROC-AUC),accuracy,precision,recall and F1-score.The key performance metrics and map validation demonstrated that the BRF model(correlation coefficient:0.988,mean absolute error:0.010,root mean square error:0.058,relative absolute error:2.964,ROC-AUC:0.947,accuracy:0.778,precision:0.819,recall:0.917 and F-1 score:0.865)outperformed the single classifiers and other bagging ensemble models for landslide susceptibility.The results show that the largest area was found under the very high susceptibility zone(33.87%),followed by the low(27.30%),high(20.68%)and moderate(18.16%)susceptibility zones.The factors,namely average annual rainfall,slope,lithology,soil texture and earthquake magnitude have been identified as the influencing factors for very high landslide susceptibility.Soil texture,lineament density and elevation have been attributed to high and moderate susceptibility.Thus,the study calls for devising suitable landslide mitigation measures in the study area.Structural measures,an immediate response system,community participation and coordination among stakeholders may help lessen the detrimental impact of landslides.The findings from this study could aid decision-makers in mitigating future catastrophes and devising suitable strategies in other geographical regions with similar geological characteristics.展开更多
Sentence classification is the process of categorizing a sentence based on the context of the sentence.Sentence categorization requires more semantic highlights than other tasks,such as dependence parsing,which requir...Sentence classification is the process of categorizing a sentence based on the context of the sentence.Sentence categorization requires more semantic highlights than other tasks,such as dependence parsing,which requires more syntactic elements.Most existing strategies focus on the general semantics of a conversation without involving the context of the sentence,recognizing the progress and comparing impacts.An ensemble pre-trained language model was taken up here to classify the conversation sentences from the conversation corpus.The conversational sentences are classified into four categories:information,question,directive,and commission.These classification label sequences are for analyzing the conversation progress and predicting the pecking order of the conversation.Ensemble of Bidirectional Encoder for Representation of Transformer(BERT),Robustly Optimized BERT pretraining Approach(RoBERTa),Generative Pre-Trained Transformer(GPT),DistilBERT and Generalized Autoregressive Pretraining for Language Understanding(XLNet)models are trained on conversation corpus with hyperparameters.Hyperparameter tuning approach is carried out for better performance on sentence classification.This Ensemble of Pre-trained Language Models with a Hyperparameter Tuning(EPLM-HT)system is trained on an annotated conversation dataset.The proposed approach outperformed compared to the base BERT,GPT,DistilBERT and XLNet transformer models.The proposed ensemble model with the fine-tuned parameters achieved an F1_score of 0.88.展开更多
Stomatopods are better known as mantis shrimp with considerable ecological importance in wide coastal waters globally. Some stomatopod species are exploited commercially, including Oratosquilla oratoria in the Northwe...Stomatopods are better known as mantis shrimp with considerable ecological importance in wide coastal waters globally. Some stomatopod species are exploited commercially, including Oratosquilla oratoria in the Northwest Pacific. Yet, few studies have published to promote accurate habitat identification of stomatopods, obstructing scientific management and conservation of these valuable organisms. This study provides an ensemble modeling framework for habitat suitability modeling of stomatopods, utilizing the O. oratoria stock in the Bohai Sea as an example. Two modeling techniques(i.e., generalized additive model(GAM) and geographical weighted regression(GWR)) were applied to select environmental predictors(especially the selection between two types of sediment metrics) that better characterize O. oratoria distribution and build separate habitat suitability models(HSM). The performance of the individual HSMs were compared on interpolation accuracy and transferability.Then, they were integrated to check whether the ensemble model outperforms either individual model, according to fishers’ knowledge and scientific survey data. As a result, grain-size metrics of sediment outperformed sediment content metrics in modeling O. oratoria habitat, possibly because grain-size metrics not only reflect the effect of substrates on burrow development, but also link to sediment heat capacity which influences individual thermoregulation. Moreover, the GWR-based HSM outperformed the GAM-based HSM in interpolation accuracy,while the latter one displayed better transferability. On balance, the ensemble HSM appeared to improve the predictive performance overall, as it could avoid dependence on a single model type and successfully identified fisher-recognized and survey-indicated suitable habitats in either sparsely sampled or well investigated areas.展开更多
Based on a simple coupled Lorenz model,we investigate how to assess a suitable initial perturbation scheme for ensemble forecasting in a multiscale system involving slow dynamics and fast dynamics.Four initial perturb...Based on a simple coupled Lorenz model,we investigate how to assess a suitable initial perturbation scheme for ensemble forecasting in a multiscale system involving slow dynamics and fast dynamics.Four initial perturbation approaches are used in the ensemble forecasting experiments:the random perturbation(RP),the bred vector(BV),the ensemble transform Kalman filter(ETKF),and the nonlinear local Lyapunov vector(NLLV)methods.Results show that,regardless of the method used,the ensemble averages behave indistinguishably from the control forecasts during the first few time steps.Due to different error growth in different time-scale systems,the ensemble averages perform better than the control forecast after very short lead times in a fast subsystem but after a relatively long period of time in a slow subsystem.Due to the coupled dynamic processes,the addition of perturbations to fast variables or to slow variables can contribute to an improvement in the forecasting skill for fast variables and slow variables.Regarding the initial perturbation approaches,the NLLVs show higher forecasting skill than the BVs or RPs overall.The NLLVs and ETKFs had nearly equivalent prediction skill,but NLLVs performed best by a narrow margin.In particular,when adding perturbations to slow variables,the independent perturbations(NLLVs and ETKFs)perform much better in ensemble prediction.These results are simply implied in a real coupled air–sea model.For the prediction of oceanic variables,using independent perturbations(NLLVs)and adding perturbations to oceanic variables are expected to result in better performance in the ensemble prediction.展开更多
The large blast furnace is essential equipment in the process of iron and steel manufacturing. Due to the complex operation process and frequent fluctuations of variables, conventional monitoring methods often bring f...The large blast furnace is essential equipment in the process of iron and steel manufacturing. Due to the complex operation process and frequent fluctuations of variables, conventional monitoring methods often bring false alarms. To address the above problem, an ensemble of greedy dynamic principal component analysis-Gaussian mixture model(EGDPCA-GMM) is proposed in this paper. First, PCA-GMM is introduced to deal with the collinearity and the non-Gaussian distribution of blast furnace data.Second, in order to explain the dynamics of data, the greedy algorithm is used to determine the extended variables and their corresponding time lags, so as to avoid introducing unnecessary noise. Then the bagging ensemble is adopted to cooperate with greedy extension to eliminate the randomness brought by the greedy algorithm and further reduce the false alarm rate(FAR) of monitoring results. Finally, the algorithm is applied to the blast furnace of a large iron and steel group in South China to verify performance.Compared with the basic algorithms, the proposed method achieves lowest FAR, while keeping missed alarm rate(MAR) remain stable.展开更多
Proper waste management models using recent technologies like computer vision,machine learning(ML),and deep learning(DL)are needed to effectively handle the massive quantity of increasing waste.Therefore,waste classif...Proper waste management models using recent technologies like computer vision,machine learning(ML),and deep learning(DL)are needed to effectively handle the massive quantity of increasing waste.Therefore,waste classification becomes a crucial topic which helps to categorize waste into hazardous or non-hazardous ones and thereby assist in the decision making of the waste management process.This study concentrates on the design of hazardous waste detection and classification using ensemble learning(HWDC-EL)technique to reduce toxicity and improve human health.The goal of the HWDC-EL technique is to detect the multiple classes of wastes,particularly hazardous and non-hazardous wastes.The HWDC-EL technique involves the ensemble of three feature extractors using Model Averaging technique namely discrete local binary patterns(DLBP),EfficientNet,and DenseNet121.In addition,the flower pollination algorithm(FPA)based hyperparameter optimizers are used to optimally adjust the parameters involved in the EfficientNet and DenseNet121 models.Moreover,a weighted voting-based ensemble classifier is derived using three machine learning algorithms namely support vector machine(SVM),extreme learning machine(ELM),and gradient boosting tree(GBT).The performance of the HWDC-EL technique is tested using a benchmark Garbage dataset and it obtains a maximum accuracy of 98.85%.展开更多
Numerous factors affect the increased temperature of a machine tool, including prolonged and high-intensity usage,tool-workpiece interaction, mechanical friction, and elevated ambient temperatures, among others. Conse...Numerous factors affect the increased temperature of a machine tool, including prolonged and high-intensity usage,tool-workpiece interaction, mechanical friction, and elevated ambient temperatures, among others. Consequently,spindle thermal displacement occurs, and machining precision suffers. To prevent the errors caused by thetemperature rise of the Spindle fromaffecting the accuracy during themachining process, typically, the factory willwarm up themachine before themanufacturing process.However, if there is noway to understand the tool spindle’sthermal deformation, the machining quality will be greatly affected. In order to solve the above problem, thisstudy aims to predict the thermal displacement of the machine tool by using intelligent algorithms. In the practicalapplication, only a few temperature sensors are used to input the information into the prediction model for realtimethermal displacement prediction. This approach has greatly improved the quality of tool processing.However,each algorithm has different performances in different environments. In this study, an ensemble model is used tointegrate Long Short-TermMemory (LSTM) with Support VectorMachine (SVM). The experimental results showthat the prediction performance of LSTM-SVM is higher than that of other machine learning algorithms.展开更多
Breast cancer is one of the leading cancers among women.It has the second-highest mortality rate in women after lung cancer.Timely detection,especially in the early stages,can help increase survival rates.However,manu...Breast cancer is one of the leading cancers among women.It has the second-highest mortality rate in women after lung cancer.Timely detection,especially in the early stages,can help increase survival rates.However,manual diagnosis of breast cancer is a tedious and time-consuming process,and the accuracy of detection is reliant on the quality of the images and the radiologist’s experience.However,computer-aided medical diagnosis has recently shown promising results,leading to the need to develop an efficient system that can aid radiologists in diagnosing breast cancer in its early stages.The research presented in this paper is focused on the multi-class classification of breast cancer.The deep transfer learning approach has been utilized to train the deep learning models,and a pre-processing technique has been used to improve the quality of the ultrasound dataset.The proposed technique utilizes two deep learning models,Mobile-NetV2 and DenseNet201,for the composition of the deep ensemble model.Deep learning models are fine-tuned along with hyperparameter tuning to achieve better results.Subsequently,entropy-based feature selection is used.Breast cancer identification using the proposed classification approach was found to attain an accuracy of 97.04%,while the sensitivity and F1 score were 96.87%and 96.76%,respectively.The performance of the proposed model is very effective and outperforms other state-of-the-art techniques presented in the literature.展开更多
The oil industries are an important part of a country’s economy.The crude oil’s price is influenced by a wide range of variables.Therefore,how accurately can countries predict its behavior and what predictors to emp...The oil industries are an important part of a country’s economy.The crude oil’s price is influenced by a wide range of variables.Therefore,how accurately can countries predict its behavior and what predictors to employ are two main questions.In this view,we propose utilizing deep learning and ensemble learning techniques to boost crude oil’s price forecasting performance.The suggested method is based on a deep learning snapshot ensemble method of the Transformer model.To examine the superiority of the proposed model,this paper compares the proposed deep learning ensemble model against different machine learning and statistical models for daily Organization of the Petroleum Exporting Countries(OPEC)oil price forecasting.Experimental results demonstrated the outperformance of the proposed method over statistical and machine learning methods.More precisely,the proposed snapshot ensemble of Transformer method achieved relative improvement in the forecasting performance compared to autoregressive integrated moving average ARIMA(1,1,1),ARIMA(0,1,1),autoregressive moving average(ARMA)(0,1),vector autoregression(VAR),random walk(RW),support vector machine(SVM),and random forests(RF)models by 99.94%,99.62%,99.87%,99.65%,7.55%,98.38%,and 99.35%,respectively,according to mean square error metric.展开更多
In forecasting real time environmental factors,large data is needed to analyse the pattern behind the data values.Air pollution is a major threat towards developing countries and it is proliferating every year.Many me...In forecasting real time environmental factors,large data is needed to analyse the pattern behind the data values.Air pollution is a major threat towards developing countries and it is proliferating every year.Many methods in time ser-ies prediction and deep learning models to estimate the severity of air pollution.Each independent variable contributing towards pollution is necessary to analyse the trend behind the air pollution in that particular locality.This approach selects multivariate time series and coalesce a real time updatable autoregressive model to forecast Particulate matter(PM)PM2.5.To perform experimental analysis the data from the Central Pollution Control Board(CPCB)is used.Prediction is car-ried out for Chennai with seven locations and estimated PM’s using the weighted ensemble method.Proposed method for air pollution prediction unveiled effective and moored performance in long term prediction.Dynamic budge with high weighted k-models are used simultaneously and devising an ensemble helps to achieve stable forecasting.Computational time of ensemble decreases with paral-lel processing in each sub model.Weighted ensemble model shows high perfor-mance in long term prediction when compared to the traditional time series models like Vector Auto-Regression(VAR),Autoregressive Integrated with Mov-ing Average(ARIMA),Autoregressive Moving Average with Extended terms(ARMEX).Evaluation metrics like Root Mean Square Error(RMSE),Mean Absolute Error(MAE)and the time to achieve the time series are compared.展开更多
Pneumonia is an acute lung infection that has caused many fatalitiesglobally. Radiologists often employ chest X-rays to identify pneumoniasince they are presently the most effective imaging method for this purpose.Com...Pneumonia is an acute lung infection that has caused many fatalitiesglobally. Radiologists often employ chest X-rays to identify pneumoniasince they are presently the most effective imaging method for this purpose.Computer-aided diagnosis of pneumonia using deep learning techniques iswidely used due to its effectiveness and performance. In the proposed method,the Synthetic Minority Oversampling Technique (SMOTE) approach is usedto eliminate the class imbalance in the X-ray dataset. To compensate forthe paucity of accessible data, pre-trained transfer learning is used, and anensemble Convolutional Neural Network (CNN) model is developed. Theensemble model consists of all possible combinations of the MobileNetv2,Visual Geometry Group (VGG16), and DenseNet169 models. MobileNetV2and DenseNet169 performed well in the Single classifier model, with anaccuracy of 94%, while the ensemble model (MobileNetV2+DenseNet169)achieved an accuracy of 96.9%. Using the data synchronous parallel modelin Distributed Tensorflow, the training process accelerated performance by98.6% and outperformed other conventional approaches.展开更多
Pneumonia is a dangerous respiratory disease due to which breathing becomes incredibly difficult and painful;thus,catching it early is crucial.Medical physicians’time is limited in outdoor situations due to many pati...Pneumonia is a dangerous respiratory disease due to which breathing becomes incredibly difficult and painful;thus,catching it early is crucial.Medical physicians’time is limited in outdoor situations due to many patients;therefore,automated systems can be a rescue.The input images from the X-ray equipment are also highly unpredictable due to variances in radiologists’experience.Therefore,radiologists require an automated system that can swiftly and accurately detect pneumonic lungs from chest x-rays.In medical classifications,deep convolution neural networks are commonly used.This research aims to use deep pretrained transfer learning models to accurately categorize CXR images into binary classes,i.e.,Normal and Pneumonia.The MDEV is a proposed novel ensemble approach that concatenates four heterogeneous transfer learning models:Mobile-Net,DenseNet-201,EfficientNet-B0,and VGG-16,which have been finetuned and trained on 5,856 CXR images.The evaluation matrices used in this research to contrast different deep transfer learning architectures include precision,accuracy,recall,AUC-roc,and f1-score.The model effectively decreases training loss while increasing accuracy.The findings conclude that the proposed MDEV model outperformed cutting-edge deep transfer learning models and obtains an overall precision of 92.26%,an accuracy of 92.15%,a recall of 90.90%,an auc-roc score of 90.9%,and f-score of 91.49%with minimal data pre-processing,data augmentation,finetuning and hyperparameter adjustment in classifying Normal and Pneumonia chests.展开更多
Covid-19 is a deadly virus that is rapidly spread around the world towards the end of the 2020.The consequences of this virus are quite frightening,especially when accompanied by an underlying disease.The novelty of t...Covid-19 is a deadly virus that is rapidly spread around the world towards the end of the 2020.The consequences of this virus are quite frightening,especially when accompanied by an underlying disease.The novelty of the virus,the constant emergence of different variants and its rapid spread have a negative impact on the control and treatment process.Although the new test kits provide almost certain results,chest X-rays are extremely important to detect the progression and degree of the disease.In addition to the Covid-19 virus,pneumonia and harmless opacity of the lungs also complicate the diagnosis.Considering the negative results caused by the virus and the treatment costs,the importance of fast and accurate diagnosis is clearly seen.In this context,deep learning methods appear as an extremely popular approach.In this study,a hybrid model design with superior properties of convolutional neural networks is presented to correctly classify the Covid-19 disease.In addition,in order to contribute to the literature,a suitable dataset with balanced case numbers that can be used in all artificial intelligence classification studies is presented.With this ensemble model design,quite remarkable results are obtained for the diagnosis of three and four-class Covid-19.The proposed model can classify normal,pneumonia,and Covid-19 with 92.6%accuracy and 82.6%for normal,pneumonia,Covid-19,and lung opacity.展开更多
The number of attacks is growing tremendously in tandem with the growth of internet technologies.As a result,protecting the private data from prying eyes has become a critical and tough undertaking.Many intrusion dete...The number of attacks is growing tremendously in tandem with the growth of internet technologies.As a result,protecting the private data from prying eyes has become a critical and tough undertaking.Many intrusion detection solutions have been offered by researchers in order to decrease the effect of these attacks.For attack detection,the prior system has created an SMSRPF(Stacking Model Significant Rule Power Factor)classifier.To provide creative instance detection,the SMSRPF combines the detection of trained classifiers such as DT(Decision Tree)and RF(Random Forest).Nevertheless,it does not generate any accuratefindings that are adequate.The suggested system has built an EWF(Ensemble Wrapper Filter)feature selection with SMSRPF classifier for attack detection so as to overcome this problem.The UNSW-NB15 dataset is used as an input in this proposed research project.Specifically,min–max normalization approach is used to pre-process the incoming data.The feature selection is then carried out using EWF.Based on the selected features,SMSRPF classifiers are utilized to detect the attacks.The SMSRPF is integrated with the trained classi-fiers such as DT and RF to create creative instance detection.After that,the testing data is classified using MCAR(Multi-Class Classification based on Association Rules).The SRPF judges the rules correctly even when the confidence and the lift measures fail.Regarding accuracy,precision,recall,f-measure,computation time,and error,the experimental findings suggest that the new system outperforms the prior systems.展开更多
The ensemble optimal interpolation (EnOI) is applied to the regional ocean modeling system (ROMS) with the ability to assimilate the along-track sea level anomaly (TSLA). This system is tested with an eddy-resol...The ensemble optimal interpolation (EnOI) is applied to the regional ocean modeling system (ROMS) with the ability to assimilate the along-track sea level anomaly (TSLA). This system is tested with an eddy-resolving system of the South China Sea (SCS). Background errors are derived from a running seasonal ensemble to account for the seasonal variability within the SCS. A fifth-order localization function with a 250 km localization radius is chosen to reduce the negative effects of sampling errors. The data assimilation system is tested from January 2004 to December 2006. The results show that the root mean square deviation (RMSD) of the sea level anomaly decreased from 10.57 to 6.70 cm, which represents a 36.6% reduction of error. The data assimilation reduces error for temperature within the upper 800 m and for salinity within the upper 200 m, although error degrades slightly at deeper depths. Surface currents are in better agreement with trajectories of surface drifters after data assimilation. The variance of sea level improves significantly in terms of both the amplitude and position of the strong and weak variance regions after assimilating TSLA. Results with AGE error (AGE) perform better than no AGE error (NoAGE) when considering the improvements of the temperature and the salinity. Furthermore, reasons for the extremely strong variability in the northern SCS in high resolution models are investigated. The results demonstrate that the strong variability of sea level in the high resolution model is caused by an extremely strong Kuroshio intrusion. Therefore, it is demonstrated that it is necessary to assimilate the TSLA in order to better simulate the SCS with high resolution models.展开更多
The final product quality is determined by cumulation, coupling and propagation of product quality variations from all stations in multi-stage manufacturing systems (MMSs). Modeling and control of variation propagat...The final product quality is determined by cumulation, coupling and propagation of product quality variations from all stations in multi-stage manufacturing systems (MMSs). Modeling and control of variation propagation is essential to improve product quality. However, the current stream of variations (SOV) theory can only solve the problem that a single SOV affects the product quality. Due to the existence of multiple variation streams, limited research has been done on the quality control in serial-parallel hybrid multi-stage manufacturing systems (SPH-MMSs). A state space model and its modeling strategies are developed to describe the multiple variation streams stack-up in an SPH-MMS. The SOV theory is extended to SPH-MMS. The dimensions of system model are reduced to the production-reality level, and the effect and feasibility of the model is validated by a machining case.展开更多
Predictive analytics have been widely used in the literature with respect to laparoscopic surgery and risk stratification.However,most predictive analytics in this field exploit generalized linearmodels for predictive...Predictive analytics have been widely used in the literature with respect to laparoscopic surgery and risk stratification.However,most predictive analytics in this field exploit generalized linearmodels for predictive purposes,which are limited by model assumptionsdincluding linearity between response variables and additive interactions between variables.In many instances,such assumptions may not hold true,and the complex relationship between predictors and response variables is usually unknown.To address this limitation,machine-learning algorithms can be employed to model the underlying data.The advantage of machine learning algorithms is that they usually do not require strict assumptions regarding data structure,and they are able to learn complex functional forms using a nonparametric approach.Furthermore,two or more machine learning algorithms can be synthesized to further improve predictive accuracy.Such a process is referred to as ensemble modeling,and it has been used broadly in various industries.However,this approach has not been widely reported in the laparoscopic surgical literature due to its complexity in both model training and interpretation.With this technical note,we provide a comprehensive overview of the ensemble-modeling technique and a step-by-step tutorial on how to implement ensemble modeling.展开更多
A single model cannot satisfy the high-precision prediction requirements given the high nonlinearity between variables.By contrast,ensemble models can effectively solve this problem.Three key factors for improving the...A single model cannot satisfy the high-precision prediction requirements given the high nonlinearity between variables.By contrast,ensemble models can effectively solve this problem.Three key factors for improving the accuracy of ensemble models are namely the high accuracy of a submodel,the diversity between subsample sets and the optimal ensemble method.This study presents an improved ensemble modeling method to improve the prediction precision and generalization capability of the model.Our proposed method first uses a bagging algorithm to generate multiple subsample sets.Second,an indicator vector is defined to describe these subsample sets.Third,subsample sets are selected on the basis of the results of agglomerative nesting clustering on indicator vectors to maximize the diversity between subsets.Subsequently,these subsample sets are placed in a stacked autoencoder for training.Finally,XGBoost algorithm,rather than the traditional simple average ensemble method,is imported to ensemble the model during modeling.Three machine learning public datasets and atmospheric column dry point dataset from a practical industrial process show that our proposed method demonstrates high precision and improved prediction ability.展开更多
The estimation of model parameters is an important subject in engineering.In this area of work,the prevailing approach is to estimate or calculate these as deterministic parameters.In this study,we consider the model ...The estimation of model parameters is an important subject in engineering.In this area of work,the prevailing approach is to estimate or calculate these as deterministic parameters.In this study,we consider the model parameters from the perspective of random variables and describe the general form of the parameter distribution inference problem.Under this framework,we propose an ensemble Bayesian method by introducing Bayesian inference and the Markov chain Monte Carlo(MCMC)method.Experiments on a finite cylindrical reactor and a 2D IAEA benchmark problem show that the proposed method converges quickly and can estimate parameters effectively,even for several correlated parameters simultaneously.Our experiments include cases of engineering software calls,demonstrating that the method can be applied to engineering,such as nuclear reactor engineering.展开更多
Landslide is a serious natural disaster next only to earthquake and flood,which will cause a great threat to people’s lives and property safety.The traditional research of landslide disaster based on experience-drive...Landslide is a serious natural disaster next only to earthquake and flood,which will cause a great threat to people’s lives and property safety.The traditional research of landslide disaster based on experience-driven or statistical model and its assessment results are subjective,difficult to quantify,and no pertinence.As a new research method for landslide susceptibility assessment,machine learning can greatly improve the landslide susceptibility model’s accuracy by constructing statistical models.Taking Western Henan for example,the study selected 16 landslide influencing factors such as topography,geological environment,hydrological conditions,and human activities,and 11 landslide factors with the most significant influence on the landslide were selected by the recursive feature elimination(RFE)method.Five machine learning methods[Support Vector Machines(SVM),Logistic Regression(LR),Random Forest(RF),Extreme Gradient Boosting(XGBoost),and Linear Discriminant Analysis(LDA)]were used to construct the spatial distribution model of landslide susceptibility.The models were evaluated by the receiver operating characteristic curve and statistical index.After analysis and comparison,the XGBoost model(AUC 0.8759)performed the best and was suitable for dealing with regression problems.The model had a high adaptability to landslide data.According to the landslide susceptibility map of the five models,the overall distribution can be observed.The extremely high and high susceptibility areas are distributed in the Funiu Mountain range in the southwest,the Xiaoshan Mountain range in the west,and the Yellow River Basin in the north.These areas have large terrain fluctuations,complicated geological structural environments and frequent human engineering activities.The extremely high and highly prone areas were 12043.3 km^(2)and 3087.45 km^(2),accounting for 47.61%and 12.20%of the total area of the study area,respectively.Our study reflects the distribution of landslide susceptibility in western Henan Province,which provides a scientific basis for regional disaster warning,prediction,and resource protection.The study has important practical significance for subsequent landslide disaster management.展开更多
文摘The Indian Himalayan region is frequently experiencing climate change-induced landslides.Thus,landslide susceptibility assessment assumes greater significance for lessening the impact of a landslide hazard.This paper makes an attempt to assess landslide susceptibility in Shimla district of the northwest Indian Himalayan region.It examined the effectiveness of random forest(RF),multilayer perceptron(MLP),sequential minimal optimization regression(SMOreg)and bagging ensemble(B-RF,BSMOreg,B-MLP)models.A landslide inventory map comprising 1052 locations of past landslide occurrences was classified into training(70%)and testing(30%)datasets.The site-specific influencing factors were selected by employing a multicollinearity test.The relationship between past landslide occurrences and influencing factors was established using the frequency ratio method.The effectiveness of machine learning models was verified through performance assessors.The landslide susceptibility maps were validated by the area under the receiver operating characteristic curves(ROC-AUC),accuracy,precision,recall and F1-score.The key performance metrics and map validation demonstrated that the BRF model(correlation coefficient:0.988,mean absolute error:0.010,root mean square error:0.058,relative absolute error:2.964,ROC-AUC:0.947,accuracy:0.778,precision:0.819,recall:0.917 and F-1 score:0.865)outperformed the single classifiers and other bagging ensemble models for landslide susceptibility.The results show that the largest area was found under the very high susceptibility zone(33.87%),followed by the low(27.30%),high(20.68%)and moderate(18.16%)susceptibility zones.The factors,namely average annual rainfall,slope,lithology,soil texture and earthquake magnitude have been identified as the influencing factors for very high landslide susceptibility.Soil texture,lineament density and elevation have been attributed to high and moderate susceptibility.Thus,the study calls for devising suitable landslide mitigation measures in the study area.Structural measures,an immediate response system,community participation and coordination among stakeholders may help lessen the detrimental impact of landslides.The findings from this study could aid decision-makers in mitigating future catastrophes and devising suitable strategies in other geographical regions with similar geological characteristics.
文摘Sentence classification is the process of categorizing a sentence based on the context of the sentence.Sentence categorization requires more semantic highlights than other tasks,such as dependence parsing,which requires more syntactic elements.Most existing strategies focus on the general semantics of a conversation without involving the context of the sentence,recognizing the progress and comparing impacts.An ensemble pre-trained language model was taken up here to classify the conversation sentences from the conversation corpus.The conversational sentences are classified into four categories:information,question,directive,and commission.These classification label sequences are for analyzing the conversation progress and predicting the pecking order of the conversation.Ensemble of Bidirectional Encoder for Representation of Transformer(BERT),Robustly Optimized BERT pretraining Approach(RoBERTa),Generative Pre-Trained Transformer(GPT),DistilBERT and Generalized Autoregressive Pretraining for Language Understanding(XLNet)models are trained on conversation corpus with hyperparameters.Hyperparameter tuning approach is carried out for better performance on sentence classification.This Ensemble of Pre-trained Language Models with a Hyperparameter Tuning(EPLM-HT)system is trained on an annotated conversation dataset.The proposed approach outperformed compared to the base BERT,GPT,DistilBERT and XLNet transformer models.The proposed ensemble model with the fine-tuned parameters achieved an F1_score of 0.88.
基金The National Natural Science Foundation of China under contract No.31902375the David and Lucile Packard Foundation+1 种基金the Innovation Team of Fishery Resources and Ecology in the Yellow Sea and Bohai Sea under contract No.2020TD01the Special Funds for Taishan Scholars Project of Shandong Province。
文摘Stomatopods are better known as mantis shrimp with considerable ecological importance in wide coastal waters globally. Some stomatopod species are exploited commercially, including Oratosquilla oratoria in the Northwest Pacific. Yet, few studies have published to promote accurate habitat identification of stomatopods, obstructing scientific management and conservation of these valuable organisms. This study provides an ensemble modeling framework for habitat suitability modeling of stomatopods, utilizing the O. oratoria stock in the Bohai Sea as an example. Two modeling techniques(i.e., generalized additive model(GAM) and geographical weighted regression(GWR)) were applied to select environmental predictors(especially the selection between two types of sediment metrics) that better characterize O. oratoria distribution and build separate habitat suitability models(HSM). The performance of the individual HSMs were compared on interpolation accuracy and transferability.Then, they were integrated to check whether the ensemble model outperforms either individual model, according to fishers’ knowledge and scientific survey data. As a result, grain-size metrics of sediment outperformed sediment content metrics in modeling O. oratoria habitat, possibly because grain-size metrics not only reflect the effect of substrates on burrow development, but also link to sediment heat capacity which influences individual thermoregulation. Moreover, the GWR-based HSM outperformed the GAM-based HSM in interpolation accuracy,while the latter one displayed better transferability. On balance, the ensemble HSM appeared to improve the predictive performance overall, as it could avoid dependence on a single model type and successfully identified fisher-recognized and survey-indicated suitable habitats in either sparsely sampled or well investigated areas.
基金jointly supported by the National Natural Science Foundation of China (Grant Nos. 42225501, 42105059)
文摘Based on a simple coupled Lorenz model,we investigate how to assess a suitable initial perturbation scheme for ensemble forecasting in a multiscale system involving slow dynamics and fast dynamics.Four initial perturbation approaches are used in the ensemble forecasting experiments:the random perturbation(RP),the bred vector(BV),the ensemble transform Kalman filter(ETKF),and the nonlinear local Lyapunov vector(NLLV)methods.Results show that,regardless of the method used,the ensemble averages behave indistinguishably from the control forecasts during the first few time steps.Due to different error growth in different time-scale systems,the ensemble averages perform better than the control forecast after very short lead times in a fast subsystem but after a relatively long period of time in a slow subsystem.Due to the coupled dynamic processes,the addition of perturbations to fast variables or to slow variables can contribute to an improvement in the forecasting skill for fast variables and slow variables.Regarding the initial perturbation approaches,the NLLVs show higher forecasting skill than the BVs or RPs overall.The NLLVs and ETKFs had nearly equivalent prediction skill,but NLLVs performed best by a narrow margin.In particular,when adding perturbations to slow variables,the independent perturbations(NLLVs and ETKFs)perform much better in ensemble prediction.These results are simply implied in a real coupled air–sea model.For the prediction of oceanic variables,using independent perturbations(NLLVs)and adding perturbations to oceanic variables are expected to result in better performance in the ensemble prediction.
基金supported by the National Natural Science Foundation of China (61903326, 61933015)。
文摘The large blast furnace is essential equipment in the process of iron and steel manufacturing. Due to the complex operation process and frequent fluctuations of variables, conventional monitoring methods often bring false alarms. To address the above problem, an ensemble of greedy dynamic principal component analysis-Gaussian mixture model(EGDPCA-GMM) is proposed in this paper. First, PCA-GMM is introduced to deal with the collinearity and the non-Gaussian distribution of blast furnace data.Second, in order to explain the dynamics of data, the greedy algorithm is used to determine the extended variables and their corresponding time lags, so as to avoid introducing unnecessary noise. Then the bagging ensemble is adopted to cooperate with greedy extension to eliminate the randomness brought by the greedy algorithm and further reduce the false alarm rate(FAR) of monitoring results. Finally, the algorithm is applied to the blast furnace of a large iron and steel group in South China to verify performance.Compared with the basic algorithms, the proposed method achieves lowest FAR, while keeping missed alarm rate(MAR) remain stable.
基金the Deanship of Scientific Research at King Khalid University for funding this work underGrant Number(RGP 2/209/42)PrincessNourah bint Abdulrahman University Researchers Supporting Project Number(PNURSP2022R136)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code:(22UQU4210118DSR27).
文摘Proper waste management models using recent technologies like computer vision,machine learning(ML),and deep learning(DL)are needed to effectively handle the massive quantity of increasing waste.Therefore,waste classification becomes a crucial topic which helps to categorize waste into hazardous or non-hazardous ones and thereby assist in the decision making of the waste management process.This study concentrates on the design of hazardous waste detection and classification using ensemble learning(HWDC-EL)technique to reduce toxicity and improve human health.The goal of the HWDC-EL technique is to detect the multiple classes of wastes,particularly hazardous and non-hazardous wastes.The HWDC-EL technique involves the ensemble of three feature extractors using Model Averaging technique namely discrete local binary patterns(DLBP),EfficientNet,and DenseNet121.In addition,the flower pollination algorithm(FPA)based hyperparameter optimizers are used to optimally adjust the parameters involved in the EfficientNet and DenseNet121 models.Moreover,a weighted voting-based ensemble classifier is derived using three machine learning algorithms namely support vector machine(SVM),extreme learning machine(ELM),and gradient boosting tree(GBT).The performance of the HWDC-EL technique is tested using a benchmark Garbage dataset and it obtains a maximum accuracy of 98.85%.
基金supported by the Ministry of Science and Technology,Taiwan,under Grant MOST 110-2218-E-194-010。
文摘Numerous factors affect the increased temperature of a machine tool, including prolonged and high-intensity usage,tool-workpiece interaction, mechanical friction, and elevated ambient temperatures, among others. Consequently,spindle thermal displacement occurs, and machining precision suffers. To prevent the errors caused by thetemperature rise of the Spindle fromaffecting the accuracy during themachining process, typically, the factory willwarm up themachine before themanufacturing process.However, if there is noway to understand the tool spindle’sthermal deformation, the machining quality will be greatly affected. In order to solve the above problem, thisstudy aims to predict the thermal displacement of the machine tool by using intelligent algorithms. In the practicalapplication, only a few temperature sensors are used to input the information into the prediction model for realtimethermal displacement prediction. This approach has greatly improved the quality of tool processing.However,each algorithm has different performances in different environments. In this study, an ensemble model is used tointegrate Long Short-TermMemory (LSTM) with Support VectorMachine (SVM). The experimental results showthat the prediction performance of LSTM-SVM is higher than that of other machine learning algorithms.
基金This research work was funded by Institutional Fund Projects under Grant No.(IFPIP:1614-611-1442)from the Ministry of Education and King Abdulaziz University,DSR,Jeddah,Saudi Arabia.
文摘Breast cancer is one of the leading cancers among women.It has the second-highest mortality rate in women after lung cancer.Timely detection,especially in the early stages,can help increase survival rates.However,manual diagnosis of breast cancer is a tedious and time-consuming process,and the accuracy of detection is reliant on the quality of the images and the radiologist’s experience.However,computer-aided medical diagnosis has recently shown promising results,leading to the need to develop an efficient system that can aid radiologists in diagnosing breast cancer in its early stages.The research presented in this paper is focused on the multi-class classification of breast cancer.The deep transfer learning approach has been utilized to train the deep learning models,and a pre-processing technique has been used to improve the quality of the ultrasound dataset.The proposed technique utilizes two deep learning models,Mobile-NetV2 and DenseNet201,for the composition of the deep ensemble model.Deep learning models are fine-tuned along with hyperparameter tuning to achieve better results.Subsequently,entropy-based feature selection is used.Breast cancer identification using the proposed classification approach was found to attain an accuracy of 97.04%,while the sensitivity and F1 score were 96.87%and 96.76%,respectively.The performance of the proposed model is very effective and outperforms other state-of-the-art techniques presented in the literature.
文摘The oil industries are an important part of a country’s economy.The crude oil’s price is influenced by a wide range of variables.Therefore,how accurately can countries predict its behavior and what predictors to employ are two main questions.In this view,we propose utilizing deep learning and ensemble learning techniques to boost crude oil’s price forecasting performance.The suggested method is based on a deep learning snapshot ensemble method of the Transformer model.To examine the superiority of the proposed model,this paper compares the proposed deep learning ensemble model against different machine learning and statistical models for daily Organization of the Petroleum Exporting Countries(OPEC)oil price forecasting.Experimental results demonstrated the outperformance of the proposed method over statistical and machine learning methods.More precisely,the proposed snapshot ensemble of Transformer method achieved relative improvement in the forecasting performance compared to autoregressive integrated moving average ARIMA(1,1,1),ARIMA(0,1,1),autoregressive moving average(ARMA)(0,1),vector autoregression(VAR),random walk(RW),support vector machine(SVM),and random forests(RF)models by 99.94%,99.62%,99.87%,99.65%,7.55%,98.38%,and 99.35%,respectively,according to mean square error metric.
文摘In forecasting real time environmental factors,large data is needed to analyse the pattern behind the data values.Air pollution is a major threat towards developing countries and it is proliferating every year.Many methods in time ser-ies prediction and deep learning models to estimate the severity of air pollution.Each independent variable contributing towards pollution is necessary to analyse the trend behind the air pollution in that particular locality.This approach selects multivariate time series and coalesce a real time updatable autoregressive model to forecast Particulate matter(PM)PM2.5.To perform experimental analysis the data from the Central Pollution Control Board(CPCB)is used.Prediction is car-ried out for Chennai with seven locations and estimated PM’s using the weighted ensemble method.Proposed method for air pollution prediction unveiled effective and moored performance in long term prediction.Dynamic budge with high weighted k-models are used simultaneously and devising an ensemble helps to achieve stable forecasting.Computational time of ensemble decreases with paral-lel processing in each sub model.Weighted ensemble model shows high perfor-mance in long term prediction when compared to the traditional time series models like Vector Auto-Regression(VAR),Autoregressive Integrated with Mov-ing Average(ARIMA),Autoregressive Moving Average with Extended terms(ARMEX).Evaluation metrics like Root Mean Square Error(RMSE),Mean Absolute Error(MAE)and the time to achieve the time series are compared.
文摘Pneumonia is an acute lung infection that has caused many fatalitiesglobally. Radiologists often employ chest X-rays to identify pneumoniasince they are presently the most effective imaging method for this purpose.Computer-aided diagnosis of pneumonia using deep learning techniques iswidely used due to its effectiveness and performance. In the proposed method,the Synthetic Minority Oversampling Technique (SMOTE) approach is usedto eliminate the class imbalance in the X-ray dataset. To compensate forthe paucity of accessible data, pre-trained transfer learning is used, and anensemble Convolutional Neural Network (CNN) model is developed. Theensemble model consists of all possible combinations of the MobileNetv2,Visual Geometry Group (VGG16), and DenseNet169 models. MobileNetV2and DenseNet169 performed well in the Single classifier model, with anaccuracy of 94%, while the ensemble model (MobileNetV2+DenseNet169)achieved an accuracy of 96.9%. Using the data synchronous parallel modelin Distributed Tensorflow, the training process accelerated performance by98.6% and outperformed other conventional approaches.
基金This research was supported by Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(2021R1I1A1A01052299).
文摘Pneumonia is a dangerous respiratory disease due to which breathing becomes incredibly difficult and painful;thus,catching it early is crucial.Medical physicians’time is limited in outdoor situations due to many patients;therefore,automated systems can be a rescue.The input images from the X-ray equipment are also highly unpredictable due to variances in radiologists’experience.Therefore,radiologists require an automated system that can swiftly and accurately detect pneumonic lungs from chest x-rays.In medical classifications,deep convolution neural networks are commonly used.This research aims to use deep pretrained transfer learning models to accurately categorize CXR images into binary classes,i.e.,Normal and Pneumonia.The MDEV is a proposed novel ensemble approach that concatenates four heterogeneous transfer learning models:Mobile-Net,DenseNet-201,EfficientNet-B0,and VGG-16,which have been finetuned and trained on 5,856 CXR images.The evaluation matrices used in this research to contrast different deep transfer learning architectures include precision,accuracy,recall,AUC-roc,and f1-score.The model effectively decreases training loss while increasing accuracy.The findings conclude that the proposed MDEV model outperformed cutting-edge deep transfer learning models and obtains an overall precision of 92.26%,an accuracy of 92.15%,a recall of 90.90%,an auc-roc score of 90.9%,and f-score of 91.49%with minimal data pre-processing,data augmentation,finetuning and hyperparameter adjustment in classifying Normal and Pneumonia chests.
文摘Covid-19 is a deadly virus that is rapidly spread around the world towards the end of the 2020.The consequences of this virus are quite frightening,especially when accompanied by an underlying disease.The novelty of the virus,the constant emergence of different variants and its rapid spread have a negative impact on the control and treatment process.Although the new test kits provide almost certain results,chest X-rays are extremely important to detect the progression and degree of the disease.In addition to the Covid-19 virus,pneumonia and harmless opacity of the lungs also complicate the diagnosis.Considering the negative results caused by the virus and the treatment costs,the importance of fast and accurate diagnosis is clearly seen.In this context,deep learning methods appear as an extremely popular approach.In this study,a hybrid model design with superior properties of convolutional neural networks is presented to correctly classify the Covid-19 disease.In addition,in order to contribute to the literature,a suitable dataset with balanced case numbers that can be used in all artificial intelligence classification studies is presented.With this ensemble model design,quite remarkable results are obtained for the diagnosis of three and four-class Covid-19.The proposed model can classify normal,pneumonia,and Covid-19 with 92.6%accuracy and 82.6%for normal,pneumonia,Covid-19,and lung opacity.
文摘The number of attacks is growing tremendously in tandem with the growth of internet technologies.As a result,protecting the private data from prying eyes has become a critical and tough undertaking.Many intrusion detection solutions have been offered by researchers in order to decrease the effect of these attacks.For attack detection,the prior system has created an SMSRPF(Stacking Model Significant Rule Power Factor)classifier.To provide creative instance detection,the SMSRPF combines the detection of trained classifiers such as DT(Decision Tree)and RF(Random Forest).Nevertheless,it does not generate any accuratefindings that are adequate.The suggested system has built an EWF(Ensemble Wrapper Filter)feature selection with SMSRPF classifier for attack detection so as to overcome this problem.The UNSW-NB15 dataset is used as an input in this proposed research project.Specifically,min–max normalization approach is used to pre-process the incoming data.The feature selection is then carried out using EWF.Based on the selected features,SMSRPF classifiers are utilized to detect the attacks.The SMSRPF is integrated with the trained classi-fiers such as DT and RF to create creative instance detection.After that,the testing data is classified using MCAR(Multi-Class Classification based on Association Rules).The SRPF judges the rules correctly even when the confidence and the lift measures fail.Regarding accuracy,precision,recall,f-measure,computation time,and error,the experimental findings suggest that the new system outperforms the prior systems.
基金The Major State Basic Research Development Program of China under contract Nos 201-1CB403606 and 2011CB403500the National Natural Science Foundation of China under contract Nos 41222038,41076011and 41206023the National Marine Environmental Forecasting Center Operational Development Foundation of the State Oceanic Administration of China under contract No.2013002
文摘The ensemble optimal interpolation (EnOI) is applied to the regional ocean modeling system (ROMS) with the ability to assimilate the along-track sea level anomaly (TSLA). This system is tested with an eddy-resolving system of the South China Sea (SCS). Background errors are derived from a running seasonal ensemble to account for the seasonal variability within the SCS. A fifth-order localization function with a 250 km localization radius is chosen to reduce the negative effects of sampling errors. The data assimilation system is tested from January 2004 to December 2006. The results show that the root mean square deviation (RMSD) of the sea level anomaly decreased from 10.57 to 6.70 cm, which represents a 36.6% reduction of error. The data assimilation reduces error for temperature within the upper 800 m and for salinity within the upper 200 m, although error degrades slightly at deeper depths. Surface currents are in better agreement with trajectories of surface drifters after data assimilation. The variance of sea level improves significantly in terms of both the amplitude and position of the strong and weak variance regions after assimilating TSLA. Results with AGE error (AGE) perform better than no AGE error (NoAGE) when considering the improvements of the temperature and the salinity. Furthermore, reasons for the extremely strong variability in the northern SCS in high resolution models are investigated. The results demonstrate that the strong variability of sea level in the high resolution model is caused by an extremely strong Kuroshio intrusion. Therefore, it is demonstrated that it is necessary to assimilate the TSLA in order to better simulate the SCS with high resolution models.
基金This project is supported by National Natural Science Foundation of China(No.50675137).
文摘The final product quality is determined by cumulation, coupling and propagation of product quality variations from all stations in multi-stage manufacturing systems (MMSs). Modeling and control of variation propagation is essential to improve product quality. However, the current stream of variations (SOV) theory can only solve the problem that a single SOV affects the product quality. Due to the existence of multiple variation streams, limited research has been done on the quality control in serial-parallel hybrid multi-stage manufacturing systems (SPH-MMSs). A state space model and its modeling strategies are developed to describe the multiple variation streams stack-up in an SPH-MMS. The SOV theory is extended to SPH-MMS. The dimensions of system model are reduced to the production-reality level, and the effect and feasibility of the model is validated by a machining case.
基金funding from RUIYI emergency medical research fund(202013)Open Foundation of Artificial Intelligence Key Laboratory of Sichuan Province(2020RYY03)+1 种基金Research project of Health and Family Planning Commission of Sichuan Province(17PJ136)funding from Key Research&Development project of Zhejiang Province(2021C03071).
文摘Predictive analytics have been widely used in the literature with respect to laparoscopic surgery and risk stratification.However,most predictive analytics in this field exploit generalized linearmodels for predictive purposes,which are limited by model assumptionsdincluding linearity between response variables and additive interactions between variables.In many instances,such assumptions may not hold true,and the complex relationship between predictors and response variables is usually unknown.To address this limitation,machine-learning algorithms can be employed to model the underlying data.The advantage of machine learning algorithms is that they usually do not require strict assumptions regarding data structure,and they are able to learn complex functional forms using a nonparametric approach.Furthermore,two or more machine learning algorithms can be synthesized to further improve predictive accuracy.Such a process is referred to as ensemble modeling,and it has been used broadly in various industries.However,this approach has not been widely reported in the laparoscopic surgical literature due to its complexity in both model training and interpretation.With this technical note,we provide a comprehensive overview of the ensemble-modeling technique and a step-by-step tutorial on how to implement ensemble modeling.
基金The authors are grateful for the support of National Natural Science Foundation of China(21878081)Fundamental Research Funds for the Central Universities under Grant of China(222201717006)the Program of Introducing Talents of Discipline to Universities(the 111 Project)under Grant B17017.
文摘A single model cannot satisfy the high-precision prediction requirements given the high nonlinearity between variables.By contrast,ensemble models can effectively solve this problem.Three key factors for improving the accuracy of ensemble models are namely the high accuracy of a submodel,the diversity between subsample sets and the optimal ensemble method.This study presents an improved ensemble modeling method to improve the prediction precision and generalization capability of the model.Our proposed method first uses a bagging algorithm to generate multiple subsample sets.Second,an indicator vector is defined to describe these subsample sets.Third,subsample sets are selected on the basis of the results of agglomerative nesting clustering on indicator vectors to maximize the diversity between subsets.Subsequently,these subsample sets are placed in a stacked autoencoder for training.Finally,XGBoost algorithm,rather than the traditional simple average ensemble method,is imported to ensemble the model during modeling.Three machine learning public datasets and atmospheric column dry point dataset from a practical industrial process show that our proposed method demonstrates high precision and improved prediction ability.
基金partially sponsored by the Natural Science Foundation of Shanghai(No.23ZR1429300)the Innovation Fund of CNNC(Lingchuang Fund)。
文摘The estimation of model parameters is an important subject in engineering.In this area of work,the prevailing approach is to estimate or calculate these as deterministic parameters.In this study,we consider the model parameters from the perspective of random variables and describe the general form of the parameter distribution inference problem.Under this framework,we propose an ensemble Bayesian method by introducing Bayesian inference and the Markov chain Monte Carlo(MCMC)method.Experiments on a finite cylindrical reactor and a 2D IAEA benchmark problem show that the proposed method converges quickly and can estimate parameters effectively,even for several correlated parameters simultaneously.Our experiments include cases of engineering software calls,demonstrating that the method can be applied to engineering,such as nuclear reactor engineering.
基金This work was financially supported by National Natural Science Foundation of China(41972262)Hebei Natural Science Foundation for Excellent Young Scholars(D2020504032)+1 种基金Central Plains Science and technology innovation leader Project(214200510030)Key research and development Project of Henan province(221111321500).
文摘Landslide is a serious natural disaster next only to earthquake and flood,which will cause a great threat to people’s lives and property safety.The traditional research of landslide disaster based on experience-driven or statistical model and its assessment results are subjective,difficult to quantify,and no pertinence.As a new research method for landslide susceptibility assessment,machine learning can greatly improve the landslide susceptibility model’s accuracy by constructing statistical models.Taking Western Henan for example,the study selected 16 landslide influencing factors such as topography,geological environment,hydrological conditions,and human activities,and 11 landslide factors with the most significant influence on the landslide were selected by the recursive feature elimination(RFE)method.Five machine learning methods[Support Vector Machines(SVM),Logistic Regression(LR),Random Forest(RF),Extreme Gradient Boosting(XGBoost),and Linear Discriminant Analysis(LDA)]were used to construct the spatial distribution model of landslide susceptibility.The models were evaluated by the receiver operating characteristic curve and statistical index.After analysis and comparison,the XGBoost model(AUC 0.8759)performed the best and was suitable for dealing with regression problems.The model had a high adaptability to landslide data.According to the landslide susceptibility map of the five models,the overall distribution can be observed.The extremely high and high susceptibility areas are distributed in the Funiu Mountain range in the southwest,the Xiaoshan Mountain range in the west,and the Yellow River Basin in the north.These areas have large terrain fluctuations,complicated geological structural environments and frequent human engineering activities.The extremely high and highly prone areas were 12043.3 km^(2)and 3087.45 km^(2),accounting for 47.61%and 12.20%of the total area of the study area,respectively.Our study reflects the distribution of landslide susceptibility in western Henan Province,which provides a scientific basis for regional disaster warning,prediction,and resource protection.The study has important practical significance for subsequent landslide disaster management.