Reliable long-term settlement prediction of a high embankment relates to mountain infrastructure safety.This study developed a novel hybrid model(NHM)that combines a joint denoising technique with an enhanced gray wol...Reliable long-term settlement prediction of a high embankment relates to mountain infrastructure safety.This study developed a novel hybrid model(NHM)that combines a joint denoising technique with an enhanced gray wolf optimizer(EGWO)-n-support vector regression(n-SVR)method.High-embankment field measurements were preprocessed using the joint denoising technique,which in-cludes complete ensemble empirical mode decomposition,singular value decomposition,and wavelet packet transform.Furthermore,high-embankment settlements were predicted using the EGWO-n-SVR method.In this method,the standard gray wolf optimizer(GWO)was improved to obtain the EGWO to better tune the n-SVR model hyperparameters.The proposed NHM was then tested in two case studies.Finally,the influences of the data division ratio and kernel function on the EGWO-n-SVR forecasting performance and prediction efficiency were investigated.The results indicate that the NHM suppresses noise and restores details in high-embankment field measurements.Simultaneously,the NHM out-performs other alternative prediction methods in prediction accuracy and robustness.This demonstrates that the proposed NHM is effective in predicting high-embankment settlements with noisy field mea-surements.Moreover,the appropriate data division ratio and kernel function for EGWO-n-SVR are 7:3 and radial basis function,respectively.展开更多
When a customer uses the software, then it is possible to occur defects that can be removed in the updated versions of the software. Hence, in the present work, a robust examination of cross-project software defect pr...When a customer uses the software, then it is possible to occur defects that can be removed in the updated versions of the software. Hence, in the present work, a robust examination of cross-project software defect prediction is elaborated through an innovative hybrid machine learning framework. The proposed technique combines an advanced deep neural network architecture with ensemble models such as Support Vector Machine (SVM), Random Forest (RF), and XGBoost. The study evaluates the performance by considering multiple software projects like CM1, JM1, KC1, and PC1 using datasets from the PROMISE Software Engineering Repository. The three hybrid models that are compared are Hybrid Model-1 (SVM, RandomForest, XGBoost, Neural Network), Hybrid Model-2 (GradientBoosting, DecisionTree, LogisticRegression, Neural Network), and Hybrid Model-3 (KNeighbors, GaussianNB, Support Vector Classification (SVC), Neural Network), and the Hybrid Model 3 surpasses the others in terms of recall, F1-score, accuracy, ROC AUC, and precision. The presented work offers valuable insights into the effectiveness of hybrid techniques for cross-project defect prediction, providing a comparative perspective on early defect identification and mitigation strategies. .展开更多
BACKGROUND Postoperative delirium,particularly prevalent in elderly patients after abdominal cancer surgery,presents significant challenges in clinical management.AIM To develop a synthetic minority oversampling techn...BACKGROUND Postoperative delirium,particularly prevalent in elderly patients after abdominal cancer surgery,presents significant challenges in clinical management.AIM To develop a synthetic minority oversampling technique(SMOTE)-based model for predicting postoperative delirium in elderly abdominal cancer patients.METHODS In this retrospective cohort study,we analyzed data from 611 elderly patients who underwent abdominal malignant tumor surgery at our hospital between September 2020 and October 2022.The incidence of postoperative delirium was recorded for 7 d post-surgery.Patients were divided into delirium and non-delirium groups based on the occurrence of postoperative delirium or not.A multivariate logistic regression model was used to identify risk factors and develop a predictive model for postoperative delirium.The SMOTE technique was applied to enhance the model by oversampling the delirium cases.The model’s predictive accuracy was then validated.RESULTS In our study involving 611 elderly patients with abdominal malignant tumors,multivariate logistic regression analysis identified significant risk factors for postoperative delirium.These included the Charlson comorbidity index,American Society of Anesthesiologists classification,history of cerebrovascular disease,surgical duration,perioperative blood transfusion,and postoperative pain score.The incidence rate of postoperative delirium in our study was 22.91%.The original predictive model(P1)exhibited an area under the receiver operating characteristic curve of 0.862.In comparison,the SMOTE-based logistic early warning model(P2),which utilized the SMOTE oversampling algorithm,showed a slightly lower but comparable area under the curve of 0.856,suggesting no significant difference in performance between the two predictive approaches.CONCLUSION This study confirms that the SMOTE-enhanced predictive model for postoperative delirium in elderly abdominal tumor patients shows performance equivalent to that of traditional methods,effectively addressing data imbalance.展开更多
The development of defect prediction plays a significant role in improving software quality. Such predictions are used to identify defective modules before the testing and to minimize the time and cost. The software w...The development of defect prediction plays a significant role in improving software quality. Such predictions are used to identify defective modules before the testing and to minimize the time and cost. The software with defects negatively impacts operational costs and finally affects customer satisfaction. Numerous approaches exist to predict software defects. However, the timely and accurate software bugs are the major challenging issues. To improve the timely and accurate software defect prediction, a novel technique called Nonparametric Statistical feature scaled QuAdratic regressive convolution Deep nEural Network (SQADEN) is introduced. The proposed SQADEN technique mainly includes two major processes namely metric or feature selection and classification. First, the SQADEN uses the nonparametric statistical Torgerson–Gower scaling technique for identifying the relevant software metrics by measuring the similarity using the dice coefficient. The feature selection process is used to minimize the time complexity of software fault prediction. With the selected metrics, software fault perdition with the help of the Quadratic Censored regressive convolution deep neural network-based classification. The deep learning classifier analyzes the training and testing samples using the contingency correlation coefficient. The softstep activation function is used to provide the final fault prediction results. To minimize the error, the Nelder–Mead method is applied to solve non-linear least-squares problems. Finally, accurate classification results with a minimum error are obtained at the output layer. Experimental evaluation is carried out with different quantitative metrics such as accuracy, precision, recall, F-measure, and time complexity. The analyzed results demonstrate the superior performance of our proposed SQADEN technique with maximum accuracy, sensitivity and specificity by 3%, 3%, 2% and 3% and minimum time and space by 13% and 15% when compared with the two state-of-the-art methods.展开更多
The wave-based method (WBM) has been applied for the prediction of mid-frequency vibrations of fiat plates. The scaling factors, Gauss point selection rule and truncation rule are introduced to insure the wave model...The wave-based method (WBM) has been applied for the prediction of mid-frequency vibrations of fiat plates. The scaling factors, Gauss point selection rule and truncation rule are introduced to insure the wave model to converge. Numerical results show that the prediction tech- nique based on WBM is with higher accuracy and smaller computational effort than the one on FEM, which implies that this new technique on WBM can be applied to higher-frequency range.展开更多
Maintaining software reliability is the key idea for conducting quality research.This can be done by having less complex applications.While developers and other experts have made signicant efforts in this context,the ...Maintaining software reliability is the key idea for conducting quality research.This can be done by having less complex applications.While developers and other experts have made signicant efforts in this context,the level of reliability is not the same as it should be.Therefore,further research into the most detailed mechanisms for evaluating and increasing software reliability is essential.A signicant aspect of growing the degree of reliable applications is the quantitative assessment of reliability.There are multiple statistical as well as soft computing methods available in literature for predicting reliability of software.However,none of these mechanisms are useful for all kinds of failure datasets and applications.Hence nding the most optimal model for reliability prediction is an important concern.This paper suggests a novel method to substantially pick the best model of reliability prediction.This method is the combination of analytic hierarchy method(AHP),hesitant fuzzy(HF)sets and technique for order of preference by similarity to ideal solution(TOPSIS).In addition,using the different iterations of the process,procedural sensitivity was also performed to validate the ndings.The ndings of the software reliability prediction models prioritization will help the developers to estimate reliability prediction based on the software type.展开更多
Decreasing the risks and geohazards associated with drilling engineering in high-temperature high-pressure(HTHP) geologic settings begins with the implementation of pre-drilling prediction techniques(PPTs). To improve...Decreasing the risks and geohazards associated with drilling engineering in high-temperature high-pressure(HTHP) geologic settings begins with the implementation of pre-drilling prediction techniques(PPTs). To improve the accuracy of geopressure prediction in HTHP hydrocarbon reservoirs offshore Hainan Island, we made a comprehensive summary of current PPTs to identify existing problems and challenges by analyzing the global distribution of HTHP hydrocarbon reservoirs, the research status of PPTs, and the geologic setting and its HTHP formation mechanism. Our research results indicate that the HTHP formation mechanism in the study area is caused by multiple factors, including rapid loading, diapir intrusions, hydrocarbon generation, and the thermal expansion of pore fluids. Due to this multi-factor interaction, a cloud of HTHP hydrocarbon reservoirs has developed in the Ying-Qiong Basin, but only traditional PPTs have been implemented, based on the assumption of conditions that do not conform to the actual geologic environment, e.g., Bellotti's law and Eaton's law. In this paper, we focus on these issues, identify some challenges and solutions, and call for further PPT research to address the drawbacks of previous works and meet the challenges associated with the deepwater technology gap. In this way, we hope to contribute to the improved accuracy of geopressure prediction prior to drilling and provide support for future HTHP drilling offshore Hainan Island.展开更多
1 Introduction As new exploration domain for oil and gas,reservoirs with low porosity and low permeability have become a hotspot in recent years(Li Daopin,1997).With the improvement of technology,low porosity and low
This article presented a new data fusion approach for reasonably predicting dynamic serviceability reliability of the long-span bridge girder.Firstly,multivariate Bayesian dynamic linear model(MBDLM)considering dynami...This article presented a new data fusion approach for reasonably predicting dynamic serviceability reliability of the long-span bridge girder.Firstly,multivariate Bayesian dynamic linear model(MBDLM)considering dynamic correlation among the multiple variables is provided to predict dynamic extreme deflections;secondly,with the proposed MBDLM,the dynamic correlation coefficients between any two performance functions can be predicted;finally,based on MBDLM and Gaussian copula technique,a new data fusion method is given to predict the serviceability reliability of the long-span bridge girder,and the monitoring extreme deflection data from an actual bridge is provided to illustrated the feasibility and application of the proposed method.展开更多
Short-term traffic flow prediction is one of the essential issues in intelligent transportation systems(ITS). A new two-stage traffic flow prediction method named AKNN-AVL method is presented, which combines an advanc...Short-term traffic flow prediction is one of the essential issues in intelligent transportation systems(ITS). A new two-stage traffic flow prediction method named AKNN-AVL method is presented, which combines an advanced k-nearest neighbor(AKNN)method and balanced binary tree(AVL) data structure to improve the prediction accuracy. The AKNN method uses pattern recognition two times in the searching process, which considers the previous sequences of traffic flow to forecast the future traffic state. Clustering method and balanced binary tree technique are introduced to build case database to reduce the searching time. To illustrate the effects of these developments, the accuracies performance of AKNN-AVL method, k-nearest neighbor(KNN) method and the auto-regressive and moving average(ARMA) method are compared. These methods are calibrated and evaluated by the real-time data from a freeway traffic detector near North 3rd Ring Road in Beijing under both normal and incident traffic conditions.The comparisons show that the AKNN-AVL method with the optimal neighbor and pattern size outperforms both KNN method and ARMA method under both normal and incident traffic conditions. In addition, the combinations of clustering method and balanced binary tree technique to the prediction method can increase the searching speed and respond rapidly to case database fluctuations.展开更多
Energy is essential to practically all exercises and is imperative for the development of personal satisfaction.So,valuable energy has been in great demand for many years,especially for using smart homes and structure...Energy is essential to practically all exercises and is imperative for the development of personal satisfaction.So,valuable energy has been in great demand for many years,especially for using smart homes and structures,as individuals quickly improve their way of life depending on current innovations.However,there is a shortage of energy,as the energy required is higher than that produced.Many new plans are being designed to meet the consumer’s energy requirements.In many regions,energy utilization in the housing area is 30%–40%.The growth of smart homes has raised the requirement for intelligence in applications such as asset management,energy-efficient automation,security,and healthcare monitoring to learn about residents’actions and forecast their future demands.To overcome the challenges of energy consumption optimization,in this study,we apply an energy management technique.Data fusion has recently attracted much energy efficiency in buildings,where numerous types of information are processed.The proposed research developed a data fusion model to predict energy consumption for accuracy and miss rate.The results of the proposed approach are compared with those of the previously published techniques and found that the prediction accuracy of the proposed method is 92%,which is higher than the previously published approaches.展开更多
In this study, the author will investigate and utilize advanced machine learning models related to two different methodologies to determine the best and most effective way to predict individuals with heart failure and...In this study, the author will investigate and utilize advanced machine learning models related to two different methodologies to determine the best and most effective way to predict individuals with heart failure and cardiovascular diseases. The first methodology involves a list of classification machine learning algorithms, and the second methodology involves the use of a deep learning algorithm known as MLP or Multilayer Perceptrons. Globally, hospitals are dealing with cases related to cardiovascular diseases and heart failure as they are major causes of death, not only for overweight individuals but also for those who do not adopt a healthy diet and lifestyle. Often, heart failures and cardiovascular diseases can be caused by many factors, including cardiomyopathy, high blood pressure, coronary heart disease, and heart inflammation [1]. Other factors, such as irregular shocks or stress, can also contribute to heart failure or a heart attack. While these events cannot be predicted, continuous data from patients’ health can help doctors predict heart failure. Therefore, this data-driven research utilizes advanced machine learning and deep learning techniques to better analyze and manipulate the data, providing doctors with informative decision-making tools regarding a person’s likelihood of experiencing heart failure. In this paper, the author employed advanced data preprocessing and cleaning techniques. Additionally, the dataset underwent testing using two different methodologies to determine the most effective machine-learning technique for producing optimal predictions. The first methodology involved employing a list of supervised classification machine learning algorithms, including Naïve Bayes (NB), KNN, logistic regression, and the SVM algorithm. The second methodology utilized a deep learning (DL) algorithm known as Multilayer Perceptrons (MLPs). This algorithm provided the author with the flexibility to experiment with different layer sizes and activation functions, such as ReLU, logistic (sigmoid), and Tanh. Both methodologies produced optimal models with high-level accuracy rates. The first methodology involves a list of supervised machine learning algorithms, including KNN, SVM, Adaboost, Logistic Regression, Naive Bayes, and Decision Tree algorithms. They achieved accuracy rates of 86%, 89%, 89%, 81%, 79%, and 99%, respectively. The author clearly explained that Decision Tree algorithm is not suitable for the dataset at hand due to overfitting issues. Therefore, it was discarded as an optimal model to be used. However, the latter methodology (Neural Network) demonstrated the most stable and optimal accuracy, achieving over 87% accuracy while adapting well to real-life situations and requiring low computing power overall. A performance assessment and evaluation were carried out based on a confusion matrix report to demonstrate feasibility and performance. The author concluded that the performance of the model in real-life situations can advance not only the medical field of science but also mathematical concepts. Additionally, the advanced preprocessing approach behind the model can provide value to the Data Science community. The model can be further developed by employing various optimization techniques to handle even larger datasets related to heart failures. Furthermore, different neural network algorithms can be tested to explore alternative approaches and yield different results.展开更多
Recent advances in dynamical climate prediction at the Institute of Atmospheric Physics, Chinese Academy of Sciences (IAP/CAS) during the last five years have been briefly described in this paper. Firstly, the second ...Recent advances in dynamical climate prediction at the Institute of Atmospheric Physics, Chinese Academy of Sciences (IAP/CAS) during the last five years have been briefly described in this paper. Firstly, the second generation of the IAP dynamical climate prediction system (IAP DCP-Ⅱ) has been described, and two sets of hindcast experiments of the summer rainfall anomalies over China for the periods of 1980-1994 with different versions of the IAP AGCM have been conducted. The comparison results show that the predictive skill of summer rainfall anomalies over China is improved with the improved IAP AGCM in which the surface albedo parameterization is modified. Furthermore, IAP DCP-II has been applied to the real-time prediction of summer rainfall anomalies over China since 1998, and the verification results show that IAP DCP-II can quite well capture the large scale patterns of the summer flood/drought situations over China during the last five years (1998-2002). Meanwhile, an investigation has demonstrated the importance of the atmospheric initial conditions on the seasonal climate prediction, along with studies on the influences from surface boundary conditions (e.g., land surface characteristics, sea surface temperature). Certain conclusions have been reached, such as, the initial atmospheric anomalies in spring may play an important role in the summer climate anomalies, and soil moisture anomalies in spring can also have a significant impact on the summer climate anomalies over East Asia. Finally, several practical techniques (e.g., ensemble technique, correction method, etc.), which lead to the increase of the prediction skill for summer rainfall anomalies over China, have also been illustrated. The paper concludes with a list of critical requirements needed for the further improvement of dynamical seasonal climate prediction.展开更多
Many business applications rely on their historical data to predict their business future. The marketing products process is one of the core processes for the business. Customer needs give a useful piece of informatio...Many business applications rely on their historical data to predict their business future. The marketing products process is one of the core processes for the business. Customer needs give a useful piece of information that help</span><span style="font-family:Verdana;"><span style="font-family:Verdana;">s</span></span><span style="font-family:Verdana;"> to market the appropriate products at the appropriate time. Moreover, services are considered recently as products. The development of education and health services </span><span style="font-family:Verdana;"><span style="font-family:Verdana;">is</span></span><span style="font-family:Verdana;"> depending on historical data. For the more, reducing online social media networks problems and crimes need a significant source of information. Data analysts need to use an efficient classification algorithm to predict the future of such businesses. However, dealing with a huge quantity of data requires great time to process. Data mining involves many useful techniques that are used to predict statistical data in a variety of business applications. The classification technique is one of the most widely used with a variety of algorithms. In this paper, various classification algorithms are revised in terms of accuracy in different areas of data mining applications. A comprehensive analysis is made after delegated reading of 20 papers in the literature. This paper aims to help data analysts to choose the most suitable classification algorithm for different business applications including business in general, online social media networks, agriculture, health, and education. Results show FFBPN is the most accurate algorithm in the business domain. The Random Forest algorithm is the most accurate in classifying online social networks (OSN) activities. Na<span style="white-space:nowrap;">ï</span>ve Bayes algorithm is the most accurate to classify agriculture datasets. OneR is the most accurate algorithm to classify instances within the health domain. The C4.5 Decision Tree algorithm is the most accurate to classify students’ records to predict degree completion time.展开更多
Based on experimental data of line heating, the methods of vector mapping, plane projection, and coordinate converting are presented to establish the spectra for line heating distortion discipline which shows the rela...Based on experimental data of line heating, the methods of vector mapping, plane projection, and coordinate converting are presented to establish the spectra for line heating distortion discipline which shows the relationship between process parameters and distortion parameters of line heating. Back-propagation network (BP-net) is used to modify the spectra.Mathematical models for optimizing line heating techniques parameters, which include two-objective functions,are constructed. To convert the multi-objective optimization into a single-objective one, the method of changing weight coefficient is used, and then the individual fitness function is built up. Taking the number of heating lines, distance between the heating lines’ border (line space), and shrink quantity of lines as three restrictive conditions,a hierarchy genetic algorithm (HGA) code is established by making use of information provided by the spectra, in which inner coding and outer coding adopt different heredity arithmetic operators in inherent operating. The numerical example shows that the spectra for line heating distortion discipline presented here can provide accurate information required by techniques parameter prediction of line heating process and the technique parameter optimization method based on HGA provided here can obtain good results for hull plate.展开更多
A method for predicting the Gibbs free energies of intermediate compounds in a binary system has been presented, based upon the regulations of the Gibbs free energies of formation of intermediate compounds in the syst...A method for predicting the Gibbs free energies of intermediate compounds in a binary system has been presented, based upon the regulations of the Gibbs free energies of formation of intermediate compounds in the system. The application of this procedure to the V-O system demonstrates that this method is feasible.展开更多
Reliable and accurate ultra-short-term prediction of wind power is vital for the operation and optimization of power systems.However,the volatility and intermittence of wind power pose uncertainties to traditional poi...Reliable and accurate ultra-short-term prediction of wind power is vital for the operation and optimization of power systems.However,the volatility and intermittence of wind power pose uncertainties to traditional point prediction,resulting in an increased risk of power system operation.To represent the uncertainty of wind power,this paper proposes a new method for ultra-short-term interval prediction of wind power based on a graph neural network(GNN)and an improved Bootstrap technique.Specifically,adjacent wind farms and local meteorological factors are modeled as the new form of a graph from the graph-theoretic perspective.Then,the graph convolutional network(GCN)and bi-directional long short-term memory(Bi-LSTM)are proposed to capture spatiotemporal features between nodes in the graph.To obtain highquality prediction intervals(PIs),an improved Bootstrap technique is designed to increase coverage percentage and narrow PIs effectively.Numerical simulations demonstrate that the proposed method can capture the spatiotemporal correlations from the graph,and the prediction results outperform popular baselines on two real-world datasets,which implies a high potential for practical applications in power systems.展开更多
基金We acknowledge the funding support from the National Natural Science Foundation of China(Grant No.51808462)the Natural Science Foundation Project of Sichuan Province,China(Grant No.2023NSFSC0346)the Science and Technology Project of Inner Mongolia Transportation Department,China(Grant No.NJ-2022-14).
文摘Reliable long-term settlement prediction of a high embankment relates to mountain infrastructure safety.This study developed a novel hybrid model(NHM)that combines a joint denoising technique with an enhanced gray wolf optimizer(EGWO)-n-support vector regression(n-SVR)method.High-embankment field measurements were preprocessed using the joint denoising technique,which in-cludes complete ensemble empirical mode decomposition,singular value decomposition,and wavelet packet transform.Furthermore,high-embankment settlements were predicted using the EGWO-n-SVR method.In this method,the standard gray wolf optimizer(GWO)was improved to obtain the EGWO to better tune the n-SVR model hyperparameters.The proposed NHM was then tested in two case studies.Finally,the influences of the data division ratio and kernel function on the EGWO-n-SVR forecasting performance and prediction efficiency were investigated.The results indicate that the NHM suppresses noise and restores details in high-embankment field measurements.Simultaneously,the NHM out-performs other alternative prediction methods in prediction accuracy and robustness.This demonstrates that the proposed NHM is effective in predicting high-embankment settlements with noisy field mea-surements.Moreover,the appropriate data division ratio and kernel function for EGWO-n-SVR are 7:3 and radial basis function,respectively.
文摘When a customer uses the software, then it is possible to occur defects that can be removed in the updated versions of the software. Hence, in the present work, a robust examination of cross-project software defect prediction is elaborated through an innovative hybrid machine learning framework. The proposed technique combines an advanced deep neural network architecture with ensemble models such as Support Vector Machine (SVM), Random Forest (RF), and XGBoost. The study evaluates the performance by considering multiple software projects like CM1, JM1, KC1, and PC1 using datasets from the PROMISE Software Engineering Repository. The three hybrid models that are compared are Hybrid Model-1 (SVM, RandomForest, XGBoost, Neural Network), Hybrid Model-2 (GradientBoosting, DecisionTree, LogisticRegression, Neural Network), and Hybrid Model-3 (KNeighbors, GaussianNB, Support Vector Classification (SVC), Neural Network), and the Hybrid Model 3 surpasses the others in terms of recall, F1-score, accuracy, ROC AUC, and precision. The presented work offers valuable insights into the effectiveness of hybrid techniques for cross-project defect prediction, providing a comparative perspective on early defect identification and mitigation strategies. .
基金Supported by Discipline Advancement Program of Shanghai Fourth People’s Hospital,No.SY-XKZT-2020-2013.
文摘BACKGROUND Postoperative delirium,particularly prevalent in elderly patients after abdominal cancer surgery,presents significant challenges in clinical management.AIM To develop a synthetic minority oversampling technique(SMOTE)-based model for predicting postoperative delirium in elderly abdominal cancer patients.METHODS In this retrospective cohort study,we analyzed data from 611 elderly patients who underwent abdominal malignant tumor surgery at our hospital between September 2020 and October 2022.The incidence of postoperative delirium was recorded for 7 d post-surgery.Patients were divided into delirium and non-delirium groups based on the occurrence of postoperative delirium or not.A multivariate logistic regression model was used to identify risk factors and develop a predictive model for postoperative delirium.The SMOTE technique was applied to enhance the model by oversampling the delirium cases.The model’s predictive accuracy was then validated.RESULTS In our study involving 611 elderly patients with abdominal malignant tumors,multivariate logistic regression analysis identified significant risk factors for postoperative delirium.These included the Charlson comorbidity index,American Society of Anesthesiologists classification,history of cerebrovascular disease,surgical duration,perioperative blood transfusion,and postoperative pain score.The incidence rate of postoperative delirium in our study was 22.91%.The original predictive model(P1)exhibited an area under the receiver operating characteristic curve of 0.862.In comparison,the SMOTE-based logistic early warning model(P2),which utilized the SMOTE oversampling algorithm,showed a slightly lower but comparable area under the curve of 0.856,suggesting no significant difference in performance between the two predictive approaches.CONCLUSION This study confirms that the SMOTE-enhanced predictive model for postoperative delirium in elderly abdominal tumor patients shows performance equivalent to that of traditional methods,effectively addressing data imbalance.
文摘The development of defect prediction plays a significant role in improving software quality. Such predictions are used to identify defective modules before the testing and to minimize the time and cost. The software with defects negatively impacts operational costs and finally affects customer satisfaction. Numerous approaches exist to predict software defects. However, the timely and accurate software bugs are the major challenging issues. To improve the timely and accurate software defect prediction, a novel technique called Nonparametric Statistical feature scaled QuAdratic regressive convolution Deep nEural Network (SQADEN) is introduced. The proposed SQADEN technique mainly includes two major processes namely metric or feature selection and classification. First, the SQADEN uses the nonparametric statistical Torgerson–Gower scaling technique for identifying the relevant software metrics by measuring the similarity using the dice coefficient. The feature selection process is used to minimize the time complexity of software fault prediction. With the selected metrics, software fault perdition with the help of the Quadratic Censored regressive convolution deep neural network-based classification. The deep learning classifier analyzes the training and testing samples using the contingency correlation coefficient. The softstep activation function is used to provide the final fault prediction results. To minimize the error, the Nelder–Mead method is applied to solve non-linear least-squares problems. Finally, accurate classification results with a minimum error are obtained at the output layer. Experimental evaluation is carried out with different quantitative metrics such as accuracy, precision, recall, F-measure, and time complexity. The analyzed results demonstrate the superior performance of our proposed SQADEN technique with maximum accuracy, sensitivity and specificity by 3%, 3%, 2% and 3% and minimum time and space by 13% and 15% when compared with the two state-of-the-art methods.
基金Project supported by the National Natural Science Foundation of China (No.10472035).
文摘The wave-based method (WBM) has been applied for the prediction of mid-frequency vibrations of fiat plates. The scaling factors, Gauss point selection rule and truncation rule are introduced to insure the wave model to converge. Numerical results show that the prediction tech- nique based on WBM is with higher accuracy and smaller computational effort than the one on FEM, which implies that this new technique on WBM can be applied to higher-frequency range.
基金funded by Grant No.12-INF2970-10 from the National Science,Technology and Innovation Plan(MAARIFAH)the King Abdul-Aziz City for Science and Technology(KACST)Kingdom of Saudi Arabia.
文摘Maintaining software reliability is the key idea for conducting quality research.This can be done by having less complex applications.While developers and other experts have made signicant efforts in this context,the level of reliability is not the same as it should be.Therefore,further research into the most detailed mechanisms for evaluating and increasing software reliability is essential.A signicant aspect of growing the degree of reliable applications is the quantitative assessment of reliability.There are multiple statistical as well as soft computing methods available in literature for predicting reliability of software.However,none of these mechanisms are useful for all kinds of failure datasets and applications.Hence nding the most optimal model for reliability prediction is an important concern.This paper suggests a novel method to substantially pick the best model of reliability prediction.This method is the combination of analytic hierarchy method(AHP),hesitant fuzzy(HF)sets and technique for order of preference by similarity to ideal solution(TOPSIS).In addition,using the different iterations of the process,procedural sensitivity was also performed to validate the ndings.The ndings of the software reliability prediction models prioritization will help the developers to estimate reliability prediction based on the software type.
基金funded by the National Basic Research Program of China (No. 2015CB251201)the NSFC-Shandong Joint Fund for Marine Science Research Centers (No. U1606401)+3 种基金the Scientific and Technological Innovation Project financially supported by Qingdao National Laboratory for Marine Science and Technology (No. 2016ASKJ13)the Major National Science and Technology Programs (No. 016ZX05024-001-002)the Natural Science Foundation of Hainan (No. ZDYF2016215)Key Science and Technology Foundation of Sanya (Nos. 2017PT13, 2017PT2014)
文摘Decreasing the risks and geohazards associated with drilling engineering in high-temperature high-pressure(HTHP) geologic settings begins with the implementation of pre-drilling prediction techniques(PPTs). To improve the accuracy of geopressure prediction in HTHP hydrocarbon reservoirs offshore Hainan Island, we made a comprehensive summary of current PPTs to identify existing problems and challenges by analyzing the global distribution of HTHP hydrocarbon reservoirs, the research status of PPTs, and the geologic setting and its HTHP formation mechanism. Our research results indicate that the HTHP formation mechanism in the study area is caused by multiple factors, including rapid loading, diapir intrusions, hydrocarbon generation, and the thermal expansion of pore fluids. Due to this multi-factor interaction, a cloud of HTHP hydrocarbon reservoirs has developed in the Ying-Qiong Basin, but only traditional PPTs have been implemented, based on the assumption of conditions that do not conform to the actual geologic environment, e.g., Bellotti's law and Eaton's law. In this paper, we focus on these issues, identify some challenges and solutions, and call for further PPT research to address the drawbacks of previous works and meet the challenges associated with the deepwater technology gap. In this way, we hope to contribute to the improved accuracy of geopressure prediction prior to drilling and provide support for future HTHP drilling offshore Hainan Island.
基金funded by Major Projects of National Science and Technology "Large Oil and Gas Fields and CBM development"(Grant No. 2016ZX05027)
文摘1 Introduction As new exploration domain for oil and gas,reservoirs with low porosity and low permeability have become a hotspot in recent years(Li Daopin,1997).With the improvement of technology,low porosity and low
基金This work was supported by Natural Science Foundation of Gansu Province of China(20JR10RA625,20JR10RA623)National Key Research and Development Project of China(Project No.2019YFC1511005)+1 种基金Fundamental Research Funds for the Central Universities(Grant No.lzujbky-2020-55)National Natural Science Foundation of China(Grant No.51608243).
文摘This article presented a new data fusion approach for reasonably predicting dynamic serviceability reliability of the long-span bridge girder.Firstly,multivariate Bayesian dynamic linear model(MBDLM)considering dynamic correlation among the multiple variables is provided to predict dynamic extreme deflections;secondly,with the proposed MBDLM,the dynamic correlation coefficients between any two performance functions can be predicted;finally,based on MBDLM and Gaussian copula technique,a new data fusion method is given to predict the serviceability reliability of the long-span bridge girder,and the monitoring extreme deflection data from an actual bridge is provided to illustrated the feasibility and application of the proposed method.
基金Project(2012CB725403)supported by the National Basic Research Program of ChinaProjects(71210001,51338008)supported by the National Natural Science Foundation of ChinaProject supported by World Capital Cities Smooth Traffic Collaborative Innovation Center and Singapore National Research Foundation Under Its Campus for Research Excellence and Technology Enterprise(CREATE)Programme
文摘Short-term traffic flow prediction is one of the essential issues in intelligent transportation systems(ITS). A new two-stage traffic flow prediction method named AKNN-AVL method is presented, which combines an advanced k-nearest neighbor(AKNN)method and balanced binary tree(AVL) data structure to improve the prediction accuracy. The AKNN method uses pattern recognition two times in the searching process, which considers the previous sequences of traffic flow to forecast the future traffic state. Clustering method and balanced binary tree technique are introduced to build case database to reduce the searching time. To illustrate the effects of these developments, the accuracies performance of AKNN-AVL method, k-nearest neighbor(KNN) method and the auto-regressive and moving average(ARMA) method are compared. These methods are calibrated and evaluated by the real-time data from a freeway traffic detector near North 3rd Ring Road in Beijing under both normal and incident traffic conditions.The comparisons show that the AKNN-AVL method with the optimal neighbor and pattern size outperforms both KNN method and ARMA method under both normal and incident traffic conditions. In addition, the combinations of clustering method and balanced binary tree technique to the prediction method can increase the searching speed and respond rapidly to case database fluctuations.
文摘Energy is essential to practically all exercises and is imperative for the development of personal satisfaction.So,valuable energy has been in great demand for many years,especially for using smart homes and structures,as individuals quickly improve their way of life depending on current innovations.However,there is a shortage of energy,as the energy required is higher than that produced.Many new plans are being designed to meet the consumer’s energy requirements.In many regions,energy utilization in the housing area is 30%–40%.The growth of smart homes has raised the requirement for intelligence in applications such as asset management,energy-efficient automation,security,and healthcare monitoring to learn about residents’actions and forecast their future demands.To overcome the challenges of energy consumption optimization,in this study,we apply an energy management technique.Data fusion has recently attracted much energy efficiency in buildings,where numerous types of information are processed.The proposed research developed a data fusion model to predict energy consumption for accuracy and miss rate.The results of the proposed approach are compared with those of the previously published techniques and found that the prediction accuracy of the proposed method is 92%,which is higher than the previously published approaches.
文摘In this study, the author will investigate and utilize advanced machine learning models related to two different methodologies to determine the best and most effective way to predict individuals with heart failure and cardiovascular diseases. The first methodology involves a list of classification machine learning algorithms, and the second methodology involves the use of a deep learning algorithm known as MLP or Multilayer Perceptrons. Globally, hospitals are dealing with cases related to cardiovascular diseases and heart failure as they are major causes of death, not only for overweight individuals but also for those who do not adopt a healthy diet and lifestyle. Often, heart failures and cardiovascular diseases can be caused by many factors, including cardiomyopathy, high blood pressure, coronary heart disease, and heart inflammation [1]. Other factors, such as irregular shocks or stress, can also contribute to heart failure or a heart attack. While these events cannot be predicted, continuous data from patients’ health can help doctors predict heart failure. Therefore, this data-driven research utilizes advanced machine learning and deep learning techniques to better analyze and manipulate the data, providing doctors with informative decision-making tools regarding a person’s likelihood of experiencing heart failure. In this paper, the author employed advanced data preprocessing and cleaning techniques. Additionally, the dataset underwent testing using two different methodologies to determine the most effective machine-learning technique for producing optimal predictions. The first methodology involved employing a list of supervised classification machine learning algorithms, including Naïve Bayes (NB), KNN, logistic regression, and the SVM algorithm. The second methodology utilized a deep learning (DL) algorithm known as Multilayer Perceptrons (MLPs). This algorithm provided the author with the flexibility to experiment with different layer sizes and activation functions, such as ReLU, logistic (sigmoid), and Tanh. Both methodologies produced optimal models with high-level accuracy rates. The first methodology involves a list of supervised machine learning algorithms, including KNN, SVM, Adaboost, Logistic Regression, Naive Bayes, and Decision Tree algorithms. They achieved accuracy rates of 86%, 89%, 89%, 81%, 79%, and 99%, respectively. The author clearly explained that Decision Tree algorithm is not suitable for the dataset at hand due to overfitting issues. Therefore, it was discarded as an optimal model to be used. However, the latter methodology (Neural Network) demonstrated the most stable and optimal accuracy, achieving over 87% accuracy while adapting well to real-life situations and requiring low computing power overall. A performance assessment and evaluation were carried out based on a confusion matrix report to demonstrate feasibility and performance. The author concluded that the performance of the model in real-life situations can advance not only the medical field of science but also mathematical concepts. Additionally, the advanced preprocessing approach behind the model can provide value to the Data Science community. The model can be further developed by employing various optimization techniques to handle even larger datasets related to heart failures. Furthermore, different neural network algorithms can be tested to explore alternative approaches and yield different results.
基金supported by the Key P roject of the National N atural Science Foundation of China(Grant Nos:40233027 and 40221503)the Key Project of the Chinese Academy of Sciences(KZCX2-203)the IAP/CAS Knowledge Innovation Project.
文摘Recent advances in dynamical climate prediction at the Institute of Atmospheric Physics, Chinese Academy of Sciences (IAP/CAS) during the last five years have been briefly described in this paper. Firstly, the second generation of the IAP dynamical climate prediction system (IAP DCP-Ⅱ) has been described, and two sets of hindcast experiments of the summer rainfall anomalies over China for the periods of 1980-1994 with different versions of the IAP AGCM have been conducted. The comparison results show that the predictive skill of summer rainfall anomalies over China is improved with the improved IAP AGCM in which the surface albedo parameterization is modified. Furthermore, IAP DCP-II has been applied to the real-time prediction of summer rainfall anomalies over China since 1998, and the verification results show that IAP DCP-II can quite well capture the large scale patterns of the summer flood/drought situations over China during the last five years (1998-2002). Meanwhile, an investigation has demonstrated the importance of the atmospheric initial conditions on the seasonal climate prediction, along with studies on the influences from surface boundary conditions (e.g., land surface characteristics, sea surface temperature). Certain conclusions have been reached, such as, the initial atmospheric anomalies in spring may play an important role in the summer climate anomalies, and soil moisture anomalies in spring can also have a significant impact on the summer climate anomalies over East Asia. Finally, several practical techniques (e.g., ensemble technique, correction method, etc.), which lead to the increase of the prediction skill for summer rainfall anomalies over China, have also been illustrated. The paper concludes with a list of critical requirements needed for the further improvement of dynamical seasonal climate prediction.
文摘Many business applications rely on their historical data to predict their business future. The marketing products process is one of the core processes for the business. Customer needs give a useful piece of information that help</span><span style="font-family:Verdana;"><span style="font-family:Verdana;">s</span></span><span style="font-family:Verdana;"> to market the appropriate products at the appropriate time. Moreover, services are considered recently as products. The development of education and health services </span><span style="font-family:Verdana;"><span style="font-family:Verdana;">is</span></span><span style="font-family:Verdana;"> depending on historical data. For the more, reducing online social media networks problems and crimes need a significant source of information. Data analysts need to use an efficient classification algorithm to predict the future of such businesses. However, dealing with a huge quantity of data requires great time to process. Data mining involves many useful techniques that are used to predict statistical data in a variety of business applications. The classification technique is one of the most widely used with a variety of algorithms. In this paper, various classification algorithms are revised in terms of accuracy in different areas of data mining applications. A comprehensive analysis is made after delegated reading of 20 papers in the literature. This paper aims to help data analysts to choose the most suitable classification algorithm for different business applications including business in general, online social media networks, agriculture, health, and education. Results show FFBPN is the most accurate algorithm in the business domain. The Random Forest algorithm is the most accurate in classifying online social networks (OSN) activities. Na<span style="white-space:nowrap;">ï</span>ve Bayes algorithm is the most accurate to classify agriculture datasets. OneR is the most accurate algorithm to classify instances within the health domain. The C4.5 Decision Tree algorithm is the most accurate to classify students’ records to predict degree completion time.
基金funded by the Department for Environment Food & Rural Affairsthe Scottish Government+2 种基金the Department of Agriculture and Rural Development for Northern Irelandthe Welsh Government as part of the UK’s Agricultural GHG Research Platform initiative
文摘Based on experimental data of line heating, the methods of vector mapping, plane projection, and coordinate converting are presented to establish the spectra for line heating distortion discipline which shows the relationship between process parameters and distortion parameters of line heating. Back-propagation network (BP-net) is used to modify the spectra.Mathematical models for optimizing line heating techniques parameters, which include two-objective functions,are constructed. To convert the multi-objective optimization into a single-objective one, the method of changing weight coefficient is used, and then the individual fitness function is built up. Taking the number of heating lines, distance between the heating lines’ border (line space), and shrink quantity of lines as three restrictive conditions,a hierarchy genetic algorithm (HGA) code is established by making use of information provided by the spectra, in which inner coding and outer coding adopt different heredity arithmetic operators in inherent operating. The numerical example shows that the spectra for line heating distortion discipline presented here can provide accurate information required by techniques parameter prediction of line heating process and the technique parameter optimization method based on HGA provided here can obtain good results for hull plate.
文摘A method for predicting the Gibbs free energies of intermediate compounds in a binary system has been presented, based upon the regulations of the Gibbs free energies of formation of intermediate compounds in the system. The application of this procedure to the V-O system demonstrates that this method is feasible.
文摘Reliable and accurate ultra-short-term prediction of wind power is vital for the operation and optimization of power systems.However,the volatility and intermittence of wind power pose uncertainties to traditional point prediction,resulting in an increased risk of power system operation.To represent the uncertainty of wind power,this paper proposes a new method for ultra-short-term interval prediction of wind power based on a graph neural network(GNN)and an improved Bootstrap technique.Specifically,adjacent wind farms and local meteorological factors are modeled as the new form of a graph from the graph-theoretic perspective.Then,the graph convolutional network(GCN)and bi-directional long short-term memory(Bi-LSTM)are proposed to capture spatiotemporal features between nodes in the graph.To obtain highquality prediction intervals(PIs),an improved Bootstrap technique is designed to increase coverage percentage and narrow PIs effectively.Numerical simulations demonstrate that the proposed method can capture the spatiotemporal correlations from the graph,and the prediction results outperform popular baselines on two real-world datasets,which implies a high potential for practical applications in power systems.