The reasonable quantification of the concrete freezing environment on the Qinghai–Tibet Plateau(QTP) is the primary issue in frost resistant concrete design, which is one of the challenges that the QTP engineering ma...The reasonable quantification of the concrete freezing environment on the Qinghai–Tibet Plateau(QTP) is the primary issue in frost resistant concrete design, which is one of the challenges that the QTP engineering managers should take into account. In this paper, we propose a more realistic method to calculate the number of concrete freeze–thaw cycles(NFTCs) on the QTP. The calculated results show that the NFTCs increase as the altitude of the meteorological station increases with the average NFTCs being 208.7. Four machine learning methods, i.e., the random forest(RF) model, generalized boosting method(GBM), generalized linear model(GLM), and generalized additive model(GAM), are used to fit the NFTCs. The root mean square error(RMSE) values of the RF, GBM, GLM, and GAM are 32.3, 4.3, 247.9, and 161.3, respectively. The R^(2) values of the RF, GBM, GLM, and GAM are 0.93, 0.99, 0.48, and 0.66, respectively. The GBM method performs the best compared to the other three methods, which was shown by the results of RMSE and R^(2) values. The quantitative results from the GBM method indicate that the lowest, medium, and highest NFTC values are distributed in the northern, central, and southern parts of the QTP, respectively. The annual NFTCs in the QTP region are mainly concentrated at 160 and above, and the average NFTCs is 200 across the QTP. Our results can provide scientific guidance and a theoretical basis for the freezing resistance design of concrete in various projects on the QTP.展开更多
Cryptocurrency price prediction has garnered significant attention due to the growing importance of digital assets in the financial landscape. This paper presents a comprehensive study on predicting future cryptocurre...Cryptocurrency price prediction has garnered significant attention due to the growing importance of digital assets in the financial landscape. This paper presents a comprehensive study on predicting future cryptocurrency prices using machine learning algorithms. Open-source historical data from various cryptocurrency exchanges is utilized. Interpolation techniques are employed to handle missing data, ensuring the completeness and reliability of the dataset. Four technical indicators are selected as features for prediction. The study explores the application of five machine learning algorithms to capture the complex patterns in the highly volatile cryptocurrency market. The findings demonstrate the strengths and limitations of the different approaches, highlighting the significance of feature engineering and algorithm selection in achieving accurate cryptocurrency price predictions. The research contributes valuable insights into the dynamic and rapidly evolving field of cryptocurrency price prediction, assisting investors and traders in making informed decisions amidst the challenges posed by the cryptocurrency market.展开更多
The current study aimed at evaluating the capabilities of seven advanced machine learning techniques(MLTs),including,Support Vector Machine(SVM),Random Forest(RF),Multivariate Adaptive Regression Spline(MARS),Artifici...The current study aimed at evaluating the capabilities of seven advanced machine learning techniques(MLTs),including,Support Vector Machine(SVM),Random Forest(RF),Multivariate Adaptive Regression Spline(MARS),Artificial Neural Network(ANN),Quadratic Discriminant Analysis(QDA),Linear Discriminant Analysis(LDA),and Naive Bayes(NB),for landslide susceptibility modeling and comparison of their performances.Coupling machine learning algorithms with spatial data types for landslide susceptibility mapping is a vitally important issue.This study was carried out using GIS and R open source software at Abha Basin,Asir Region,Saudi Arabia.First,a total of 243 landslide locations were identified at Abha Basin to prepare the landslide inventory map using different data sources.All the landslide areas were randomly separated into two groups with a ratio of 70%for training and 30%for validating purposes.Twelve landslide-variables were generated for landslide susceptibility modeling,which include altitude,lithology,distance to faults,normalized difference vegetation index(NDVI),landuse/landcover(LULC),distance to roads,slope angle,distance to streams,profile curvature,plan curvature,slope length(LS),and slope-aspect.The area under curve(AUC-ROC)approach has been applied to evaluate,validate,and compare the MLTs performance.The results indicated that AUC values for seven MLTs range from 89.0%for QDA to 95.1%for RF.Our findings showed that the RF(AUC=95.1%)and LDA(AUC=941.7%)have produced the best performances in comparison to other MLTs.The outcome of this study and the landslide susceptibility maps would be useful for environmental protection.展开更多
The risk of rockbursts is one of the main threats in hard coal mines. Compared to other underground mines, the number of factors contributing to the rockburst at underground coal mines is much greater.Factors such as ...The risk of rockbursts is one of the main threats in hard coal mines. Compared to other underground mines, the number of factors contributing to the rockburst at underground coal mines is much greater.Factors such as the coal seam tendency to rockbursts, the thickness of the coal seam, and the stress level in the seam have to be considered, but also the entire coal seam-surrounding rock system has to be evaluated when trying to predict the rockbursts. However, in hard coal mines, there are stroke or stress-stroke rockbursts in which the fracture of a thick layer of sandstone plays an essential role in predicting rockbursts. The occurrence of rockbursts in coal mines is complex, and their prediction is even more difficult than in other mines. In recent years, the interest in machine learning algorithms for solving complex nonlinear problems has increased, which also applies to geosciences. This study attempts to use machine learning algorithms, i.e. neural network, decision tree, random forest, gradient boosting, and extreme gradient boosting(XGB), to assess the rockburst hazard of an active hard coal mine in the Upper Silesian Coal Basin. The rock mass bursting tendency index WTGthat describes the tendency of the seam-surrounding rock system to rockbursts and the anomaly of the vertical stress component were applied for this purpose. Especially, the decision tree and neural network models were proved to be effective in correctly distinguishing rockbursts from tremors, after which the excavation was not damaged. On average, these models correctly classified about 80% of the rockbursts in the testing datasets.展开更多
Big data analytic techniques associated with machine learning algorithms are playing an increasingly important role in various application fields,including stock market investment.However,few studies have focused on f...Big data analytic techniques associated with machine learning algorithms are playing an increasingly important role in various application fields,including stock market investment.However,few studies have focused on forecasting daily stock market returns,especially when using powerful machine learning techniques,such as deep neural networks(DNNs),to perform the analyses.DNNs employ various deep learning algorithms based on the combination of network structure,activation function,and model parameters,with their performance depending on the format of the data representation.This paper presents a comprehensive big data analytics process to predict the daily return direction of the SPDR S&P 500 ETF(ticker symbol:SPY)based on 60 financial and economic features.DNNs and traditional artificial neural networks(ANNs)are then deployed over the entire preprocessed but untransformed dataset,along with two datasets transformed via principal component analysis(PCA),to predict the daily direction of future stock market index returns.While controlling for overfitting,a pattern for the classification accuracy of the DNNs is detected and demonstrated as the number of the hidden layers increases gradually from 12 to 1000.Moreover,a set of hypothesis testing procedures are implemented on the classification,and the simulation results show that the DNNs using two PCA-represented datasets give significantly higher classification accuracy than those using the entire untransformed dataset,as well as several other hybrid machine learning algorithms.In addition,the trading strategies guided by the DNN classification process based on PCA-represented data perform slightly better than the others tested,including in a comparison against two standard benchmarks.展开更多
Some countries have announced national benchmark rates,while others have been working on the recent trend in which the London Interbank Offered Rate will be retired at the end of 2021.Considering that Turkey announced...Some countries have announced national benchmark rates,while others have been working on the recent trend in which the London Interbank Offered Rate will be retired at the end of 2021.Considering that Turkey announced the Turkish Lira Overnight Reference Interest Rate(TLREF),this study examines the determinants of TLREF.In this context,three global determinants,five country-level macroeconomic determinants,and the COVID-19 pandemic are considered by using daily data between December 28,2018,and December 31,2020,by performing machine learning algorithms and Ordinary Least Square.The empirical results show that(1)the most significant determinant is the amount of securities bought by Central Banks;(2)country-level macroeconomic factors have a higher impact whereas global factors are less important,and the pandemic does not have a significant effect;(3)Random Forest is the most accurate prediction model.Taking action by considering the study’s findings can help support economic growth by achieving low-level benchmark rates.展开更多
This study aims to empirically analyze teaching-learning-based optimization(TLBO)and machine learning algorithms using k-means and fuzzy c-means(FCM)algorithms for their individual performance evaluation in terms of c...This study aims to empirically analyze teaching-learning-based optimization(TLBO)and machine learning algorithms using k-means and fuzzy c-means(FCM)algorithms for their individual performance evaluation in terms of clustering and classification.In the first phase,the clustering(k-means and FCM)algorithms were employed independently and the clustering accuracy was evaluated using different computationalmeasures.During the second phase,the non-clustered data obtained from the first phase were preprocessed with TLBO.TLBO was performed using k-means(TLBO-KM)and FCM(TLBO-FCM)(TLBO-KM/FCM)algorithms.The objective function was determined by considering both minimization and maximization criteria.Non-clustered data obtained from the first phase were further utilized and fed as input for threshold optimization.Five benchmark datasets were considered from theUniversity of California,Irvine(UCI)Machine Learning Repository for comparative study and experimentation.These are breast cancer Wisconsin(BCW),Pima Indians Diabetes,Heart-Statlog,Hepatitis,and Cleveland Heart Disease datasets.The combined average accuracy obtained collectively is approximately 99.4%in case of TLBO-KM and 98.6%in case of TLBOFCM.This approach is also capable of finding the dominating attributes.The findings indicate that TLBO-KM/FCM,considering different computational measures,perform well on the non-clustered data where k-means and FCM,if employed independently,fail to provide significant results.Evaluating different feature sets,the TLBO-KM/FCM and SVM(GS)clearly outperformed all other classifiers in terms of sensitivity,specificity and accuracy.TLBOKM/FCM attained the highest average sensitivity(98.7%),highest average specificity(98.4%)and highest average accuracy(99.4%)for 10-fold cross validation with different test data.展开更多
This investigation assessed the efficacy of 10 widely used machine learning algorithms(MLA)comprising the least absolute shrinkage and selection operator(LASSO),generalized linear model(GLM),stepwise generalized linea...This investigation assessed the efficacy of 10 widely used machine learning algorithms(MLA)comprising the least absolute shrinkage and selection operator(LASSO),generalized linear model(GLM),stepwise generalized linear model(SGLM),elastic net(ENET),partial least square(PLS),ridge regression,support vector machine(SVM),classification and regression trees(CART),bagged CART,and random forest(RF)for gully erosion susceptibility mapping(GESM)in Iran.The location of 462 previously existing gully erosion sites were mapped through widespread field investigations,of which 70%(323)and 30%(139)of observations were arbitrarily divided for algorithm calibration and validation.Twelve controlling factors for gully erosion,namely,soil texture,annual mean rainfall,digital elevation model(DEM),drainage density,slope,lithology,topographic wetness index(TWI),distance from rivers,aspect,distance from roads,plan curvature,and profile curvature were ranked in terms of their importance using each MLA.The MLA were compared using a training dataset for gully erosion and statistical measures such as RMSE(root mean square error),MAE(mean absolute error),and R-squared.Based on the comparisons among MLA,the RF algorithm exhibited the minimum RMSE and MAE and the maximum value of R-squared,and was therefore selected as the best model.The variable importance evaluation using the RF model revealed that distance from rivers had the highest significance in influencing the occurrence of gully erosion whereas plan curvature had the least importance.According to the GESM generated using RF,most of the study area is predicted to have a low(53.72%)or moderate(29.65%)susceptibility to gully erosion,whereas only a small area is identified to have a high(12.56%)or very high(4.07%)susceptibility.The outcome generated by RF model is validated using the ROC(Receiver Operating Characteristics)curve approach,which returned an area under the curve(AUC)of 0.985,proving the excellent forecasting ability of the model.The GESM prepared using the RF algorithm can aid decision-makers in targeting remedial actions for minimizing the damage caused by gully erosion.展开更多
Periodontitis is closely related to many systemic diseases linked by different periodontal pathogens.To unravel the relationship between periodontitis and systemic diseases,it is very important to correctly discrimina...Periodontitis is closely related to many systemic diseases linked by different periodontal pathogens.To unravel the relationship between periodontitis and systemic diseases,it is very important to correctly discriminate major periodontal pathogens.To realize convenient,effcient,and high-accuracy bacterial species classification,the authors use Raman spectroscopy combined with machine learning algorithms to distinguish three major periodontal pathogens Porphyromonas gingivalis(Pg),Fusobacterium nucleatum(Fn),and Aggregatibacter actinomycetemcomitans(Aa).The result shows that this novel method can successfully discriminate the three abovementioned periodontal pathogens.Moreover,the classification accuracies for the three categories of the original data were 94.7%at the sample level and 93.9%at the spectrum level by the machine learning algorithm extra trees.This study provides a fast,simple,and accurate method which is very beneficial to differentiate periodontal pathogens.展开更多
Climate change and human activities have reduced the area and degraded the functions and services of wetlands in China.To protect and restore wetlands,it is urgent to predict the spatial distribution of potential wetl...Climate change and human activities have reduced the area and degraded the functions and services of wetlands in China.To protect and restore wetlands,it is urgent to predict the spatial distribution of potential wetlands.In this study,the distribution of potential wetlands in China was simulated by integrating the advantages of Google Earth Engine with geographic big data and machine learning algorithms.Based on a potential wetland database with 46,000 samples and an indicator system of 30 hydrologic,soil,vegetation,and topographic factors,a simulation model was constructed by machine learning algorithms.The accuracy of the random forest model for simulating the distribution of potential wetlands in China was good,with an area under the receiver operating characteristic curve value of 0.851.The area of potential wetlands was 332,702 km^(2),with 39.0%of potential wetlands in Northeast China.Geographic features were notable,and potential wetlands were mainly concentrated in areas with 400-600 mm precipitation,semi-hydric and hydric soils,meadow and marsh vegetation,altitude less than 700 m,and slope less than 3°.The results provide an important reference for wetland remote sensing mapping and a scientific basis for wetland management in China.展开更多
BACKGROUND Synchronous liver metastasis(SLM)is a significant contributor to morbidity in colorectal cancer(CRC).There are no effective predictive device integration algorithms to predict adverse SLM events during the ...BACKGROUND Synchronous liver metastasis(SLM)is a significant contributor to morbidity in colorectal cancer(CRC).There are no effective predictive device integration algorithms to predict adverse SLM events during the diagnosis of CRC.AIM To explore the risk factors for SLM in CRC and construct a visual prediction model based on gray-level co-occurrence matrix(GLCM)features collected from magnetic resonance imaging(MRI).METHODS Our study retrospectively enrolled 392 patients with CRC from Yichang Central People’s Hospital from January 2015 to May 2023.Patients were randomly divided into a training and validation group(3:7).The clinical parameters and GLCM features extracted from MRI were included as candidate variables.The prediction model was constructed using a generalized linear regression model,random forest model(RFM),and artificial neural network model.Receiver operating characteristic curves and decision curves were used to evaluate the prediction model.RESULTS Among the 392 patients,48 had SLM(12.24%).We obtained fourteen GLCM imaging data for variable screening of SLM prediction models.Inverse difference,mean sum,sum entropy,sum variance,sum of squares,energy,and difference variance were listed as candidate variables,and the prediction efficiency(area under the curve)of the subsequent RFM in the training set and internal validation set was 0.917[95%confidence interval(95%CI):0.866-0.968]and 0.09(95%CI:0.858-0.960),respectively.CONCLUSION A predictive model combining GLCM image features with machine learning can predict SLM in CRC.This model can assist clinicians in making timely and personalized clinical decisions.展开更多
In this paper,models to predict hot spot temperature and to estimate cooling air’s working parameters of racks in data centers were established using machine learning algorithms based on simulation data.First,simulat...In this paper,models to predict hot spot temperature and to estimate cooling air’s working parameters of racks in data centers were established using machine learning algorithms based on simulation data.First,simulation models of typical racks were established in computational fluid dynamics(CFD).The model was validated with field test results and results in literature,error of which was less than 3%.Then,the CFD model was used to simulate thermal environments of a typical rack considering different factors,such as servers’power,which is from 3.3 kW to 20.1 kW,cooling air’s inlet velocity,which is from 1.0 m/s to 3.0 m/s,and cooling air’s inlet temperature,which is from 16℃ to 26℃ The highest temperature in the rack,also called hot spot temperature,was selected for each case.Next,a prediction model of hot spot temperature was built using machine learning algorithms,with servers’power,cooling air’s inlet velocity and cooling air’s inlet temperature as inputs,and the hot spot temperatures as outputs.Finally,based on the prediction model,an operating parameters estimation model was established to recommend cooling air’s inlet temperatures and velocities,which can not only keep the hot spot temperature at the safety value,but are also energy saving.展开更多
Reconfigurable Intelligent Surfaces(RIS)have emerged as a promising technology for improving the reliability of massive MIMO communication networks.However,conventional RIS suffer from poor Spectral Efficiency(SE)and ...Reconfigurable Intelligent Surfaces(RIS)have emerged as a promising technology for improving the reliability of massive MIMO communication networks.However,conventional RIS suffer from poor Spectral Efficiency(SE)and high energy consumption,leading to complex Hybrid Precoding(HP)designs.To address these issues,we propose a new low-complexity HP model,named Dynamic Hybrid Relay Reflecting RIS based Hybrid Precoding(DHRR-RIS-HP).Our approach combines active and passive elements to cancel out the downsides of both conventional designs.We first design a DHRR-RIS and optimize the pilot and Channel State Information(CSI)estimation using an adaptive threshold method and Adaptive Back Propagation Neural Network(ABPNN)algorithm,respectively,to reduce the Bit Error Rate(BER)and energy consumption.To optimize the data stream,we cluster them into private and public streams using Enhanced Fuzzy C-Means(EFCM)algorithm,and schedule them based on priority and emergency level.To maximize the sum rate and SE,we perform digital precoder optimization at the Base Station(BS)side using Deep Deterministic Policy Gradient(DDPG)algorithm and analog precoder optimization at the DHRR-RIS using Fire Hawk Optimization(FHO)algorithm.We implement our proposed work using MATLAB R2020a and compare it with existing works using several validation metrics.Our results show that our proposed work outperforms existing works in terms of SE,Weighted Sum Rate(WSR),and BER.展开更多
Model parameters estimation is a pivotal issue for runoff modeling in ungauged catchments.The nonlinear relationship between model parameters and catchment descriptors is a major obstacle for parameter regionalization...Model parameters estimation is a pivotal issue for runoff modeling in ungauged catchments.The nonlinear relationship between model parameters and catchment descriptors is a major obstacle for parameter regionalization,which is the most widely used approach.Runoff modeling was studied in 38 catchments located in the Yellow–Huai–Hai River Basin(YHHRB).The values of the Nash–Sutcliffe efficiency coefficient(NSE),coefficient of determination(R2),and percent bias(PBIAS)indicated the acceptable performance of the soil and water assessment tool(SWAT)model in the YHHRB.Nine descriptors belonging to the categories of climate,soil,vegetation,and topography were used to express the catchment characteristics related to the hydrological processes.The quantitative relationships between the parameters of the SWAT model and the catchment descriptors were analyzed by six regression-based models,including linear regression(LR)equations,support vector regression(SVR),random forest(RF),k-nearest neighbor(kNN),decision tree(DT),and radial basis function(RBF).Each of the 38 catchments was assumed to be an ungauged catchment in turn.Then,the parameters in each target catchment were estimated by the constructed regression models based on the remaining 37 donor catchments.Furthermore,the similaritybased regionalization scheme was used for comparison with the regression-based approach.The results indicated that the runoff with the highest accuracy was modeled by the SVR-based scheme in ungauged catchments.Compared with the traditional LR-based approach,the accuracy of the runoff modeling in ungauged catchments was improved by the machine learning algorithms because of the outstanding capability to deal with nonlinear relationships.The performances of different approaches were similar in humid regions,while the advantages of the machine learning techniques were more evident in arid regions.When the study area contained nested catchments,the best result was calculated with the similarity-based parameter regionalization scheme because of the high catchment density and short spatial distance.The new findings could improve flood forecasting and water resources planning in regions that lack observed data.展开更多
In this study,experimental wear losses under different loads and sliding distances of AISI 1020 steel surfaces coated with(wt.%)50FeCrC‐20FeW‐30FeB and 70FeCrC‐30FeB powder mixtures by plasma transfer arc welding w...In this study,experimental wear losses under different loads and sliding distances of AISI 1020 steel surfaces coated with(wt.%)50FeCrC‐20FeW‐30FeB and 70FeCrC‐30FeB powder mixtures by plasma transfer arc welding were determined.The dataset comprised 99 different wear amount measurements obtained experimentally in the laboratory.The linear regression(LR),support vector machine(SVM),and Gaussian process regression(GPR)algorithms are used for predicting wear quantities.A success rate of 0.93 was obtained from the LR algorithm and 0.96 from the SVM and GPR algorithms.展开更多
This paper empirically evaluates the several machine learning algorithms adaptable for lung cancer detection linked with IoT devices.In this work,a review of nearly 65 papers for predicting different diseases,using ma...This paper empirically evaluates the several machine learning algorithms adaptable for lung cancer detection linked with IoT devices.In this work,a review of nearly 65 papers for predicting different diseases,using machine learning algorithms,has been done.The analysis mainly focuses on various machine learning algorithms used for detecting several diseases in order to search for a gap toward the future improvement for detecting lung cancer in medical IoT.Each technique was analyzed on each step,and the overall drawbacks are pointed out.In addition,it also analyzes the type of data used for predicting the concerned disease,whether it is the benchmark or manually collected data.Finally,research directions have been identified and depicted based on the various existing methodologies.This will be helpful for the upcoming researchers to detect the cancerous patients accurately in early stages without any flaws.展开更多
Emotion represents the feeling of an individual in a given situation. There are various ways to express the emotions of an individual. It can be categorized into verbal expressions, written expressions, facial express...Emotion represents the feeling of an individual in a given situation. There are various ways to express the emotions of an individual. It can be categorized into verbal expressions, written expressions, facial expressions and gestures. Among these various ways of expressing the emotion, the written method is a challenging task to extract the emotions, as the data is in the form of textual dat. Finding the different kinds of emotions is also a tedious task as it requires a lot of pre preparations of the textual data taken for the research. This research work is carried out to analyse and extract the emotions hidden in text data. The text data taken for the analysis is from the social media dataset. Using the raw text data directly from the social media will not serve the purpose. Therefore, the text data has to be pre-processed and then utilised for further processing. Pre-processing makes the text data more efficient and would infer valuable insights of the emotions hidden in it. The preprocessing steps also help to manage the text data for identifying the emotions conveyed in the text. This work proposes to deduct the emotions taken from the social media text data by applying the machine learning algorithm. Finally, the usefulness of the emotions is suggested for various stake holders, to find the attitude of individuals at that moment, the data is produced. .展开更多
Due to the development of the novel materials,the past two decades have witnessed the rapid advances of soft electronics.The soft electronics have huge potential in the physical sign monitoring and health care.One of ...Due to the development of the novel materials,the past two decades have witnessed the rapid advances of soft electronics.The soft electronics have huge potential in the physical sign monitoring and health care.One of the important advantages of soft electronics is forming good interface with skin,which can increase the user scale and improve the signal quality.Therefore,it is easy to build the specific dataset,which is important to improve the performance of machine learning algorithm.At the same time,with the assistance of machine learning algorithm,the soft electronics have become more and more intelligent to realize real-time analysis and diagnosis.The soft electronics and machining learning algorithms complement each other very well.It is indubitable that the soft electronics will bring us to a healthier and more intelligent world in the near future.Therefore,in this review,we will give a careful introduction about the new soft material,physiological signal detected by soft devices,and the soft devices assisted by machine learning algorithm.Some soft materials will be discussed such as two-dimensional material,carbon nanotube,nanowire,nanomesh,and hydrogel.Then,soft sensors will be discussed according to the physiological signal types(pulse,respiration,human motion,intraocular pressure,phonation,etc.).After that,the soft electronics assisted by various algorithms will be reviewed,including some classical algorithms and powerful neural network algorithms.Especially,the soft device assisted by neural network will be introduced carefully.Finally,the outlook,challenge,and conclusion of soft system powered by machine learning algorithm will be discussed.展开更多
Traditional wireless sensor networks(WSNs)are not suitable for rough terrains that are difficult or impossible to access by humans.Smart dust is a technology that works with the combination of many tiny sensors which ...Traditional wireless sensor networks(WSNs)are not suitable for rough terrains that are difficult or impossible to access by humans.Smart dust is a technology that works with the combination of many tiny sensors which is highly useful for obtaining remote sensing information from rough terrains.The tiny sensors are sprinkled in large numbers on rough terrains using airborne distribution through drones or aircraftwithout manually setting their locations.Although it is clear that a number of remote sensing applications can benefit from this technology,but the small size of smart dust fundamentally restricts the integration of advanced hardware on tiny sensors.This raises many challenges including how to estimate the location of events sensed by the smart dusts.Existing solutions on estimating the location of events sensed by the smart dusts are not suitable for monitoring rough terrains as these solutions depend on relay sensors and laser patterns which have their own limitations in terms of power constraint and uneven surfaces.The study proposes a novel machine learning based localization algorithm for estimating the location of events.The approach utilizes timestamps(time of arrival)of sensed events received at base stations by assembling them into a multidimensional vector and input to a machine learning classifier for estimating the location.Due to the unavailability of real smart dusts,we built a simulator for analysing the accuracy of the proposed approach formonitoring forest fire.The experiments on the simulator show reasonable accuracy of the approach.展开更多
In this study, the author will investigate and utilize advanced machine learning models related to two different methodologies to determine the best and most effective way to predict individuals with heart failure and...In this study, the author will investigate and utilize advanced machine learning models related to two different methodologies to determine the best and most effective way to predict individuals with heart failure and cardiovascular diseases. The first methodology involves a list of classification machine learning algorithms, and the second methodology involves the use of a deep learning algorithm known as MLP or Multilayer Perceptrons. Globally, hospitals are dealing with cases related to cardiovascular diseases and heart failure as they are major causes of death, not only for overweight individuals but also for those who do not adopt a healthy diet and lifestyle. Often, heart failures and cardiovascular diseases can be caused by many factors, including cardiomyopathy, high blood pressure, coronary heart disease, and heart inflammation [1]. Other factors, such as irregular shocks or stress, can also contribute to heart failure or a heart attack. While these events cannot be predicted, continuous data from patients’ health can help doctors predict heart failure. Therefore, this data-driven research utilizes advanced machine learning and deep learning techniques to better analyze and manipulate the data, providing doctors with informative decision-making tools regarding a person’s likelihood of experiencing heart failure. In this paper, the author employed advanced data preprocessing and cleaning techniques. Additionally, the dataset underwent testing using two different methodologies to determine the most effective machine-learning technique for producing optimal predictions. The first methodology involved employing a list of supervised classification machine learning algorithms, including Naïve Bayes (NB), KNN, logistic regression, and the SVM algorithm. The second methodology utilized a deep learning (DL) algorithm known as Multilayer Perceptrons (MLPs). This algorithm provided the author with the flexibility to experiment with different layer sizes and activation functions, such as ReLU, logistic (sigmoid), and Tanh. Both methodologies produced optimal models with high-level accuracy rates. The first methodology involves a list of supervised machine learning algorithms, including KNN, SVM, Adaboost, Logistic Regression, Naive Bayes, and Decision Tree algorithms. They achieved accuracy rates of 86%, 89%, 89%, 81%, 79%, and 99%, respectively. The author clearly explained that Decision Tree algorithm is not suitable for the dataset at hand due to overfitting issues. Therefore, it was discarded as an optimal model to be used. However, the latter methodology (Neural Network) demonstrated the most stable and optimal accuracy, achieving over 87% accuracy while adapting well to real-life situations and requiring low computing power overall. A performance assessment and evaluation were carried out based on a confusion matrix report to demonstrate feasibility and performance. The author concluded that the performance of the model in real-life situations can advance not only the medical field of science but also mathematical concepts. Additionally, the advanced preprocessing approach behind the model can provide value to the Data Science community. The model can be further developed by employing various optimization techniques to handle even larger datasets related to heart failures. Furthermore, different neural network algorithms can be tested to explore alternative approaches and yield different results.展开更多
基金supported by Shandong Provincial Natural Science Foundation (grant number: ZR2023MD036)Key Research and Development Project in Shandong Province (grant number: 2019GGX101064)project for excellent youth foundation of the innovation teacher team, Shandong (grant number: 2022KJ310)。
文摘The reasonable quantification of the concrete freezing environment on the Qinghai–Tibet Plateau(QTP) is the primary issue in frost resistant concrete design, which is one of the challenges that the QTP engineering managers should take into account. In this paper, we propose a more realistic method to calculate the number of concrete freeze–thaw cycles(NFTCs) on the QTP. The calculated results show that the NFTCs increase as the altitude of the meteorological station increases with the average NFTCs being 208.7. Four machine learning methods, i.e., the random forest(RF) model, generalized boosting method(GBM), generalized linear model(GLM), and generalized additive model(GAM), are used to fit the NFTCs. The root mean square error(RMSE) values of the RF, GBM, GLM, and GAM are 32.3, 4.3, 247.9, and 161.3, respectively. The R^(2) values of the RF, GBM, GLM, and GAM are 0.93, 0.99, 0.48, and 0.66, respectively. The GBM method performs the best compared to the other three methods, which was shown by the results of RMSE and R^(2) values. The quantitative results from the GBM method indicate that the lowest, medium, and highest NFTC values are distributed in the northern, central, and southern parts of the QTP, respectively. The annual NFTCs in the QTP region are mainly concentrated at 160 and above, and the average NFTCs is 200 across the QTP. Our results can provide scientific guidance and a theoretical basis for the freezing resistance design of concrete in various projects on the QTP.
文摘Cryptocurrency price prediction has garnered significant attention due to the growing importance of digital assets in the financial landscape. This paper presents a comprehensive study on predicting future cryptocurrency prices using machine learning algorithms. Open-source historical data from various cryptocurrency exchanges is utilized. Interpolation techniques are employed to handle missing data, ensuring the completeness and reliability of the dataset. Four technical indicators are selected as features for prediction. The study explores the application of five machine learning algorithms to capture the complex patterns in the highly volatile cryptocurrency market. The findings demonstrate the strengths and limitations of the different approaches, highlighting the significance of feature engineering and algorithm selection in achieving accurate cryptocurrency price predictions. The research contributes valuable insights into the dynamic and rapidly evolving field of cryptocurrency price prediction, assisting investors and traders in making informed decisions amidst the challenges posed by the cryptocurrency market.
文摘The current study aimed at evaluating the capabilities of seven advanced machine learning techniques(MLTs),including,Support Vector Machine(SVM),Random Forest(RF),Multivariate Adaptive Regression Spline(MARS),Artificial Neural Network(ANN),Quadratic Discriminant Analysis(QDA),Linear Discriminant Analysis(LDA),and Naive Bayes(NB),for landslide susceptibility modeling and comparison of their performances.Coupling machine learning algorithms with spatial data types for landslide susceptibility mapping is a vitally important issue.This study was carried out using GIS and R open source software at Abha Basin,Asir Region,Saudi Arabia.First,a total of 243 landslide locations were identified at Abha Basin to prepare the landslide inventory map using different data sources.All the landslide areas were randomly separated into two groups with a ratio of 70%for training and 30%for validating purposes.Twelve landslide-variables were generated for landslide susceptibility modeling,which include altitude,lithology,distance to faults,normalized difference vegetation index(NDVI),landuse/landcover(LULC),distance to roads,slope angle,distance to streams,profile curvature,plan curvature,slope length(LS),and slope-aspect.The area under curve(AUC-ROC)approach has been applied to evaluate,validate,and compare the MLTs performance.The results indicated that AUC values for seven MLTs range from 89.0%for QDA to 95.1%for RF.Our findings showed that the RF(AUC=95.1%)and LDA(AUC=941.7%)have produced the best performances in comparison to other MLTs.The outcome of this study and the landslide susceptibility maps would be useful for environmental protection.
基金supported by the Ministry of Science and Higher Education, Republic of Poland (Statutory Activity of the Central Mining Institute, Grant No. 11133010)
文摘The risk of rockbursts is one of the main threats in hard coal mines. Compared to other underground mines, the number of factors contributing to the rockburst at underground coal mines is much greater.Factors such as the coal seam tendency to rockbursts, the thickness of the coal seam, and the stress level in the seam have to be considered, but also the entire coal seam-surrounding rock system has to be evaluated when trying to predict the rockbursts. However, in hard coal mines, there are stroke or stress-stroke rockbursts in which the fracture of a thick layer of sandstone plays an essential role in predicting rockbursts. The occurrence of rockbursts in coal mines is complex, and their prediction is even more difficult than in other mines. In recent years, the interest in machine learning algorithms for solving complex nonlinear problems has increased, which also applies to geosciences. This study attempts to use machine learning algorithms, i.e. neural network, decision tree, random forest, gradient boosting, and extreme gradient boosting(XGB), to assess the rockburst hazard of an active hard coal mine in the Upper Silesian Coal Basin. The rock mass bursting tendency index WTGthat describes the tendency of the seam-surrounding rock system to rockbursts and the anomaly of the vertical stress component were applied for this purpose. Especially, the decision tree and neural network models were proved to be effective in correctly distinguishing rockbursts from tremors, after which the excavation was not damaged. On average, these models correctly classified about 80% of the rockbursts in the testing datasets.
文摘Big data analytic techniques associated with machine learning algorithms are playing an increasingly important role in various application fields,including stock market investment.However,few studies have focused on forecasting daily stock market returns,especially when using powerful machine learning techniques,such as deep neural networks(DNNs),to perform the analyses.DNNs employ various deep learning algorithms based on the combination of network structure,activation function,and model parameters,with their performance depending on the format of the data representation.This paper presents a comprehensive big data analytics process to predict the daily return direction of the SPDR S&P 500 ETF(ticker symbol:SPY)based on 60 financial and economic features.DNNs and traditional artificial neural networks(ANNs)are then deployed over the entire preprocessed but untransformed dataset,along with two datasets transformed via principal component analysis(PCA),to predict the daily direction of future stock market index returns.While controlling for overfitting,a pattern for the classification accuracy of the DNNs is detected and demonstrated as the number of the hidden layers increases gradually from 12 to 1000.Moreover,a set of hypothesis testing procedures are implemented on the classification,and the simulation results show that the DNNs using two PCA-represented datasets give significantly higher classification accuracy than those using the entire untransformed dataset,as well as several other hybrid machine learning algorithms.In addition,the trading strategies guided by the DNN classification process based on PCA-represented data perform slightly better than the others tested,including in a comparison against two standard benchmarks.
文摘Some countries have announced national benchmark rates,while others have been working on the recent trend in which the London Interbank Offered Rate will be retired at the end of 2021.Considering that Turkey announced the Turkish Lira Overnight Reference Interest Rate(TLREF),this study examines the determinants of TLREF.In this context,three global determinants,five country-level macroeconomic determinants,and the COVID-19 pandemic are considered by using daily data between December 28,2018,and December 31,2020,by performing machine learning algorithms and Ordinary Least Square.The empirical results show that(1)the most significant determinant is the amount of securities bought by Central Banks;(2)country-level macroeconomic factors have a higher impact whereas global factors are less important,and the pandemic does not have a significant effect;(3)Random Forest is the most accurate prediction model.Taking action by considering the study’s findings can help support economic growth by achieving low-level benchmark rates.
文摘This study aims to empirically analyze teaching-learning-based optimization(TLBO)and machine learning algorithms using k-means and fuzzy c-means(FCM)algorithms for their individual performance evaluation in terms of clustering and classification.In the first phase,the clustering(k-means and FCM)algorithms were employed independently and the clustering accuracy was evaluated using different computationalmeasures.During the second phase,the non-clustered data obtained from the first phase were preprocessed with TLBO.TLBO was performed using k-means(TLBO-KM)and FCM(TLBO-FCM)(TLBO-KM/FCM)algorithms.The objective function was determined by considering both minimization and maximization criteria.Non-clustered data obtained from the first phase were further utilized and fed as input for threshold optimization.Five benchmark datasets were considered from theUniversity of California,Irvine(UCI)Machine Learning Repository for comparative study and experimentation.These are breast cancer Wisconsin(BCW),Pima Indians Diabetes,Heart-Statlog,Hepatitis,and Cleveland Heart Disease datasets.The combined average accuracy obtained collectively is approximately 99.4%in case of TLBO-KM and 98.6%in case of TLBOFCM.This approach is also capable of finding the dominating attributes.The findings indicate that TLBO-KM/FCM,considering different computational measures,perform well on the non-clustered data where k-means and FCM,if employed independently,fail to provide significant results.Evaluating different feature sets,the TLBO-KM/FCM and SVM(GS)clearly outperformed all other classifiers in terms of sensitivity,specificity and accuracy.TLBOKM/FCM attained the highest average sensitivity(98.7%),highest average specificity(98.4%)and highest average accuracy(99.4%)for 10-fold cross validation with different test data.
基金supported by the College of Agriculture,Shiraz University(Grant No.97GRC1M271143)funding from the UK Biotechnology and Biological Sciences Research Council(BBSRC)funded by BBSRC grant award BBS/E/C/000I0330–Soil to Nutrition project 3–Sustainable intensification:optimisation at multiple scales。
文摘This investigation assessed the efficacy of 10 widely used machine learning algorithms(MLA)comprising the least absolute shrinkage and selection operator(LASSO),generalized linear model(GLM),stepwise generalized linear model(SGLM),elastic net(ENET),partial least square(PLS),ridge regression,support vector machine(SVM),classification and regression trees(CART),bagged CART,and random forest(RF)for gully erosion susceptibility mapping(GESM)in Iran.The location of 462 previously existing gully erosion sites were mapped through widespread field investigations,of which 70%(323)and 30%(139)of observations were arbitrarily divided for algorithm calibration and validation.Twelve controlling factors for gully erosion,namely,soil texture,annual mean rainfall,digital elevation model(DEM),drainage density,slope,lithology,topographic wetness index(TWI),distance from rivers,aspect,distance from roads,plan curvature,and profile curvature were ranked in terms of their importance using each MLA.The MLA were compared using a training dataset for gully erosion and statistical measures such as RMSE(root mean square error),MAE(mean absolute error),and R-squared.Based on the comparisons among MLA,the RF algorithm exhibited the minimum RMSE and MAE and the maximum value of R-squared,and was therefore selected as the best model.The variable importance evaluation using the RF model revealed that distance from rivers had the highest significance in influencing the occurrence of gully erosion whereas plan curvature had the least importance.According to the GESM generated using RF,most of the study area is predicted to have a low(53.72%)or moderate(29.65%)susceptibility to gully erosion,whereas only a small area is identified to have a high(12.56%)or very high(4.07%)susceptibility.The outcome generated by RF model is validated using the ROC(Receiver Operating Characteristics)curve approach,which returned an area under the curve(AUC)of 0.985,proving the excellent forecasting ability of the model.The GESM prepared using the RF algorithm can aid decision-makers in targeting remedial actions for minimizing the damage caused by gully erosion.
基金funded by the Major Program of Social Science Foundation of Tianjin Municipal Education Commission(2019JWZD53).
文摘Periodontitis is closely related to many systemic diseases linked by different periodontal pathogens.To unravel the relationship between periodontitis and systemic diseases,it is very important to correctly discriminate major periodontal pathogens.To realize convenient,effcient,and high-accuracy bacterial species classification,the authors use Raman spectroscopy combined with machine learning algorithms to distinguish three major periodontal pathogens Porphyromonas gingivalis(Pg),Fusobacterium nucleatum(Fn),and Aggregatibacter actinomycetemcomitans(Aa).The result shows that this novel method can successfully discriminate the three abovementioned periodontal pathogens.Moreover,the classification accuracies for the three categories of the original data were 94.7%at the sample level and 93.9%at the spectrum level by the machine learning algorithm extra trees.This study provides a fast,simple,and accurate method which is very beneficial to differentiate periodontal pathogens.
基金supported by the Natural Science Foundation of Jilin Province,China[YDZJ202301ZYTS218]the National Natural Science Foundation of China[42301430,42222103,42171379,U2243230,and 42101379]+1 种基金the Youth Innovation Promotion Association of the Chinese Academy of Sciences[2017277 and 2021227]the Professional Association of the Alliance of International Science Organizations[ANSO-PA-2020-14].
文摘Climate change and human activities have reduced the area and degraded the functions and services of wetlands in China.To protect and restore wetlands,it is urgent to predict the spatial distribution of potential wetlands.In this study,the distribution of potential wetlands in China was simulated by integrating the advantages of Google Earth Engine with geographic big data and machine learning algorithms.Based on a potential wetland database with 46,000 samples and an indicator system of 30 hydrologic,soil,vegetation,and topographic factors,a simulation model was constructed by machine learning algorithms.The accuracy of the random forest model for simulating the distribution of potential wetlands in China was good,with an area under the receiver operating characteristic curve value of 0.851.The area of potential wetlands was 332,702 km^(2),with 39.0%of potential wetlands in Northeast China.Geographic features were notable,and potential wetlands were mainly concentrated in areas with 400-600 mm precipitation,semi-hydric and hydric soils,meadow and marsh vegetation,altitude less than 700 m,and slope less than 3°.The results provide an important reference for wetland remote sensing mapping and a scientific basis for wetland management in China.
文摘BACKGROUND Synchronous liver metastasis(SLM)is a significant contributor to morbidity in colorectal cancer(CRC).There are no effective predictive device integration algorithms to predict adverse SLM events during the diagnosis of CRC.AIM To explore the risk factors for SLM in CRC and construct a visual prediction model based on gray-level co-occurrence matrix(GLCM)features collected from magnetic resonance imaging(MRI).METHODS Our study retrospectively enrolled 392 patients with CRC from Yichang Central People’s Hospital from January 2015 to May 2023.Patients were randomly divided into a training and validation group(3:7).The clinical parameters and GLCM features extracted from MRI were included as candidate variables.The prediction model was constructed using a generalized linear regression model,random forest model(RFM),and artificial neural network model.Receiver operating characteristic curves and decision curves were used to evaluate the prediction model.RESULTS Among the 392 patients,48 had SLM(12.24%).We obtained fourteen GLCM imaging data for variable screening of SLM prediction models.Inverse difference,mean sum,sum entropy,sum variance,sum of squares,energy,and difference variance were listed as candidate variables,and the prediction efficiency(area under the curve)of the subsequent RFM in the training set and internal validation set was 0.917[95%confidence interval(95%CI):0.866-0.968]and 0.09(95%CI:0.858-0.960),respectively.CONCLUSION A predictive model combining GLCM image features with machine learning can predict SLM in CRC.This model can assist clinicians in making timely and personalized clinical decisions.
基金The authors appreciate support of the project from China Electronics Engineering Design Institute CO.,LTD.(No.SDIC2021-08)from the Beijing Natural Science Foundation(No.4212040).
文摘In this paper,models to predict hot spot temperature and to estimate cooling air’s working parameters of racks in data centers were established using machine learning algorithms based on simulation data.First,simulation models of typical racks were established in computational fluid dynamics(CFD).The model was validated with field test results and results in literature,error of which was less than 3%.Then,the CFD model was used to simulate thermal environments of a typical rack considering different factors,such as servers’power,which is from 3.3 kW to 20.1 kW,cooling air’s inlet velocity,which is from 1.0 m/s to 3.0 m/s,and cooling air’s inlet temperature,which is from 16℃ to 26℃ The highest temperature in the rack,also called hot spot temperature,was selected for each case.Next,a prediction model of hot spot temperature was built using machine learning algorithms,with servers’power,cooling air’s inlet velocity and cooling air’s inlet temperature as inputs,and the hot spot temperatures as outputs.Finally,based on the prediction model,an operating parameters estimation model was established to recommend cooling air’s inlet temperatures and velocities,which can not only keep the hot spot temperature at the safety value,but are also energy saving.
文摘Reconfigurable Intelligent Surfaces(RIS)have emerged as a promising technology for improving the reliability of massive MIMO communication networks.However,conventional RIS suffer from poor Spectral Efficiency(SE)and high energy consumption,leading to complex Hybrid Precoding(HP)designs.To address these issues,we propose a new low-complexity HP model,named Dynamic Hybrid Relay Reflecting RIS based Hybrid Precoding(DHRR-RIS-HP).Our approach combines active and passive elements to cancel out the downsides of both conventional designs.We first design a DHRR-RIS and optimize the pilot and Channel State Information(CSI)estimation using an adaptive threshold method and Adaptive Back Propagation Neural Network(ABPNN)algorithm,respectively,to reduce the Bit Error Rate(BER)and energy consumption.To optimize the data stream,we cluster them into private and public streams using Enhanced Fuzzy C-Means(EFCM)algorithm,and schedule them based on priority and emergency level.To maximize the sum rate and SE,we perform digital precoder optimization at the Base Station(BS)side using Deep Deterministic Policy Gradient(DDPG)algorithm and analog precoder optimization at the DHRR-RIS using Fire Hawk Optimization(FHO)algorithm.We implement our proposed work using MATLAB R2020a and compare it with existing works using several validation metrics.Our results show that our proposed work outperforms existing works in terms of SE,Weighted Sum Rate(WSR),and BER.
基金funded by the National Key Research and Development Program of China(2017YFA0605002,2017YFA0605004,and 2016YFA0601501)the National Natural Science Foundation of China(41961124007,51779145,and 41830863)“Six top talents”in Jiangsu Province(RJFW-031)。
文摘Model parameters estimation is a pivotal issue for runoff modeling in ungauged catchments.The nonlinear relationship between model parameters and catchment descriptors is a major obstacle for parameter regionalization,which is the most widely used approach.Runoff modeling was studied in 38 catchments located in the Yellow–Huai–Hai River Basin(YHHRB).The values of the Nash–Sutcliffe efficiency coefficient(NSE),coefficient of determination(R2),and percent bias(PBIAS)indicated the acceptable performance of the soil and water assessment tool(SWAT)model in the YHHRB.Nine descriptors belonging to the categories of climate,soil,vegetation,and topography were used to express the catchment characteristics related to the hydrological processes.The quantitative relationships between the parameters of the SWAT model and the catchment descriptors were analyzed by six regression-based models,including linear regression(LR)equations,support vector regression(SVR),random forest(RF),k-nearest neighbor(kNN),decision tree(DT),and radial basis function(RBF).Each of the 38 catchments was assumed to be an ungauged catchment in turn.Then,the parameters in each target catchment were estimated by the constructed regression models based on the remaining 37 donor catchments.Furthermore,the similaritybased regionalization scheme was used for comparison with the regression-based approach.The results indicated that the runoff with the highest accuracy was modeled by the SVR-based scheme in ungauged catchments.Compared with the traditional LR-based approach,the accuracy of the runoff modeling in ungauged catchments was improved by the machine learning algorithms because of the outstanding capability to deal with nonlinear relationships.The performances of different approaches were similar in humid regions,while the advantages of the machine learning techniques were more evident in arid regions.When the study area contained nested catchments,the best result was calculated with the similarity-based parameter regionalization scheme because of the high catchment density and short spatial distance.The new findings could improve flood forecasting and water resources planning in regions that lack observed data.
文摘In this study,experimental wear losses under different loads and sliding distances of AISI 1020 steel surfaces coated with(wt.%)50FeCrC‐20FeW‐30FeB and 70FeCrC‐30FeB powder mixtures by plasma transfer arc welding were determined.The dataset comprised 99 different wear amount measurements obtained experimentally in the laboratory.The linear regression(LR),support vector machine(SVM),and Gaussian process regression(GPR)algorithms are used for predicting wear quantities.A success rate of 0.93 was obtained from the LR algorithm and 0.96 from the SVM and GPR algorithms.
文摘This paper empirically evaluates the several machine learning algorithms adaptable for lung cancer detection linked with IoT devices.In this work,a review of nearly 65 papers for predicting different diseases,using machine learning algorithms,has been done.The analysis mainly focuses on various machine learning algorithms used for detecting several diseases in order to search for a gap toward the future improvement for detecting lung cancer in medical IoT.Each technique was analyzed on each step,and the overall drawbacks are pointed out.In addition,it also analyzes the type of data used for predicting the concerned disease,whether it is the benchmark or manually collected data.Finally,research directions have been identified and depicted based on the various existing methodologies.This will be helpful for the upcoming researchers to detect the cancerous patients accurately in early stages without any flaws.
文摘Emotion represents the feeling of an individual in a given situation. There are various ways to express the emotions of an individual. It can be categorized into verbal expressions, written expressions, facial expressions and gestures. Among these various ways of expressing the emotion, the written method is a challenging task to extract the emotions, as the data is in the form of textual dat. Finding the different kinds of emotions is also a tedious task as it requires a lot of pre preparations of the textual data taken for the research. This research work is carried out to analyse and extract the emotions hidden in text data. The text data taken for the analysis is from the social media dataset. Using the raw text data directly from the social media will not serve the purpose. Therefore, the text data has to be pre-processed and then utilised for further processing. Pre-processing makes the text data more efficient and would infer valuable insights of the emotions hidden in it. The preprocessing steps also help to manage the text data for identifying the emotions conveyed in the text. This work proposes to deduct the emotions taken from the social media text data by applying the machine learning algorithm. Finally, the usefulness of the emotions is suggested for various stake holders, to find the attitude of individuals at that moment, the data is produced. .
基金supported by National Natural Science Foundation of China(No.62201624,32000939,21775168,22174167,51861145202,U20A20168)the Guangdong Basic and Applied Basic Research Foundation(2019A1515111183)+3 种基金Shenzhen Research Funding Program(JCYJ20190807160401657,JCYJ201908073000608,JCYJ20150831192224146)the National Key R&D Program(2018YFC2001202)the support of the Research Fund from Tsinghua University Initiative Scientific Research Programthe support from Key Laboratory of Sensing Technology and Biomedical Instruments of Guangdong Province(No.2020B1212060077)。
文摘Due to the development of the novel materials,the past two decades have witnessed the rapid advances of soft electronics.The soft electronics have huge potential in the physical sign monitoring and health care.One of the important advantages of soft electronics is forming good interface with skin,which can increase the user scale and improve the signal quality.Therefore,it is easy to build the specific dataset,which is important to improve the performance of machine learning algorithm.At the same time,with the assistance of machine learning algorithm,the soft electronics have become more and more intelligent to realize real-time analysis and diagnosis.The soft electronics and machining learning algorithms complement each other very well.It is indubitable that the soft electronics will bring us to a healthier and more intelligent world in the near future.Therefore,in this review,we will give a careful introduction about the new soft material,physiological signal detected by soft devices,and the soft devices assisted by machine learning algorithm.Some soft materials will be discussed such as two-dimensional material,carbon nanotube,nanowire,nanomesh,and hydrogel.Then,soft sensors will be discussed according to the physiological signal types(pulse,respiration,human motion,intraocular pressure,phonation,etc.).After that,the soft electronics assisted by various algorithms will be reviewed,including some classical algorithms and powerful neural network algorithms.Especially,the soft device assisted by neural network will be introduced carefully.Finally,the outlook,challenge,and conclusion of soft system powered by machine learning algorithm will be discussed.
基金This research is supported by Universiti Brunei Darussalam(UBD)under FIC allied research grant program.
文摘Traditional wireless sensor networks(WSNs)are not suitable for rough terrains that are difficult or impossible to access by humans.Smart dust is a technology that works with the combination of many tiny sensors which is highly useful for obtaining remote sensing information from rough terrains.The tiny sensors are sprinkled in large numbers on rough terrains using airborne distribution through drones or aircraftwithout manually setting their locations.Although it is clear that a number of remote sensing applications can benefit from this technology,but the small size of smart dust fundamentally restricts the integration of advanced hardware on tiny sensors.This raises many challenges including how to estimate the location of events sensed by the smart dusts.Existing solutions on estimating the location of events sensed by the smart dusts are not suitable for monitoring rough terrains as these solutions depend on relay sensors and laser patterns which have their own limitations in terms of power constraint and uneven surfaces.The study proposes a novel machine learning based localization algorithm for estimating the location of events.The approach utilizes timestamps(time of arrival)of sensed events received at base stations by assembling them into a multidimensional vector and input to a machine learning classifier for estimating the location.Due to the unavailability of real smart dusts,we built a simulator for analysing the accuracy of the proposed approach formonitoring forest fire.The experiments on the simulator show reasonable accuracy of the approach.
文摘In this study, the author will investigate and utilize advanced machine learning models related to two different methodologies to determine the best and most effective way to predict individuals with heart failure and cardiovascular diseases. The first methodology involves a list of classification machine learning algorithms, and the second methodology involves the use of a deep learning algorithm known as MLP or Multilayer Perceptrons. Globally, hospitals are dealing with cases related to cardiovascular diseases and heart failure as they are major causes of death, not only for overweight individuals but also for those who do not adopt a healthy diet and lifestyle. Often, heart failures and cardiovascular diseases can be caused by many factors, including cardiomyopathy, high blood pressure, coronary heart disease, and heart inflammation [1]. Other factors, such as irregular shocks or stress, can also contribute to heart failure or a heart attack. While these events cannot be predicted, continuous data from patients’ health can help doctors predict heart failure. Therefore, this data-driven research utilizes advanced machine learning and deep learning techniques to better analyze and manipulate the data, providing doctors with informative decision-making tools regarding a person’s likelihood of experiencing heart failure. In this paper, the author employed advanced data preprocessing and cleaning techniques. Additionally, the dataset underwent testing using two different methodologies to determine the most effective machine-learning technique for producing optimal predictions. The first methodology involved employing a list of supervised classification machine learning algorithms, including Naïve Bayes (NB), KNN, logistic regression, and the SVM algorithm. The second methodology utilized a deep learning (DL) algorithm known as Multilayer Perceptrons (MLPs). This algorithm provided the author with the flexibility to experiment with different layer sizes and activation functions, such as ReLU, logistic (sigmoid), and Tanh. Both methodologies produced optimal models with high-level accuracy rates. The first methodology involves a list of supervised machine learning algorithms, including KNN, SVM, Adaboost, Logistic Regression, Naive Bayes, and Decision Tree algorithms. They achieved accuracy rates of 86%, 89%, 89%, 81%, 79%, and 99%, respectively. The author clearly explained that Decision Tree algorithm is not suitable for the dataset at hand due to overfitting issues. Therefore, it was discarded as an optimal model to be used. However, the latter methodology (Neural Network) demonstrated the most stable and optimal accuracy, achieving over 87% accuracy while adapting well to real-life situations and requiring low computing power overall. A performance assessment and evaluation were carried out based on a confusion matrix report to demonstrate feasibility and performance. The author concluded that the performance of the model in real-life situations can advance not only the medical field of science but also mathematical concepts. Additionally, the advanced preprocessing approach behind the model can provide value to the Data Science community. The model can be further developed by employing various optimization techniques to handle even larger datasets related to heart failures. Furthermore, different neural network algorithms can be tested to explore alternative approaches and yield different results.