期刊文献+
共找到15,744篇文章
< 1 2 250 >
每页显示 20 50 100
Enhancing Deep Learning Soil Moisture Forecasting Models by Integrating Physics-based Models 被引量:1
1
作者 Lu LI Yongjiu DAI +5 位作者 Zhongwang WEI Wei SHANGGUAN Nan WEI Yonggen ZHANG Qingliang LI Xian-Xiang LI 《Advances in Atmospheric Sciences》 SCIE CAS CSCD 2024年第7期1326-1341,共16页
Accurate soil moisture(SM)prediction is critical for understanding hydrological processes.Physics-based(PB)models exhibit large uncertainties in SM predictions arising from uncertain parameterizations and insufficient... Accurate soil moisture(SM)prediction is critical for understanding hydrological processes.Physics-based(PB)models exhibit large uncertainties in SM predictions arising from uncertain parameterizations and insufficient representation of land-surface processes.In addition to PB models,deep learning(DL)models have been widely used in SM predictions recently.However,few pure DL models have notably high success rates due to lacking physical information.Thus,we developed hybrid models to effectively integrate the outputs of PB models into DL models to improve SM predictions.To this end,we first developed a hybrid model based on the attention mechanism to take advantage of PB models at each forecast time scale(attention model).We further built an ensemble model that combined the advantages of different hybrid schemes(ensemble model).We utilized SM forecasts from the Global Forecast System to enhance the convolutional long short-term memory(ConvLSTM)model for 1–16 days of SM predictions.The performances of the proposed hybrid models were investigated and compared with two existing hybrid models.The results showed that the attention model could leverage benefits of PB models and achieved the best predictability of drought events among the different hybrid models.Moreover,the ensemble model performed best among all hybrid models at all forecast time scales and different soil conditions.It is highlighted that the ensemble model outperformed the pure DL model over 79.5%of in situ stations for 16-day predictions.These findings suggest that our proposed hybrid models can adequately exploit the benefits of PB model outputs to aid DL models in making SM predictions. 展开更多
关键词 soil moisture forecasting hybrid model deep learning ConvLSTM attention mechanism
下载PDF
Integrated Machine Learning and Deep Learning Models for Cardiovascular Disease Risk Prediction: A Comprehensive Comparative Study
2
作者 Shadman Mahmood Khan Pathan Sakan Binte Imran 《Journal of Intelligent Learning Systems and Applications》 2024年第1期12-22,共11页
Cardiovascular Diseases (CVDs) pose a significant global health challenge, necessitating accurate risk prediction for effective preventive measures. This comprehensive comparative study explores the performance of tra... Cardiovascular Diseases (CVDs) pose a significant global health challenge, necessitating accurate risk prediction for effective preventive measures. This comprehensive comparative study explores the performance of traditional Machine Learning (ML) and Deep Learning (DL) models in predicting CVD risk, utilizing a meticulously curated dataset derived from health records. Rigorous preprocessing, including normalization and outlier removal, enhances model robustness. Diverse ML models (Logistic Regression, Random Forest, Support Vector Machine, K-Nearest Neighbor, Decision Tree, and Gradient Boosting) are compared with a Long Short-Term Memory (LSTM) neural network for DL. Evaluation metrics include accuracy, ROC AUC, computation time, and memory usage. Results identify the Gradient Boosting Classifier and LSTM as top performers, demonstrating high accuracy and ROC AUC scores. Comparative analyses highlight model strengths and limitations, contributing valuable insights for optimizing predictive strategies. This study advances predictive analytics for cardiovascular health, with implications for personalized medicine. The findings underscore the versatility of intelligent systems in addressing health challenges, emphasizing the broader applications of ML and DL in disease identification beyond cardiovascular health. 展开更多
关键词 Cardiovascular Disease Machine learning deep learning Predictive Modeling Risk Assessment Comparative Analysis Gradient Boosting LSTM
下载PDF
Use of machine learning models for the prognostication of liver transplantation: A systematic review 被引量:2
3
作者 Gidion Chongo Jonathan Soldera 《World Journal of Transplantation》 2024年第1期164-188,共25页
BACKGROUND Liver transplantation(LT)is a life-saving intervention for patients with end-stage liver disease.However,the equitable allocation of scarce donor organs remains a formidable challenge.Prognostic tools are p... BACKGROUND Liver transplantation(LT)is a life-saving intervention for patients with end-stage liver disease.However,the equitable allocation of scarce donor organs remains a formidable challenge.Prognostic tools are pivotal in identifying the most suitable transplant candidates.Traditionally,scoring systems like the model for end-stage liver disease have been instrumental in this process.Nevertheless,the landscape of prognostication is undergoing a transformation with the integration of machine learning(ML)and artificial intelligence models.AIM To assess the utility of ML models in prognostication for LT,comparing their performance and reliability to established traditional scoring systems.METHODS Following the Preferred Reporting Items for Systematic Reviews and Meta-Analysis guidelines,we conducted a thorough and standardized literature search using the PubMed/MEDLINE database.Our search imposed no restrictions on publication year,age,or gender.Exclusion criteria encompassed non-English studies,review articles,case reports,conference papers,studies with missing data,or those exhibiting evident methodological flaws.RESULTS Our search yielded a total of 64 articles,with 23 meeting the inclusion criteria.Among the selected studies,60.8%originated from the United States and China combined.Only one pediatric study met the criteria.Notably,91%of the studies were published within the past five years.ML models consistently demonstrated satisfactory to excellent area under the receiver operating characteristic curve values(ranging from 0.6 to 1)across all studies,surpassing the performance of traditional scoring systems.Random forest exhibited superior predictive capabilities for 90-d mortality following LT,sepsis,and acute kidney injury(AKI).In contrast,gradient boosting excelled in predicting the risk of graft-versus-host disease,pneumonia,and AKI.CONCLUSION This study underscores the potential of ML models in guiding decisions related to allograft allocation and LT,marking a significant evolution in the field of prognostication. 展开更多
关键词 Liver transplantation Machine learning models PROGNOSTICATION Allograft allocation Artificial intelligence
下载PDF
Effectiveness of hybrid ensemble machine learning models for landslide susceptibility analysis:Evidence from Shimla district of North-west Indian Himalayan region
4
作者 SHARMA Aastha SAJJAD Haroon +2 位作者 RAHAMAN Md Hibjur SAHA Tamal Kanti BHUYAN Nirsobha 《Journal of Mountain Science》 SCIE CSCD 2024年第7期2368-2393,共26页
The Indian Himalayan region is frequently experiencing climate change-induced landslides.Thus,landslide susceptibility assessment assumes greater significance for lessening the impact of a landslide hazard.This paper ... The Indian Himalayan region is frequently experiencing climate change-induced landslides.Thus,landslide susceptibility assessment assumes greater significance for lessening the impact of a landslide hazard.This paper makes an attempt to assess landslide susceptibility in Shimla district of the northwest Indian Himalayan region.It examined the effectiveness of random forest(RF),multilayer perceptron(MLP),sequential minimal optimization regression(SMOreg)and bagging ensemble(B-RF,BSMOreg,B-MLP)models.A landslide inventory map comprising 1052 locations of past landslide occurrences was classified into training(70%)and testing(30%)datasets.The site-specific influencing factors were selected by employing a multicollinearity test.The relationship between past landslide occurrences and influencing factors was established using the frequency ratio method.The effectiveness of machine learning models was verified through performance assessors.The landslide susceptibility maps were validated by the area under the receiver operating characteristic curves(ROC-AUC),accuracy,precision,recall and F1-score.The key performance metrics and map validation demonstrated that the BRF model(correlation coefficient:0.988,mean absolute error:0.010,root mean square error:0.058,relative absolute error:2.964,ROC-AUC:0.947,accuracy:0.778,precision:0.819,recall:0.917 and F-1 score:0.865)outperformed the single classifiers and other bagging ensemble models for landslide susceptibility.The results show that the largest area was found under the very high susceptibility zone(33.87%),followed by the low(27.30%),high(20.68%)and moderate(18.16%)susceptibility zones.The factors,namely average annual rainfall,slope,lithology,soil texture and earthquake magnitude have been identified as the influencing factors for very high landslide susceptibility.Soil texture,lineament density and elevation have been attributed to high and moderate susceptibility.Thus,the study calls for devising suitable landslide mitigation measures in the study area.Structural measures,an immediate response system,community participation and coordination among stakeholders may help lessen the detrimental impact of landslides.The findings from this study could aid decision-makers in mitigating future catastrophes and devising suitable strategies in other geographical regions with similar geological characteristics. 展开更多
关键词 Landslide susceptibility Site-specific factors Machine learning models Hybrid ensemble learning Geospatial techniques Himalayan region
下载PDF
Cybernet Model:A New Deep Learning Model for Cyber DDoS Attacks Detection and Recognition
5
作者 Azar Abid Salih Maiwan Bahjat Abdulrazaq 《Computers, Materials & Continua》 SCIE EI 2024年第1期1275-1295,共21页
Cyberspace is extremely dynamic,with new attacks arising daily.Protecting cybersecurity controls is vital for network security.Deep Learning(DL)models find widespread use across various fields,with cybersecurity being... Cyberspace is extremely dynamic,with new attacks arising daily.Protecting cybersecurity controls is vital for network security.Deep Learning(DL)models find widespread use across various fields,with cybersecurity being one of the most crucial due to their rapid cyberattack detection capabilities on networks and hosts.The capabilities of DL in feature learning and analyzing extensive data volumes lead to the recognition of network traffic patterns.This study presents novel lightweight DL models,known as Cybernet models,for the detection and recognition of various cyber Distributed Denial of Service(DDoS)attacks.These models were constructed to have a reasonable number of learnable parameters,i.e.,less than 225,000,hence the name“lightweight.”This not only helps reduce the number of computations required but also results in faster training and inference times.Additionally,these models were designed to extract features in parallel from 1D Convolutional Neural Networks(CNN)and Long Short-Term Memory(LSTM),which makes them unique compared to earlier existing architectures and results in better performance measures.To validate their robustness and effectiveness,they were tested on the CIC-DDoS2019 dataset,which is an imbalanced and large dataset that contains different types of DDoS attacks.Experimental results revealed that bothmodels yielded promising results,with 99.99% for the detectionmodel and 99.76% for the recognition model in terms of accuracy,precision,recall,and F1 score.Furthermore,they outperformed the existing state-of-the-art models proposed for the same task.Thus,the proposed models can be used in cyber security research domains to successfully identify different types of attacks with a high detection and recognition rate. 展开更多
关键词 deep learning CNN LSTM Cybernet model DDoS recognition
下载PDF
Advancing automated pupillometry:a practical deep learning model utilizing infrared pupil images
6
作者 Dai Guangzheng Yu Sile +2 位作者 Liu Ziming Yan Hairu He Xingru 《国际眼科杂志》 CAS 2024年第10期1522-1528,共7页
AIM:To establish pupil diameter measurement algorithms based on infrared images that can be used in real-world clinical settings.METHODS:A total of 188 patients from outpatient clinic at He Eye Specialist Shenyang Hos... AIM:To establish pupil diameter measurement algorithms based on infrared images that can be used in real-world clinical settings.METHODS:A total of 188 patients from outpatient clinic at He Eye Specialist Shenyang Hospital from Spetember to December 2022 were included,and 13470 infrared pupil images were collected for the study.All infrared images for pupil segmentation were labeled using the Labelme software.The computation of pupil diameter is divided into four steps:image pre-processing,pupil identification and localization,pupil segmentation,and diameter calculation.Two major models are used in the computation process:the modified YoloV3 and Deeplabv 3+models,which must be trained beforehand.RESULTS:The test dataset included 1348 infrared pupil images.On the test dataset,the modified YoloV3 model had a detection rate of 99.98% and an average precision(AP)of 0.80 for pupils.The DeeplabV3+model achieved a background intersection over union(IOU)of 99.23%,a pupil IOU of 93.81%,and a mean IOU of 96.52%.The pupil diameters in the test dataset ranged from 20 to 56 pixels,with a mean of 36.06±6.85 pixels.The absolute error in pupil diameters between predicted and actual values ranged from 0 to 7 pixels,with a mean absolute error(MAE)of 1.06±0.96 pixels.CONCLUSION:This study successfully demonstrates a robust infrared image-based pupil diameter measurement algorithm,proven to be highly accurate and reliable for clinical application. 展开更多
关键词 PUPIL infrared image algorithm deep learning model
下载PDF
Prediction of impurity spectrum function by deep learning algorithm
7
作者 刘婷 韩榕生 陈亮 《Chinese Physics B》 SCIE EI CAS CSCD 2024年第5期52-63,共12页
By using the numerical renormalization group(NRG)method,we construct a large dataset with about one million spectral functions of the Anderson quantum impurity model.The dataset contains the density of states(DOS)of t... By using the numerical renormalization group(NRG)method,we construct a large dataset with about one million spectral functions of the Anderson quantum impurity model.The dataset contains the density of states(DOS)of the host material,the strength of Coulomb interaction between on-site electrons(U),and the hybridization between the host material and the impurity site(Γ).The continued DOS and spectral functions are stored with Chebyshev coefficients and wavelet functions,respectively.From this dataset,we build seven different machine learning networks to predict the spectral function from the input data,DOS,U,andΓ.Three different evaluation indexes,mean absolute error(MAE),relative error(RE)and root mean square error(RMSE),are used to analyze the prediction abilities of different network models.Detailed analysis shows that,for the two kinds of widely used recurrent neural networks(RNNs),gate recurrent unit(GRU)has better performance than the long short term memory(LSTM)network.A combination of bidirectional GRU(BiGRU)and GRU has the best performance among GRU,BiGRU,LSTM,and BiLSTM.The MAE peak of BiGRU+GRU reaches 0.00037.We have also tested a one-dimensional convolutional neural network(1DCNN)with 20 hidden layers and a residual neural network(ResNet),we find that the 1DCNN has almost the same performance of the BiGRU+GRU network for the original dataset,while the robustness testing seems to be a little weak than BiGRU+GRU when we test all these models on two other independent datasets.The ResNet has the worst performance among all the seven network models.The datasets presented in this paper,including the large data set of the spectral function of Anderson quantum impurity model,are openly available at https://doi.org/10.57760/sciencedb.j00113.00192. 展开更多
关键词 machine learning Anderson impurity model spectral function
下载PDF
Analysis and comparison of retinal vascular parameters under different glucose metabolic status based on deep learning
8
作者 Yan Jiang Di Gong +7 位作者 Xiao-Hong Chen Lin Yang Jing-Jing Xu Qi-Jie Wei Bin-Bin Chen Yong-Jiang Cai Wen-Qun Xi Zhe Zhang 《International Journal of Ophthalmology(English edition)》 SCIE CAS 2024年第9期1581-1591,共11页
AIM:To develop a deep learning-based model for automatic retinal vascular segmentation,analyzing and comparing parameters under diverse glucose metabolic status(normal,prediabetes,diabetes)and to assess the potential ... AIM:To develop a deep learning-based model for automatic retinal vascular segmentation,analyzing and comparing parameters under diverse glucose metabolic status(normal,prediabetes,diabetes)and to assess the potential of artificial intelligence(AI)in image segmentation and retinal vascular parameters for predicting prediabetes and diabetes.METHODS:Retinal fundus photos from 200 normal individuals,200 prediabetic patients,and 200 diabetic patients(600 eyes in total)were used.The U-Net network served as the foundational architecture for retinal arteryvein segmentation.An automatic segmentation and evaluation system for retinal vascular parameters was trained,encompassing 26 parameters.RESULTS:Significant differences were found in retinal vascular parameters across normal,prediabetes,and diabetes groups,including artery diameter(P=0.008),fractal dimension(P=0.000),vein curvature(P=0.003),C-zone artery branching vessel count(P=0.049),C-zone vein branching vessel count(P=0.041),artery branching angle(P=0.005),vein branching angle(P=0.001),artery angle asymmetry degree(P=0.003),vessel length density(P=0.000),and vessel area density(P=0.000),totaling 10 parameters.CONCLUSION:The deep learning-based model facilitates retinal vascular parameter identification and quantification,revealing significant differences.These parameters exhibit potential as biomarkers for prediabetes and diabetes. 展开更多
关键词 deep learning retinal vascular parameters segmentation model DIABETES PREDIABETES
下载PDF
A deep learning fusion model for accurate classification of brain tumours in Magnetic Resonance images
9
作者 Nechirvan Asaad Zebari Chira Nadheef Mohammed +8 位作者 Dilovan Asaad Zebari Mazin Abed Mohammed Diyar Qader Zeebaree Haydar Abdulameer Marhoon Karrar Hameed Abdulkareem Seifedine Kadry Wattana Viriyasitavat Jan Nedoma Radek Martinek 《CAAI Transactions on Intelligence Technology》 SCIE EI 2024年第4期790-804,共15页
Detecting brain tumours is complex due to the natural variation in their location, shape, and intensity in images. While having accurate detection and segmentation of brain tumours would be beneficial, current methods... Detecting brain tumours is complex due to the natural variation in their location, shape, and intensity in images. While having accurate detection and segmentation of brain tumours would be beneficial, current methods still need to solve this problem despite the numerous available approaches. Precise analysis of Magnetic Resonance Imaging (MRI) is crucial for detecting, segmenting, and classifying brain tumours in medical diagnostics. Magnetic Resonance Imaging is a vital component in medical diagnosis, and it requires precise, efficient, careful, efficient, and reliable image analysis techniques. The authors developed a Deep Learning (DL) fusion model to classify brain tumours reliably. Deep Learning models require large amounts of training data to achieve good results, so the researchers utilised data augmentation techniques to increase the dataset size for training models. VGG16, ResNet50, and convolutional deep belief networks networks extracted deep features from MRI images. Softmax was used as the classifier, and the training set was supplemented with intentionally created MRI images of brain tumours in addition to the genuine ones. The features of two DL models were combined in the proposed model to generate a fusion model, which significantly increased classification accuracy. An openly accessible dataset from the internet was used to test the model's performance, and the experimental results showed that the proposed fusion model achieved a classification accuracy of 98.98%. Finally, the results were compared with existing methods, and the proposed model outperformed them significantly. 展开更多
关键词 brain tumour deep learning feature fusion model MRI images multi‐classification
下载PDF
Advances in Deep-Learning-based Precipitation Nowcasting Techniques
10
作者 ZHENG Qun LIU Qi +1 位作者 LAO Ping LU Zhen-ci 《Journal of Tropical Meteorology》 SCIE 2024年第3期337-350,共14页
Precipitation nowcasting,as a crucial component of weather forecasting,focuses on predicting very short-range precipitation,typically within six hours.This approach relies heavily on real-time observations rather than... Precipitation nowcasting,as a crucial component of weather forecasting,focuses on predicting very short-range precipitation,typically within six hours.This approach relies heavily on real-time observations rather than numerical weather models.The core concept involves the spatio-temporal extrapolation of current precipitation fields derived from ground radar echoes and/or satellite images,which was generally actualized by employing computer image or vision techniques.Recently,with stirring breakthroughs in artificial intelligence(AI)techniques,deep learning(DL)methods have been used as the basis for developing novel approaches to precipitation nowcasting.Notable progress has been obtained in recent years,manifesting the strong potential of DL-based nowcasting models for their advantages in both prediction accuracy and computational cost.This paper provides an overview of these precipitation nowcasting approaches,from which two stages along the advancing in this field emerge.Classic models that were established on an elementary neural network dominated in the first stage,while large meteorological models that were based on complex network architectures prevailed in the second.In particular,the nowcasting accuracy of such data-driven models has been greatly increased by imposing suitable physical constraints.The integration of AI models and physical models seems to be a promising way to improve precipitation nowcasting techniques further. 展开更多
关键词 precipitation nowcasting deep learning neural network classic model large model
下载PDF
Mesh representation matters:investigating the influence of different mesh features on perceptual and spatial fidelity of deep 3D morphable models
11
作者 Robert KOSK Richard SOUTHERN +3 位作者 Lihua YOU Shaojun BIAN Willem KOKKE Greg MAGUIRE 《虚拟现实与智能硬件(中英文)》 EI 2024年第5期383-395,共13页
Background Deep 3D morphable models(deep 3DMMs)play an essential role in computer vision.They are used in facial synthesis,compression,reconstruction and animation,avatar creation,virtual try-on,facial recognition sys... Background Deep 3D morphable models(deep 3DMMs)play an essential role in computer vision.They are used in facial synthesis,compression,reconstruction and animation,avatar creation,virtual try-on,facial recognition systems and medical imaging.These applications require high spatial and perceptual quality of synthesised meshes.Despite their significance,these models have not been compared with different mesh representations and evaluated jointly with point-wise distance and perceptual metrics.Methods We compare the influence of different mesh representation features to various deep 3DMMs on spatial and perceptual fidelity of the reconstructed meshes.This paper proves the hypothesis that building deep 3DMMs from meshes represented with global representations leads to lower spatial reconstruction error measured with L_(1) and L_(2) norm metrics and underperforms on perceptual metrics.In contrast,using differential mesh representations which describe differential surface properties yields lower perceptual FMPD and DAME and higher spatial fidelity error.The influence of mesh feature normalisation and standardisation is also compared and analysed from perceptual and spatial fidelity perspectives.Results The results presented in this paper provide guidance in selecting mesh representations to build deep 3DMMs accordingly to spatial and perceptual quality objectives and propose combinations of mesh representations and deep 3DMMs which improve either perceptual or spatial fidelity of existing methods. 展开更多
关键词 Shape modelling deep 3D morphable models Representation learning Feature engineering Perceptual metrics
下载PDF
Low rank optimization for efficient deep learning:making a balance between compact architecture and fast training
12
作者 OU Xinwei CHEN Zhangxin +1 位作者 ZHU Ce LIU Yipeng 《Journal of Systems Engineering and Electronics》 SCIE CSCD 2024年第3期509-531,F0002,共24页
Deep neural networks(DNNs)have achieved great success in many data processing applications.However,high computational complexity and storage cost make deep learning difficult to be used on resource-constrained devices... Deep neural networks(DNNs)have achieved great success in many data processing applications.However,high computational complexity and storage cost make deep learning difficult to be used on resource-constrained devices,and it is not environmental-friendly with much power cost.In this paper,we focus on low-rank optimization for efficient deep learning techniques.In the space domain,DNNs are compressed by low rank approximation of the network parameters,which directly reduces the storage requirement with a smaller number of network parameters.In the time domain,the network parameters can be trained in a few subspaces,which enables efficient training for fast convergence.The model compression in the spatial domain is summarized into three categories as pre-train,pre-set,and compression-aware methods,respectively.With a series of integrable techniques discussed,such as sparse pruning,quantization,and entropy coding,we can ensemble them in an integration framework with lower computational complexity and storage.In addition to summary of recent technical advances,we have two findings for motivating future works.One is that the effective rank,derived from the Shannon entropy of the normalized singular values,outperforms other conventional sparse measures such as the?_1 norm for network compression.The other is a spatial and temporal balance for tensorized neural networks.For accelerating the training of tensorized neural networks,it is crucial to leverage redundancy for both model compression and subspace training. 展开更多
关键词 model compression subspace training effective rank low rank tensor optimization efficient deep learning
下载PDF
Prognostic prediction models for postoperative patients with stageⅠtoⅢcolorectal cancer based on machine learning
13
作者 Xiao-Lin Ji Shuo Xu +5 位作者 Xiao-Yu Li Jin-Huan Xu Rong-Shuang Han Ying-Jie Guo Li-Ping Duan Zi-Bin Tian 《World Journal of Gastrointestinal Oncology》 SCIE 2024年第12期4597-4613,共17页
BACKGROUND Colorectal cancer(CRC)is characterized by high heterogeneity,aggressiveness,and high morbidity and mortality rates.With machine learning(ML)algorithms,patient,tumor,and treatment features can be used to dev... BACKGROUND Colorectal cancer(CRC)is characterized by high heterogeneity,aggressiveness,and high morbidity and mortality rates.With machine learning(ML)algorithms,patient,tumor,and treatment features can be used to develop and validate models for predicting survival.In addition,important variables can be screened and different applications can be provided that could serve as vital references when making clinical decisions and potentially improving patient outcomes in clinical settings.AIM To construct prognostic prediction models and screen important variables for patients with stageⅠtoⅢCRC.METHODS More than 1000 postoperative CRC patients were grouped according to survival time(with cutoff values of 3 years and 5 years)and assigned to training and testing cohorts(7:3).For each 3-category survival time,predictions were made by 4 ML algorithms(all-variable and important variable-only datasets),each of which was validated via 5-fold cross-validation and bootstrap validation.Important variables were screened with multivariable regression methods.Model performance was evaluated and compared before and after variable screening with the area under the curve(AUC).SHapley Additive exPlanations(SHAP)further demonstrated the impact of important variables on model decision-making.Nomograms were constructed for practical model application.RESULTS Our ML models performed well;the model performance before and after important parameter identification was consistent,and variable screening was effective.The highest pre-and postscreening model AUCs 95%confidence intervals in the testing set were 0.87(0.81-0.92)and 0.89(0.84-0.93)for overall survival,0.75(0.69-0.82)and 0.73(0.64-0.81)for disease-free survival,0.95(0.88-1.00)and 0.88(0.75-0.97)for recurrence-free survival,and 0.76(0.47-0.95)and 0.80(0.53-0.94)for distant metastasis-free survival.Repeated cross-validation and bootstrap validation were performed in both the training and testing datasets.The SHAP values of the important variables were consistent with the clinicopathological characteristics of patients with tumors.The nomograms were created.CONCLUSION We constructed a comprehensive,high-accuracy,important variable-based ML architecture for predicting the 3-category survival times.This architecture could serve as a vital reference for managing CRC patients. 展开更多
关键词 Colorectal cancer Machine learning Prognostic prediction model Survival times Important variables
下载PDF
Unified deep learning model for predicting fundus fluorescein angiography image from fundus structure image
14
作者 Yiwei Chen Yi He +3 位作者 Hong Ye Lina Xing Xin Zhang Guohua Shi 《Journal of Innovative Optical Health Sciences》 SCIE EI CSCD 2024年第3期105-113,共9页
The prediction of fundus fluorescein angiography(FFA)images from fundus structural images is a cutting-edge research topic in ophthalmological image processing.Prediction comprises estimating FFA from fundus camera im... The prediction of fundus fluorescein angiography(FFA)images from fundus structural images is a cutting-edge research topic in ophthalmological image processing.Prediction comprises estimating FFA from fundus camera imaging,single-phase FFA from scanning laser ophthalmoscopy(SLO),and three-phase FFA also from SLO.Although many deep learning models are available,a single model can only perform one or two of these prediction tasks.To accomplish three prediction tasks using a unified method,we propose a unified deep learning model for predicting FFA images from fundus structure images using a supervised generative adversarial network.The three prediction tasks are processed as follows:data preparation,network training under FFA supervision,and FFA image prediction from fundus structure images on a test set.By comparing the FFA images predicted by our model,pix2pix,and CycleGAN,we demonstrate the remarkable progress achieved by our proposal.The high performance of our model is validated in terms of the peak signal-to-noise ratio,structural similarity index,and mean squared error. 展开更多
关键词 Fundus fluorescein angiography image fundus structure image image translation unified deep learning model generative adversarial networks
下载PDF
Predictive value of machine learning models for lymph node metastasis in gastric cancer: A two-center study
15
作者 Tong Lu Miao Lu +4 位作者 Dong Wu Yuan-Yuan Ding Hao-Nan Liu Tao-Tao Li Da-Qing Song 《World Journal of Gastrointestinal Surgery》 SCIE 2024年第1期85-94,共10页
BACKGROUND Gastric cancer is one of the most common malignant tumors in the digestive system,ranking sixth in incidence and fourth in mortality worldwide.Since 42.5%of metastatic lymph nodes in gastric cancer belong t... BACKGROUND Gastric cancer is one of the most common malignant tumors in the digestive system,ranking sixth in incidence and fourth in mortality worldwide.Since 42.5%of metastatic lymph nodes in gastric cancer belong to nodule type and peripheral type,the application of imaging diagnosis is restricted.AIM To establish models for predicting the risk of lymph node metastasis in gastric cancer patients using machine learning(ML)algorithms and to evaluate their pre-dictive performance in clinical practice.METHODS Data of a total of 369 patients who underwent radical gastrectomy at the Depart-ment of General Surgery of Affiliated Hospital of Xuzhou Medical University(Xuzhou,China)from March 2016 to November 2019 were collected and retro-spectively analyzed as the training group.In addition,data of 123 patients who underwent radical gastrectomy at the Department of General Surgery of Jining First People’s Hospital(Jining,China)were collected and analyzed as the verifi-cation group.Seven ML models,including decision tree,random forest,support vector machine(SVM),gradient boosting machine,naive Bayes,neural network,and logistic regression,were developed to evaluate the occurrence of lymph node metastasis in patients with gastric cancer.The ML models were established fo-llowing ten cross-validation iterations using the training dataset,and subsequently,each model was assessed using the test dataset.The models’performance was evaluated by comparing the area under the receiver operating characteristic curve of each model.RESULTS Among the seven ML models,except for SVM,the other ones exhibited higher accuracy and reliability,and the influences of various risk factors on the models are intuitive.CONCLUSION The ML models developed exhibit strong predictive capabilities for lymph node metastasis in gastric cancer,which can aid in personalized clinical diagnosis and treatment. 展开更多
关键词 Machine learning Prediction model Gastric cancer Lymph node metastasis
下载PDF
Comparative study of different machine learning models in landslide susceptibility assessment: A case study of Conghua District, Guangzhou, China
16
作者 Ao Zhang Xin-wen Zhao +8 位作者 Xing-yuezi Zhao Xiao-zhan Zheng Min Zeng Xuan Huang Pan Wu Tuo Jiang Shi-chang Wang Jun He Yi-yong Li 《China Geology》 CAS CSCD 2024年第1期104-115,共12页
Machine learning is currently one of the research hotspots in the field of landslide prediction.To clarify and evaluate the differences in characteristics and prediction effects of different machine learning models,Co... Machine learning is currently one of the research hotspots in the field of landslide prediction.To clarify and evaluate the differences in characteristics and prediction effects of different machine learning models,Conghua District,which is the most prone to landslide disasters in Guangzhou,was selected for landslide susceptibility evaluation.The evaluation factors were selected by using correlation analysis and variance expansion factor method.Applying four machine learning methods namely Logistic Regression(LR),Random Forest(RF),Support Vector Machines(SVM),and Extreme Gradient Boosting(XGB),landslide models were constructed.Comparative analysis and evaluation of the model were conducted through statistical indices and receiver operating characteristic(ROC)curves.The results showed that LR,RF,SVM,and XGB models have good predictive performance for landslide susceptibility,with the area under curve(AUC)values of 0.752,0.965,0.996,and 0.998,respectively.XGB model had the highest predictive ability,followed by RF model,SVM model,and LR model.The frequency ratio(FR)accuracy of LR,RF,SVM,and XGB models was 0.775,0.842,0.759,and 0.822,respectively.RF and XGB models were superior to LR and SVM models,indicating that the integrated algorithm has better predictive ability than a single classification algorithm in regional landslide classification problems. 展开更多
关键词 Landslides susceptibility assessment Machine learning Logistic Regression Random Forest Support Vector Machines XGBoost Assessment model Geological disaster investigation and prevention engineering
下载PDF
Surface Defect Detection and Evaluation Method of Large Wind Turbine Blades Based on an Improved Deeplabv3+Deep Learning Model
17
作者 Wanrun Li Wenhai Zhao +1 位作者 Tongtong Wang Yongfeng Du 《Structural Durability & Health Monitoring》 EI 2024年第5期553-575,共23页
The accumulation of defects on wind turbine blade surfaces can lead to irreversible damage,impacting the aero-dynamic performance of the blades.To address the challenge of detecting and quantifying surface defects on ... The accumulation of defects on wind turbine blade surfaces can lead to irreversible damage,impacting the aero-dynamic performance of the blades.To address the challenge of detecting and quantifying surface defects on wind turbine blades,a blade surface defect detection and quantification method based on an improved Deeplabv3+deep learning model is proposed.Firstly,an improved method for wind turbine blade surface defect detection,utilizing Mobilenetv2 as the backbone feature extraction network,is proposed based on an original Deeplabv3+deep learning model to address the issue of limited robustness.Secondly,through integrating the concept of pre-trained weights from transfer learning and implementing a freeze training strategy,significant improvements have been made to enhance both the training speed and model training accuracy of this deep learning model.Finally,based on segmented blade surface defect images,a method for quantifying blade defects is proposed.This method combines image stitching algorithms to achieve overall quantification and risk assessment of the entire blade.Test results show that the improved Deeplabv3+deep learning model reduces training time by approximately 43.03%compared to the original model,while achieving mAP and MIoU values of 96.87%and 96.93%,respectively.Moreover,it demonstrates robustness in detecting different surface defects on blades across different back-grounds.The application of a blade surface defect quantification method enables the precise quantification of dif-ferent defects and facilitates the assessment of risk levels associated with defect measurements across the entire blade.This method enables non-contact,long-distance,high-precision detection and quantification of surface defects on the blades,providing a reference for assessing surface defects on wind turbine blades. 展开更多
关键词 Structural health monitoring computer vision blade surface defects detection deeplabv3+ deep learning model
下载PDF
Extensive identification of landslide boundaries using remote sensing images and deep learning method
18
作者 Chang-dong Li Peng-fei Feng +3 位作者 Xi-hui Jiang Shuang Zhang Jie Meng Bing-chen Li 《China Geology》 CAS CSCD 2024年第2期277-290,共14页
The frequent occurrence of extreme weather events has rendered numerous landslides to a global natural disaster issue.It is crucial to rapidly and accurately determine the boundaries of landslides for geohazards evalu... The frequent occurrence of extreme weather events has rendered numerous landslides to a global natural disaster issue.It is crucial to rapidly and accurately determine the boundaries of landslides for geohazards evaluation and emergency response.Therefore,the Skip Connection DeepLab neural network(SCDnn),a deep learning model based on 770 optical remote sensing images of landslide,is proposed to improve the accuracy of landslide boundary detection.The SCDnn model is optimized for the over-segmentation issue which occurs in conventional deep learning models when there is a significant degree of similarity between topographical geomorphic features.SCDnn exhibits notable improvements in landslide feature extraction and semantic segmentation by combining an enhanced Atrous Spatial Pyramid Convolutional Block(ASPC)with a coding structure that reduces model complexity.The experimental results demonstrate that SCDnn can identify landslide boundaries in 119 images with MIoU values between 0.8and 0.9;while 52 images with MIoU values exceeding 0.9,which exceeds the identification accuracy of existing techniques.This work can offer a novel technique for the automatic extensive identification of landslide boundaries in remote sensing images in addition to establishing the groundwork for future inve stigations and applications in related domains. 展开更多
关键词 GEOHAZARD Landslide boundary detection Remote sensing image deep learning model Steep slope Large annual rainfall Human settlements INFRASTRUCTURE Agricultural land Eastern Tibetan Plateau Geological hazards survey engineering
下载PDF
A Survey on Deep Learning-Based 2D Human Pose Estimation Models
19
作者 Sani Salisu A.S.A.Mohamed +2 位作者 M.H.Jaafar Ainun S.B.Pauzi Hussain A.Younis 《Computers, Materials & Continua》 SCIE EI 2023年第8期2385-2400,共16页
In this article,a comprehensive survey of deep learning-based(DLbased)human pose estimation(HPE)that can help researchers in the domain of computer vision is presented.HPE is among the fastest-growing research domains... In this article,a comprehensive survey of deep learning-based(DLbased)human pose estimation(HPE)that can help researchers in the domain of computer vision is presented.HPE is among the fastest-growing research domains of computer vision and is used in solving several problems for human endeavours.After the detailed introduction,three different human body modes followed by the main stages of HPE and two pipelines of twodimensional(2D)HPE are presented.The details of the four components of HPE are also presented.The keypoints output format of two popular 2D HPE datasets and the most cited DL-based HPE articles from the year of breakthrough are both shown in tabular form.This study intends to highlight the limitations of published reviews and surveys respecting presenting a systematic review of the current DL-based solution to the 2D HPE model.Furthermore,a detailed and meaningful survey that will guide new and existing researchers on DL-based 2D HPE models is achieved.Finally,some future research directions in the field of HPE,such as limited data on disabled persons and multi-training DL-based models,are revealed to encourage researchers and promote the growth of HPE research. 展开更多
关键词 Human pose estimation deep learning 2D DATASET models body parts
下载PDF
Hemodynamic Analysis and Diagnosis Based on Multi-Deep Learning Models
20
作者 Xing Deng Feipeng Da Haijian Shao 《Fluid Dynamics & Materials Processing》 EI 2023年第6期1369-1383,共15页
This study employs nine distinct deep learning models to categorize 12,444 blood cell images and automatically extract from them relevant information with an accuracy that is beyond that achievable with traditional te... This study employs nine distinct deep learning models to categorize 12,444 blood cell images and automatically extract from them relevant information with an accuracy that is beyond that achievable with traditional techniques.The work is intended to improve current methods for the assessment of human health through measurement of the distribution of four types of blood cells,namely,eosinophils,neutrophils,monocytes,and lymphocytes,known for their relationship with human body damage,inflammatory regions,and organ illnesses,in particular,and with the health of the immune system and other hazards,such as cardiovascular disease or infections,more in general.The results of the experiments show that the deep learning models can automatically extract features from the blood cell images and properly classify them with an accuracy of 98%,97%,and 89%,respectively,with regard to the training,verification,and testing of the corresponding datasets. 展开更多
关键词 Blood cell analysis deep learning models classification-detection
下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部