Huntington'sdisease(HD)isahereditary neurodegenerative disorder for which there is currently no effectivetreatmentavailable.Consequently,the development of appropriate disease models is critical to thoroughly inve...Huntington'sdisease(HD)isahereditary neurodegenerative disorder for which there is currently no effectivetreatmentavailable.Consequently,the development of appropriate disease models is critical to thoroughly investigate disease progression.The genetic basis of HD involves the abnormal expansion of CAG repeats in the huntingtin(HTT)gene,leading to the expansion of a polyglutamine repeat in the HTT protein.Mutant HTT carrying the expanded polyglutamine repeat undergoes misfolding and forms aggregates in the brain,which precipitate selective neuronal loss in specific brain regions.Animal models play an important role in elucidating the pathogenesis of neurodegenerative disorders such as HD and in identifying potential therapeutic targets.Due to the marked species differences between rodents and larger animals,substantial efforts have been directed toward establishing large animal models for HD research.These models are pivotal for advancing the discovery of novel therapeutic targets,enhancing effective drug delivery methods,and improving treatment outcomes.We have explored the advantages of utilizing large animal models,particularly pigs,in previous reviews.Since then,however,significant progress has been made in developing more sophisticated animal models that faithfully replicate the typical pathology of HD.In the current review,we provide a comprehensive overview of large animal models of HD,incorporating recent findings regarding the establishment of HD knock-in(KI)pigs and their genetic therapy.We also explore the utilization of large animal models in HD research,with a focus on sheep,non-human primates(NHPs),and pigs.Our objective is to provide valuable insights into the application of these large animal models for the investigation and treatment of neurodegenerative disorders.展开更多
Hereditary hearing loss(HHL),a genetic disorder that impairs auditory function,significantly affects quality of life and incurs substantial economic losses for society.To investigate the underlying causes of HHL and e...Hereditary hearing loss(HHL),a genetic disorder that impairs auditory function,significantly affects quality of life and incurs substantial economic losses for society.To investigate the underlying causes of HHL and evaluate therapeutic outcomes,appropriate animal models are necessary.Pigs have been extensively used as valuable large animal models in biomedical research.In this review,we highlight the advantages of pig models in terms of ear anatomy,inner ear morphology,and electrophysiological characteristics,as well as recent advancements in the development of distinct genetically modified porcine models of hearing loss.Additionally,we discuss the prospects,challenges,and recommendations regarding the use pig models in HHL research.Overall,this review provides insights and perspectives for future studies on HHL using porcine models.展开更多
Neurodegenerative diseases(NDs)are a group of debilitating neurological disorders that primarily affect elderly populations and include Alzheimer's disease(AD),Parkinson's disease(PD),Huntington's disease(...Neurodegenerative diseases(NDs)are a group of debilitating neurological disorders that primarily affect elderly populations and include Alzheimer's disease(AD),Parkinson's disease(PD),Huntington's disease(HD),and amyotrophic lateral sclerosis(ALS).Currently,there are no therapies available that can delay,stop,or reverse the pathological progression of NDs in clinical settings.As the population ages,NDs are imposing a huge burden on public health systems and affected families.Animal models are important tools for preclinical investigations to understand disease pathogenesis and test potential treatments.While numerous rodent models of NDs have been developed to enhance our understanding of disease mechanisms,the limited success of translating findings from animal models to clinical practice suggests that there is still a need to bridge this translation gap.Old World nonhuman primates(NHPs),such as rhesus,cynomolgus,and vervet monkeys,are phylogenetically,physiologically,biochemically,and behaviorally most relevant to humans.This is particularly evident in the similarity of the structure and function of their central nervous systems,rendering such species uniquely valuable for neuroscience research.Recently,the development of several genetically modified NHP models of NDs has successfully recapitulated key pathologies and revealed novel mechanisms.This review focuses on the efficacy of NHPs in modeling NDs and the novel pathological insights gained,as well as the challenges associated with the generation of such models and the complexities involved in their subsequent analysis.展开更多
BACKGROUND Liver transplantation(LT)is a life-saving intervention for patients with end-stage liver disease.However,the equitable allocation of scarce donor organs remains a formidable challenge.Prognostic tools are p...BACKGROUND Liver transplantation(LT)is a life-saving intervention for patients with end-stage liver disease.However,the equitable allocation of scarce donor organs remains a formidable challenge.Prognostic tools are pivotal in identifying the most suitable transplant candidates.Traditionally,scoring systems like the model for end-stage liver disease have been instrumental in this process.Nevertheless,the landscape of prognostication is undergoing a transformation with the integration of machine learning(ML)and artificial intelligence models.AIM To assess the utility of ML models in prognostication for LT,comparing their performance and reliability to established traditional scoring systems.METHODS Following the Preferred Reporting Items for Systematic Reviews and Meta-Analysis guidelines,we conducted a thorough and standardized literature search using the PubMed/MEDLINE database.Our search imposed no restrictions on publication year,age,or gender.Exclusion criteria encompassed non-English studies,review articles,case reports,conference papers,studies with missing data,or those exhibiting evident methodological flaws.RESULTS Our search yielded a total of 64 articles,with 23 meeting the inclusion criteria.Among the selected studies,60.8%originated from the United States and China combined.Only one pediatric study met the criteria.Notably,91%of the studies were published within the past five years.ML models consistently demonstrated satisfactory to excellent area under the receiver operating characteristic curve values(ranging from 0.6 to 1)across all studies,surpassing the performance of traditional scoring systems.Random forest exhibited superior predictive capabilities for 90-d mortality following LT,sepsis,and acute kidney injury(AKI).In contrast,gradient boosting excelled in predicting the risk of graft-versus-host disease,pneumonia,and AKI.CONCLUSION This study underscores the potential of ML models in guiding decisions related to allograft allocation and LT,marking a significant evolution in the field of prognostication.展开更多
The Indian Himalayan region is frequently experiencing climate change-induced landslides.Thus,landslide susceptibility assessment assumes greater significance for lessening the impact of a landslide hazard.This paper ...The Indian Himalayan region is frequently experiencing climate change-induced landslides.Thus,landslide susceptibility assessment assumes greater significance for lessening the impact of a landslide hazard.This paper makes an attempt to assess landslide susceptibility in Shimla district of the northwest Indian Himalayan region.It examined the effectiveness of random forest(RF),multilayer perceptron(MLP),sequential minimal optimization regression(SMOreg)and bagging ensemble(B-RF,BSMOreg,B-MLP)models.A landslide inventory map comprising 1052 locations of past landslide occurrences was classified into training(70%)and testing(30%)datasets.The site-specific influencing factors were selected by employing a multicollinearity test.The relationship between past landslide occurrences and influencing factors was established using the frequency ratio method.The effectiveness of machine learning models was verified through performance assessors.The landslide susceptibility maps were validated by the area under the receiver operating characteristic curves(ROC-AUC),accuracy,precision,recall and F1-score.The key performance metrics and map validation demonstrated that the BRF model(correlation coefficient:0.988,mean absolute error:0.010,root mean square error:0.058,relative absolute error:2.964,ROC-AUC:0.947,accuracy:0.778,precision:0.819,recall:0.917 and F-1 score:0.865)outperformed the single classifiers and other bagging ensemble models for landslide susceptibility.The results show that the largest area was found under the very high susceptibility zone(33.87%),followed by the low(27.30%),high(20.68%)and moderate(18.16%)susceptibility zones.The factors,namely average annual rainfall,slope,lithology,soil texture and earthquake magnitude have been identified as the influencing factors for very high landslide susceptibility.Soil texture,lineament density and elevation have been attributed to high and moderate susceptibility.Thus,the study calls for devising suitable landslide mitigation measures in the study area.Structural measures,an immediate response system,community participation and coordination among stakeholders may help lessen the detrimental impact of landslides.The findings from this study could aid decision-makers in mitigating future catastrophes and devising suitable strategies in other geographical regions with similar geological characteristics.展开更多
El Niño-Southern Oscillation(ENSO)is the strongest interannual climate mode influencing the coupled ocean-atmosphere system in the tropical Pacific,and numerous dynamical and statistical models have been develope...El Niño-Southern Oscillation(ENSO)is the strongest interannual climate mode influencing the coupled ocean-atmosphere system in the tropical Pacific,and numerous dynamical and statistical models have been developed to simulate and predict it.In some simplified coupled ocean-atmosphere models,the relationship between sea surface temperature(SST)anomalies and wind stress(τ)anomalies can be constructed by statistical methods,such as singular value decomposition(SVD).In recent years,the applications of artificial intelligence(AI)to climate modeling have shown promising prospects,and the integrations of AI-based models with dynamical models are active areas of research.This study constructs U-Net models for representing the relationship between SSTAs andτanomalies in the tropical Pacific;the UNet-derivedτmodel,denoted asτUNet,is then used to replace the original SVD-basedτmodel of an intermediate coupled model(ICM),forming a newly AI-integrated ICM,referred to as ICM-UNet.The simulation results obtained from ICM-UNet demonstrate their ability to represent the spatiotemporal variability of oceanic and atmospheric anomaly fields in the equatorial Pacific.In the ocean-only case study,theτUNet-derived wind stress anomaly fields are used to force the ocean component of the ICM,the results of which also indicate reasonable simulations of typical ENSO events.These results demonstrate the feasibility of integrating an AI-derived model with a physics-based dynamical model for ENSO modeling studies.Furthermore,the successful integration of the dynamical ocean models with the AI-based atmospheric wind model provides a novel approach to ocean-atmosphere interaction modeling studies.展开更多
Postoperative pancreatic fistula(POPF)is a frequent complication after pancre-atectomy,leading to increased morbidity and mortality.Optimizing prediction models for POPF has emerged as a critical focus in surgical res...Postoperative pancreatic fistula(POPF)is a frequent complication after pancre-atectomy,leading to increased morbidity and mortality.Optimizing prediction models for POPF has emerged as a critical focus in surgical research.Although over sixty models following pancreaticoduodenectomy,predominantly reliant on a variety of clinical,surgical,and radiological parameters,have been documented,their predictive accuracy remains suboptimal in external validation and across diverse populations.As models after distal pancreatectomy continue to be pro-gressively reported,their external validation is eagerly anticipated.Conversely,POPF prediction after central pancreatectomy is in its nascent stage,warranting urgent need for further development and validation.The potential of machine learning and big data analytics offers promising prospects for enhancing the accuracy of prediction models by incorporating an extensive array of variables and optimizing algorithm performance.Moreover,there is potential for the development of personalized prediction models based on patient-or pancreas-specific factors and postoperative serum or drain fluid biomarkers to improve accuracy in identifying individuals at risk of POPF.In the future,prospective multicenter studies and the integration of novel imaging technologies,such as artificial intelligence-based radiomics,may further refine predictive models.Addressing these issues is anticipated to revolutionize risk stratification,clinical decision-making,and postoperative management in patients undergoing pancre-atectomy.展开更多
The problematic use of social media has numerous negative impacts on individuals'daily lives,interpersonal relationships,physical and mental health,and more.Currently,there are few methods and tools to alleviate p...The problematic use of social media has numerous negative impacts on individuals'daily lives,interpersonal relationships,physical and mental health,and more.Currently,there are few methods and tools to alleviate problematic social media,and their potential is yet to be fully realized.Emerging large language models(LLMs)are becoming increasingly popular for providing information and assistance to people and are being applied in many aspects of life.In mitigating problematic social media use,LLMs such as ChatGPT can play a positive role by serving as conversational partners and outlets for users,providing personalized information and resources,monitoring and intervening in problematic social media use,and more.In this process,we should recognize both the enormous potential and endless possibilities of LLMs such as ChatGPT,leveraging their advantages to better address problematic social media use,while also acknowledging the limitations and potential pitfalls of ChatGPT technology,such as errors,limitations in issue resolution,privacy and security concerns,and potential overreliance.When we leverage the advantages of LLMs to address issues in social media usage,we must adopt a cautious and ethical approach,being vigilant of the potential adverse effects that LLMs may have in addressing problematic social media use to better harness technology to serve individuals and society.展开更多
BACKGROUND Gastric cancer is one of the most common malignant tumors in the digestive system,ranking sixth in incidence and fourth in mortality worldwide.Since 42.5%of metastatic lymph nodes in gastric cancer belong t...BACKGROUND Gastric cancer is one of the most common malignant tumors in the digestive system,ranking sixth in incidence and fourth in mortality worldwide.Since 42.5%of metastatic lymph nodes in gastric cancer belong to nodule type and peripheral type,the application of imaging diagnosis is restricted.AIM To establish models for predicting the risk of lymph node metastasis in gastric cancer patients using machine learning(ML)algorithms and to evaluate their pre-dictive performance in clinical practice.METHODS Data of a total of 369 patients who underwent radical gastrectomy at the Depart-ment of General Surgery of Affiliated Hospital of Xuzhou Medical University(Xuzhou,China)from March 2016 to November 2019 were collected and retro-spectively analyzed as the training group.In addition,data of 123 patients who underwent radical gastrectomy at the Department of General Surgery of Jining First People’s Hospital(Jining,China)were collected and analyzed as the verifi-cation group.Seven ML models,including decision tree,random forest,support vector machine(SVM),gradient boosting machine,naive Bayes,neural network,and logistic regression,were developed to evaluate the occurrence of lymph node metastasis in patients with gastric cancer.The ML models were established fo-llowing ten cross-validation iterations using the training dataset,and subsequently,each model was assessed using the test dataset.The models’performance was evaluated by comparing the area under the receiver operating characteristic curve of each model.RESULTS Among the seven ML models,except for SVM,the other ones exhibited higher accuracy and reliability,and the influences of various risk factors on the models are intuitive.CONCLUSION The ML models developed exhibit strong predictive capabilities for lymph node metastasis in gastric cancer,which can aid in personalized clinical diagnosis and treatment.展开更多
Drone or unmanned aerial vehicle(UAV)technology has undergone significant changes.The technology allows UAV to carry out a wide range of tasks with an increasing level of sophistication,since drones can cover a large ...Drone or unmanned aerial vehicle(UAV)technology has undergone significant changes.The technology allows UAV to carry out a wide range of tasks with an increasing level of sophistication,since drones can cover a large area with cameras.Meanwhile,the increasing number of computer vision applications utilizing deep learning provides a unique insight into such applications.The primary target in UAV-based detection applications is humans,yet aerial recordings are not included in the massive datasets used to train object detectors,which makes it necessary to gather the model data from such platforms.You only look once(YOLO)version 4,RetinaNet,faster region-based convolutional neural network(R-CNN),and cascade R-CNN are several well-known detectors that have been studied in the past using a variety of datasets to replicate rescue scenes.Here,we used the search and rescue(SAR)dataset to train the you only look once version 5(YOLOv5)algorithm to validate its speed,accuracy,and low false detection rate.In comparison to YOLOv4 and R-CNN,the highest mean average accuracy of 96.9%is obtained by YOLOv5.For comparison,experimental findings utilizing the SAR and the human rescue imaging database on land(HERIDAL)datasets are presented.The results show that the YOLOv5-based approach is the most successful human detection model for SAR missions.展开更多
Sunshine duration (S) based empirical equations have been employed in this study to estimate the daily global solar radiation on a horizontal surface (G) for six meteorological stations in Burundi. Those equations inc...Sunshine duration (S) based empirical equations have been employed in this study to estimate the daily global solar radiation on a horizontal surface (G) for six meteorological stations in Burundi. Those equations include the Ångström-Prescott linear model and four amongst its derivatives, i.e. logarithmic, exponential, power and quadratic functions. Monthly mean values of daily global solar radiation and sunshine duration data for a period of 20 to 23 years, from the Geographical Institute of Burundi (IGEBU), have been used. For any of the six stations, ten single or double linear regressions have been developed from the above-said five functions, to relate in terms of monthly mean values, the daily clearness index () to each of the next two kinds of relative sunshine duration (RSD): and . In those ratios, G<sub>0</sub>, S<sub>0 </sub>and stand for the extraterrestrial daily solar radiation on a horizontal surface, the day length and the modified day length taking into account the natural site’s horizon, respectively. According to the calculated mean values of the clearness index and the RSD, each station experiences a high number of fairly clear (or partially cloudy) days. Estimated values of the dependent variable (y) in each developed linear regression, have been compared to measured values in terms of the coefficients of correlation (R) and of determination (R<sub>2</sub>), the mean bias error (MBE), the root mean square error (RMSE) and the t-statistics. Mean values of these statistical indicators have been used to rank, according to decreasing performance level, firstly the ten developed equations per station on account of the overall six stations, secondly the six stations on account of the overall ten equations. Nevertheless, the obtained values of those indicators lay in the next ranges for all the developed sixty equations:;;;, with . These results lead to assert that any of the sixty developed linear regressions (and thus equations in terms of and ), fits very adequately measured data, and should be used to estimate monthly average daily global solar radiation with sunshine duration for the relevant station. It is also found that using as RSD, is slightly more advantageous than using for estimating the monthly average daily clearness index, . Moreover, values of statistical indicators of this study match adequately data from other works on the same kinds of empirical equations.展开更多
Tree height (H) in a natural stand or forest plantation is a fundamental variable in management, and the use of mathematical expressions that estimate H as a function of diameter at breast height (d) or variables at t...Tree height (H) in a natural stand or forest plantation is a fundamental variable in management, and the use of mathematical expressions that estimate H as a function of diameter at breast height (d) or variables at the stand level is a valuable support tool in forest inventories. The objective was to fit and propose a generalized H-d model for Pinus montezumae and Pinus pseudostrobus established in forest plantations of Nuevo San Juan Parangaricutiro, Michoacan, Mexico. Using nonlinear least squares (NLS), 10 generalized H-d models were fitted to 883 and 1226 pairs of H-d data from Pinus montezumae and Pinus pseudostrobus, respectively. The best model was refitted with the maximum likelihood mixed effects model (MEM) approach by including the site as a classification variable and a known variance structure. The Wang and Tang equation was selected as the best model with NLS;the MEM with an additive effect on two of its parameters and an exponential variance function improved the fit statistics for Pinus montezumae and Pinus pseudostrobus, respectively. The model validation showed equality of means among the estimates for both species and an independent subsample. The calibration of the MEM at the plot level was efficient and might increase the applicability of these results. The inclusion of dominant height in the MEM approach helped to reduce bias in the estimates and also to better explain the variability among plots.展开更多
In this work, four empirical models of statistical thickness, namely the models of Harkins and Jura, Hasley, Carbon Black and Jaroniec, were compared in order to determine the textural properties (external surface and...In this work, four empirical models of statistical thickness, namely the models of Harkins and Jura, Hasley, Carbon Black and Jaroniec, were compared in order to determine the textural properties (external surface and surface of micropores) of a clay concrete without molasses and clay concretes stabilized with 8%, 12% and 16% molasses. The results obtained show that Hasley’s model can be used to obtain the external surfaces. However, it does not allow the surface of the micropores to be obtained, and is not suitable for the case of simple clay concrete (without molasses) and for clay concretes stabilized with molasses. The Carbon Black, Jaroniec and Harkins and Jura models can be used for clay concrete and stabilized clay concrete. However, the Carbon Black model is the most relevant for clay concrete and the Harkins and Jura model is for molasses-stabilized clay concrete. These last two models augur well for future research.展开更多
Cardiovascular Diseases (CVDs) pose a significant global health challenge, necessitating accurate risk prediction for effective preventive measures. This comprehensive comparative study explores the performance of tra...Cardiovascular Diseases (CVDs) pose a significant global health challenge, necessitating accurate risk prediction for effective preventive measures. This comprehensive comparative study explores the performance of traditional Machine Learning (ML) and Deep Learning (DL) models in predicting CVD risk, utilizing a meticulously curated dataset derived from health records. Rigorous preprocessing, including normalization and outlier removal, enhances model robustness. Diverse ML models (Logistic Regression, Random Forest, Support Vector Machine, K-Nearest Neighbor, Decision Tree, and Gradient Boosting) are compared with a Long Short-Term Memory (LSTM) neural network for DL. Evaluation metrics include accuracy, ROC AUC, computation time, and memory usage. Results identify the Gradient Boosting Classifier and LSTM as top performers, demonstrating high accuracy and ROC AUC scores. Comparative analyses highlight model strengths and limitations, contributing valuable insights for optimizing predictive strategies. This study advances predictive analytics for cardiovascular health, with implications for personalized medicine. The findings underscore the versatility of intelligent systems in addressing health challenges, emphasizing the broader applications of ML and DL in disease identification beyond cardiovascular health.展开更多
We present an approach to classify medical text at a sentence level automatically.Given the inherent complexity of medical text classification,we employ adapters based on pre-trained language models to extract informa...We present an approach to classify medical text at a sentence level automatically.Given the inherent complexity of medical text classification,we employ adapters based on pre-trained language models to extract information from medical text,facilitating more accurate classification while minimizing the number of trainable parameters.Extensive experiments conducted on various datasets demonstrate the effectiveness of our approach.展开更多
Patient-derived tumor xenograft(PDX)models,a method involving the surgical extraction of tumor tissues from cancer patients and subsequent transplantation into immunodeficient mice,have emerged as a pivotal approach i...Patient-derived tumor xenograft(PDX)models,a method involving the surgical extraction of tumor tissues from cancer patients and subsequent transplantation into immunodeficient mice,have emerged as a pivotal approach in translational research,particularly in advancing precision medicine.As the first stage of PDX development,the patient-derived orthotopic xenograft(PDOX)models implant tumor tissue in mice in the corresponding anatomical locations of the patient.The PDOX models have several advantages,including high fidelity to the original tumor,heightened drug sensitivity,and an elevated rate of successful transplantation.However,the PDOX models present significant challenges,requiring advanced surgical techniques and resourceintensive imaging technologies,which limit its application.And then,the humanized mouse models,as well as the zebrafish models,were developed.Humanized mouse models contain a human immune environment resembling the tumor and immune system interplay.The humanized mouse models are a hot topic in PDX model research.Regarding zebrafish patient-derived tumor xenografts(zPDX)and patient-derived organoids(PDO)as promising models for studying cancer and drug discovery,zPDX models are used to transplant tumors into zebrafish as novel personalized medical animal models with the advantage of reducing patient waiting time.PDO models provide a cost-effective approach for drug testing that replicates the in vivo environment and preserves important tumor-related information for patients.The present review highlights the functional characteristics of each new phase of PDX and provides insights into the challenges and prospective developments in this rapidly evolving field.展开更多
The geometry of a landslide dam plays a critical role in its stability and failure mode,and is influenced by the damming process.However,there is a lack of understanding of the factors that affect the 3D geometry of a...The geometry of a landslide dam plays a critical role in its stability and failure mode,and is influenced by the damming process.However,there is a lack of understanding of the factors that affect the 3D geometry of a landslide dam.To address this gap,we conducted a study using the smoothed particle hydrodynamics numerical method to investigate the evolution of landslide dams.Our study included 17 numerical simulations to examine the effects of several factors on the geometry of landslide dams,including valley inclination,sliding angle,landslide velocity,and landslide mass repose angle.Based on this,three rapid prediction models were established for calculating the maximum height,the minimum height,and the maximum width of a landslide dam.The results show that the downstream width of a landslide dam remarkably increases with the valley inclination.The position of the maximum dam height along the valley direction is independent of external factors and is always located in the middle of the landslide width area.In contrast,that position of the maximum dam height across the valley direction is significantly influenced by the sliding angle and landslide velocity.To validate our models,we applied them to three typical landslide dams and found that the calculated values of the landslide dam geometry were in good agreement with the actual values.The findings of the current study provide a better understanding of the evolution and geometry of landslide dams,giving crucial guidance for the prediction and early warning of landslide dam disasters.展开更多
Maize (Zea mays L.) is one of the three major food crops and an important source of carbohydrates for maintaining food security around the world.Plant height (H),stem diameter (SD),leaf area index (LAI) and dry matter...Maize (Zea mays L.) is one of the three major food crops and an important source of carbohydrates for maintaining food security around the world.Plant height (H),stem diameter (SD),leaf area index (LAI) and dry matter (DM) are important growth parameters that influence maize production.However,the combined effect of temperature and light on maize growth is rarely considered in crop growth models.Ten maize growth models based on the modified logistic growth equation (Mlog) and the Mitscherlich growth equation (Mit) were proposed to simulate the H,SD,LAI and DM of maize under different mulching practices based on experimental data from 2015–2018.Either the accumulative growing degree-days (AGDD),helio thermal units (HTU),photothermal units (PTU) or photoperiod thermal units (PPTU,first proposed here) was used as a single driving factor in the models;or AGDD was combined with either accumulative actual solar hours (ASS),accumulative photoperiod response (APR,first proposed here) or accumulative maximum possible sunshine hours (ADL) as the dual driving factors in the models.The model performances were evaluated using seven statistical indicators and a global performance index.The results showed that the three mulching practices significantly increased the maize growth rates and the maximum values of the growth curves compared with non-mulching.Among the four single factor-driven models,the overall performance of the Mlog_(PTU)Model was the best,followed by the Mlog_(AGDD)Model.The Mlog_(PPTU)Model was better than the Mlog_(AGDD)Model in simulating SD and LAI.Among the 10 models,the overall performance of the Mlog_(AGDD–APR)Model was the best,followed by the Mlog_(AGDD–ASS)Model.Specifically,the Mlog_(AGDD–APR)Model performed the best in simulating H and LAI,while the Mlog_(AGDD–ADL)and Mlog_(AGDD–ASS)models performed the best in simulating SD and DM,respectively.In conclusion,the modified logistic growth equations with AGDD and either APR,ASS or ADL as the dual driving factors outperformed the commonly used modified logistic growth model with AGDD as a single driving factor in simulating maize growth.展开更多
Fully connected neural networks(FCNNs)have been developed for the closure of subgrid-scale(SGS)stress and SGS heat flux in large-eddy simulations of compressible turbulent channel flow.The FCNNbased SGS model trained ...Fully connected neural networks(FCNNs)have been developed for the closure of subgrid-scale(SGS)stress and SGS heat flux in large-eddy simulations of compressible turbulent channel flow.The FCNNbased SGS model trained using data with Mach number Ma=3.0 and Reynolds number Re=3000 was applied to situations with different Mach numbers and Reynolds numbers.The input variables of the neural network model were the filtered velocity gradients and temperature gradients at a single spatial grid point.The a priori test showed that the FCNN model had a correlation coefficient larger than 0.91 and a relative error smaller than 0.43,with much better reconstructions of SGS unclosed terms than the dynamic Smagorinsky model(DSM).In a posteriori test,the behavior of the FCNN model was marginally better than that of the DSM in predicting the mean velocity profiles,mean temperature profiles,turbulent intensities,total Reynolds stress,total Reynolds heat flux,and mean SGS flux of kinetic energy,and outperformed the Smagorinsky model.展开更多
The proliferation of maliciously coded documents as file transfers increase has led to a rise in sophisticated attacks.Portable Document Format(PDF)files have emerged as a major attack vector for malware due to their ...The proliferation of maliciously coded documents as file transfers increase has led to a rise in sophisticated attacks.Portable Document Format(PDF)files have emerged as a major attack vector for malware due to their adaptability and wide usage.Detecting malware in PDF files is challenging due to its ability to include various harmful elements such as embedded scripts,exploits,and malicious URLs.This paper presents a comparative analysis of machine learning(ML)techniques,including Naive Bayes(NB),K-Nearest Neighbor(KNN),Average One Dependency Estimator(A1DE),RandomForest(RF),and SupportVectorMachine(SVM)forPDFmalware detection.The study utilizes a dataset obtained from the Canadian Institute for Cyber-security and employs different testing criteria,namely percentage splitting and 10-fold cross-validation.The performance of the techniques is evaluated using F1-score,precision,recall,and accuracy measures.The results indicate that KNNoutperforms other models,achieving an accuracy of 99.8599%using 10-fold cross-validation.The findings highlight the effectiveness of ML models in accurately detecting PDF malware and provide insights for developing robust systems to protect against malicious activities.展开更多
基金supported by the National Key Research and Development Program of China (2021YFA0805300,2021YFA0805200)National Natural Science Foundation of China (32170981,82371874,82394422,82171244,82071421,82271902)+1 种基金Guangzhou Key Research Program on Brain Science (202007030008)Department of Science and Technology of Guangdong Province (2021ZT09Y007,2020B121201006,2018B030337001)。
文摘Huntington'sdisease(HD)isahereditary neurodegenerative disorder for which there is currently no effectivetreatmentavailable.Consequently,the development of appropriate disease models is critical to thoroughly investigate disease progression.The genetic basis of HD involves the abnormal expansion of CAG repeats in the huntingtin(HTT)gene,leading to the expansion of a polyglutamine repeat in the HTT protein.Mutant HTT carrying the expanded polyglutamine repeat undergoes misfolding and forms aggregates in the brain,which precipitate selective neuronal loss in specific brain regions.Animal models play an important role in elucidating the pathogenesis of neurodegenerative disorders such as HD and in identifying potential therapeutic targets.Due to the marked species differences between rodents and larger animals,substantial efforts have been directed toward establishing large animal models for HD research.These models are pivotal for advancing the discovery of novel therapeutic targets,enhancing effective drug delivery methods,and improving treatment outcomes.We have explored the advantages of utilizing large animal models,particularly pigs,in previous reviews.Since then,however,significant progress has been made in developing more sophisticated animal models that faithfully replicate the typical pathology of HD.In the current review,we provide a comprehensive overview of large animal models of HD,incorporating recent findings regarding the establishment of HD knock-in(KI)pigs and their genetic therapy.We also explore the utilization of large animal models in HD research,with a focus on sheep,non-human primates(NHPs),and pigs.Our objective is to provide valuable insights into the application of these large animal models for the investigation and treatment of neurodegenerative disorders.
基金supported by the National Key Research and Development Program of China (2021YFA0805902,2022YFF0710703)National Natural Science Foundation of China (32201257)+1 种基金Science and Technology Innovation Project of Xiongan New Area (2022XAGG0121)Young Elite Scientists Sponsorship Program by the China Association for Science and Technology (2019QNRC001)。
文摘Hereditary hearing loss(HHL),a genetic disorder that impairs auditory function,significantly affects quality of life and incurs substantial economic losses for society.To investigate the underlying causes of HHL and evaluate therapeutic outcomes,appropriate animal models are necessary.Pigs have been extensively used as valuable large animal models in biomedical research.In this review,we highlight the advantages of pig models in terms of ear anatomy,inner ear morphology,and electrophysiological characteristics,as well as recent advancements in the development of distinct genetically modified porcine models of hearing loss.Additionally,we discuss the prospects,challenges,and recommendations regarding the use pig models in HHL research.Overall,this review provides insights and perspectives for future studies on HHL using porcine models.
基金supported by the National Key Research and Development Program of China (2021YFF0702201)National Natural Science Foundation of China (81873736,31872779,81830032)+2 种基金Guangzhou Key Research Program on Brain Science (202007030008)Department of Science and Technology of Guangdong Province (2021ZT09Y007,2020B121201006,2018B030337001,2021A1515012526)Natural Science Foundation of Guangdong Province (2021A1515012526,2022A1515012651)。
文摘Neurodegenerative diseases(NDs)are a group of debilitating neurological disorders that primarily affect elderly populations and include Alzheimer's disease(AD),Parkinson's disease(PD),Huntington's disease(HD),and amyotrophic lateral sclerosis(ALS).Currently,there are no therapies available that can delay,stop,or reverse the pathological progression of NDs in clinical settings.As the population ages,NDs are imposing a huge burden on public health systems and affected families.Animal models are important tools for preclinical investigations to understand disease pathogenesis and test potential treatments.While numerous rodent models of NDs have been developed to enhance our understanding of disease mechanisms,the limited success of translating findings from animal models to clinical practice suggests that there is still a need to bridge this translation gap.Old World nonhuman primates(NHPs),such as rhesus,cynomolgus,and vervet monkeys,are phylogenetically,physiologically,biochemically,and behaviorally most relevant to humans.This is particularly evident in the similarity of the structure and function of their central nervous systems,rendering such species uniquely valuable for neuroscience research.Recently,the development of several genetically modified NHP models of NDs has successfully recapitulated key pathologies and revealed novel mechanisms.This review focuses on the efficacy of NHPs in modeling NDs and the novel pathological insights gained,as well as the challenges associated with the generation of such models and the complexities involved in their subsequent analysis.
文摘BACKGROUND Liver transplantation(LT)is a life-saving intervention for patients with end-stage liver disease.However,the equitable allocation of scarce donor organs remains a formidable challenge.Prognostic tools are pivotal in identifying the most suitable transplant candidates.Traditionally,scoring systems like the model for end-stage liver disease have been instrumental in this process.Nevertheless,the landscape of prognostication is undergoing a transformation with the integration of machine learning(ML)and artificial intelligence models.AIM To assess the utility of ML models in prognostication for LT,comparing their performance and reliability to established traditional scoring systems.METHODS Following the Preferred Reporting Items for Systematic Reviews and Meta-Analysis guidelines,we conducted a thorough and standardized literature search using the PubMed/MEDLINE database.Our search imposed no restrictions on publication year,age,or gender.Exclusion criteria encompassed non-English studies,review articles,case reports,conference papers,studies with missing data,or those exhibiting evident methodological flaws.RESULTS Our search yielded a total of 64 articles,with 23 meeting the inclusion criteria.Among the selected studies,60.8%originated from the United States and China combined.Only one pediatric study met the criteria.Notably,91%of the studies were published within the past five years.ML models consistently demonstrated satisfactory to excellent area under the receiver operating characteristic curve values(ranging from 0.6 to 1)across all studies,surpassing the performance of traditional scoring systems.Random forest exhibited superior predictive capabilities for 90-d mortality following LT,sepsis,and acute kidney injury(AKI).In contrast,gradient boosting excelled in predicting the risk of graft-versus-host disease,pneumonia,and AKI.CONCLUSION This study underscores the potential of ML models in guiding decisions related to allograft allocation and LT,marking a significant evolution in the field of prognostication.
文摘The Indian Himalayan region is frequently experiencing climate change-induced landslides.Thus,landslide susceptibility assessment assumes greater significance for lessening the impact of a landslide hazard.This paper makes an attempt to assess landslide susceptibility in Shimla district of the northwest Indian Himalayan region.It examined the effectiveness of random forest(RF),multilayer perceptron(MLP),sequential minimal optimization regression(SMOreg)and bagging ensemble(B-RF,BSMOreg,B-MLP)models.A landslide inventory map comprising 1052 locations of past landslide occurrences was classified into training(70%)and testing(30%)datasets.The site-specific influencing factors were selected by employing a multicollinearity test.The relationship between past landslide occurrences and influencing factors was established using the frequency ratio method.The effectiveness of machine learning models was verified through performance assessors.The landslide susceptibility maps were validated by the area under the receiver operating characteristic curves(ROC-AUC),accuracy,precision,recall and F1-score.The key performance metrics and map validation demonstrated that the BRF model(correlation coefficient:0.988,mean absolute error:0.010,root mean square error:0.058,relative absolute error:2.964,ROC-AUC:0.947,accuracy:0.778,precision:0.819,recall:0.917 and F-1 score:0.865)outperformed the single classifiers and other bagging ensemble models for landslide susceptibility.The results show that the largest area was found under the very high susceptibility zone(33.87%),followed by the low(27.30%),high(20.68%)and moderate(18.16%)susceptibility zones.The factors,namely average annual rainfall,slope,lithology,soil texture and earthquake magnitude have been identified as the influencing factors for very high landslide susceptibility.Soil texture,lineament density and elevation have been attributed to high and moderate susceptibility.Thus,the study calls for devising suitable landslide mitigation measures in the study area.Structural measures,an immediate response system,community participation and coordination among stakeholders may help lessen the detrimental impact of landslides.The findings from this study could aid decision-makers in mitigating future catastrophes and devising suitable strategies in other geographical regions with similar geological characteristics.
基金supported by the National Natural Science Foundation of China(NFSCGrant No.42030410)+2 种基金Laoshan Laboratory(No.LSKJ202202402)the Strategic Priority Research Program of the Chinese Academy of Sciences(Grant No.XDB40000000)the Startup Foundation for Introducing Talent of NUIST.
文摘El Niño-Southern Oscillation(ENSO)is the strongest interannual climate mode influencing the coupled ocean-atmosphere system in the tropical Pacific,and numerous dynamical and statistical models have been developed to simulate and predict it.In some simplified coupled ocean-atmosphere models,the relationship between sea surface temperature(SST)anomalies and wind stress(τ)anomalies can be constructed by statistical methods,such as singular value decomposition(SVD).In recent years,the applications of artificial intelligence(AI)to climate modeling have shown promising prospects,and the integrations of AI-based models with dynamical models are active areas of research.This study constructs U-Net models for representing the relationship between SSTAs andτanomalies in the tropical Pacific;the UNet-derivedτmodel,denoted asτUNet,is then used to replace the original SVD-basedτmodel of an intermediate coupled model(ICM),forming a newly AI-integrated ICM,referred to as ICM-UNet.The simulation results obtained from ICM-UNet demonstrate their ability to represent the spatiotemporal variability of oceanic and atmospheric anomaly fields in the equatorial Pacific.In the ocean-only case study,theτUNet-derived wind stress anomaly fields are used to force the ocean component of the ICM,the results of which also indicate reasonable simulations of typical ENSO events.These results demonstrate the feasibility of integrating an AI-derived model with a physics-based dynamical model for ENSO modeling studies.Furthermore,the successful integration of the dynamical ocean models with the AI-based atmospheric wind model provides a novel approach to ocean-atmosphere interaction modeling studies.
文摘Postoperative pancreatic fistula(POPF)is a frequent complication after pancre-atectomy,leading to increased morbidity and mortality.Optimizing prediction models for POPF has emerged as a critical focus in surgical research.Although over sixty models following pancreaticoduodenectomy,predominantly reliant on a variety of clinical,surgical,and radiological parameters,have been documented,their predictive accuracy remains suboptimal in external validation and across diverse populations.As models after distal pancreatectomy continue to be pro-gressively reported,their external validation is eagerly anticipated.Conversely,POPF prediction after central pancreatectomy is in its nascent stage,warranting urgent need for further development and validation.The potential of machine learning and big data analytics offers promising prospects for enhancing the accuracy of prediction models by incorporating an extensive array of variables and optimizing algorithm performance.Moreover,there is potential for the development of personalized prediction models based on patient-or pancreas-specific factors and postoperative serum or drain fluid biomarkers to improve accuracy in identifying individuals at risk of POPF.In the future,prospective multicenter studies and the integration of novel imaging technologies,such as artificial intelligence-based radiomics,may further refine predictive models.Addressing these issues is anticipated to revolutionize risk stratification,clinical decision-making,and postoperative management in patients undergoing pancre-atectomy.
文摘The problematic use of social media has numerous negative impacts on individuals'daily lives,interpersonal relationships,physical and mental health,and more.Currently,there are few methods and tools to alleviate problematic social media,and their potential is yet to be fully realized.Emerging large language models(LLMs)are becoming increasingly popular for providing information and assistance to people and are being applied in many aspects of life.In mitigating problematic social media use,LLMs such as ChatGPT can play a positive role by serving as conversational partners and outlets for users,providing personalized information and resources,monitoring and intervening in problematic social media use,and more.In this process,we should recognize both the enormous potential and endless possibilities of LLMs such as ChatGPT,leveraging their advantages to better address problematic social media use,while also acknowledging the limitations and potential pitfalls of ChatGPT technology,such as errors,limitations in issue resolution,privacy and security concerns,and potential overreliance.When we leverage the advantages of LLMs to address issues in social media usage,we must adopt a cautious and ethical approach,being vigilant of the potential adverse effects that LLMs may have in addressing problematic social media use to better harness technology to serve individuals and society.
文摘BACKGROUND Gastric cancer is one of the most common malignant tumors in the digestive system,ranking sixth in incidence and fourth in mortality worldwide.Since 42.5%of metastatic lymph nodes in gastric cancer belong to nodule type and peripheral type,the application of imaging diagnosis is restricted.AIM To establish models for predicting the risk of lymph node metastasis in gastric cancer patients using machine learning(ML)algorithms and to evaluate their pre-dictive performance in clinical practice.METHODS Data of a total of 369 patients who underwent radical gastrectomy at the Depart-ment of General Surgery of Affiliated Hospital of Xuzhou Medical University(Xuzhou,China)from March 2016 to November 2019 were collected and retro-spectively analyzed as the training group.In addition,data of 123 patients who underwent radical gastrectomy at the Department of General Surgery of Jining First People’s Hospital(Jining,China)were collected and analyzed as the verifi-cation group.Seven ML models,including decision tree,random forest,support vector machine(SVM),gradient boosting machine,naive Bayes,neural network,and logistic regression,were developed to evaluate the occurrence of lymph node metastasis in patients with gastric cancer.The ML models were established fo-llowing ten cross-validation iterations using the training dataset,and subsequently,each model was assessed using the test dataset.The models’performance was evaluated by comparing the area under the receiver operating characteristic curve of each model.RESULTS Among the seven ML models,except for SVM,the other ones exhibited higher accuracy and reliability,and the influences of various risk factors on the models are intuitive.CONCLUSION The ML models developed exhibit strong predictive capabilities for lymph node metastasis in gastric cancer,which can aid in personalized clinical diagnosis and treatment.
文摘Drone or unmanned aerial vehicle(UAV)technology has undergone significant changes.The technology allows UAV to carry out a wide range of tasks with an increasing level of sophistication,since drones can cover a large area with cameras.Meanwhile,the increasing number of computer vision applications utilizing deep learning provides a unique insight into such applications.The primary target in UAV-based detection applications is humans,yet aerial recordings are not included in the massive datasets used to train object detectors,which makes it necessary to gather the model data from such platforms.You only look once(YOLO)version 4,RetinaNet,faster region-based convolutional neural network(R-CNN),and cascade R-CNN are several well-known detectors that have been studied in the past using a variety of datasets to replicate rescue scenes.Here,we used the search and rescue(SAR)dataset to train the you only look once version 5(YOLOv5)algorithm to validate its speed,accuracy,and low false detection rate.In comparison to YOLOv4 and R-CNN,the highest mean average accuracy of 96.9%is obtained by YOLOv5.For comparison,experimental findings utilizing the SAR and the human rescue imaging database on land(HERIDAL)datasets are presented.The results show that the YOLOv5-based approach is the most successful human detection model for SAR missions.
文摘Sunshine duration (S) based empirical equations have been employed in this study to estimate the daily global solar radiation on a horizontal surface (G) for six meteorological stations in Burundi. Those equations include the Ångström-Prescott linear model and four amongst its derivatives, i.e. logarithmic, exponential, power and quadratic functions. Monthly mean values of daily global solar radiation and sunshine duration data for a period of 20 to 23 years, from the Geographical Institute of Burundi (IGEBU), have been used. For any of the six stations, ten single or double linear regressions have been developed from the above-said five functions, to relate in terms of monthly mean values, the daily clearness index () to each of the next two kinds of relative sunshine duration (RSD): and . In those ratios, G<sub>0</sub>, S<sub>0 </sub>and stand for the extraterrestrial daily solar radiation on a horizontal surface, the day length and the modified day length taking into account the natural site’s horizon, respectively. According to the calculated mean values of the clearness index and the RSD, each station experiences a high number of fairly clear (or partially cloudy) days. Estimated values of the dependent variable (y) in each developed linear regression, have been compared to measured values in terms of the coefficients of correlation (R) and of determination (R<sub>2</sub>), the mean bias error (MBE), the root mean square error (RMSE) and the t-statistics. Mean values of these statistical indicators have been used to rank, according to decreasing performance level, firstly the ten developed equations per station on account of the overall six stations, secondly the six stations on account of the overall ten equations. Nevertheless, the obtained values of those indicators lay in the next ranges for all the developed sixty equations:;;;, with . These results lead to assert that any of the sixty developed linear regressions (and thus equations in terms of and ), fits very adequately measured data, and should be used to estimate monthly average daily global solar radiation with sunshine duration for the relevant station. It is also found that using as RSD, is slightly more advantageous than using for estimating the monthly average daily clearness index, . Moreover, values of statistical indicators of this study match adequately data from other works on the same kinds of empirical equations.
文摘Tree height (H) in a natural stand or forest plantation is a fundamental variable in management, and the use of mathematical expressions that estimate H as a function of diameter at breast height (d) or variables at the stand level is a valuable support tool in forest inventories. The objective was to fit and propose a generalized H-d model for Pinus montezumae and Pinus pseudostrobus established in forest plantations of Nuevo San Juan Parangaricutiro, Michoacan, Mexico. Using nonlinear least squares (NLS), 10 generalized H-d models were fitted to 883 and 1226 pairs of H-d data from Pinus montezumae and Pinus pseudostrobus, respectively. The best model was refitted with the maximum likelihood mixed effects model (MEM) approach by including the site as a classification variable and a known variance structure. The Wang and Tang equation was selected as the best model with NLS;the MEM with an additive effect on two of its parameters and an exponential variance function improved the fit statistics for Pinus montezumae and Pinus pseudostrobus, respectively. The model validation showed equality of means among the estimates for both species and an independent subsample. The calibration of the MEM at the plot level was efficient and might increase the applicability of these results. The inclusion of dominant height in the MEM approach helped to reduce bias in the estimates and also to better explain the variability among plots.
文摘In this work, four empirical models of statistical thickness, namely the models of Harkins and Jura, Hasley, Carbon Black and Jaroniec, were compared in order to determine the textural properties (external surface and surface of micropores) of a clay concrete without molasses and clay concretes stabilized with 8%, 12% and 16% molasses. The results obtained show that Hasley’s model can be used to obtain the external surfaces. However, it does not allow the surface of the micropores to be obtained, and is not suitable for the case of simple clay concrete (without molasses) and for clay concretes stabilized with molasses. The Carbon Black, Jaroniec and Harkins and Jura models can be used for clay concrete and stabilized clay concrete. However, the Carbon Black model is the most relevant for clay concrete and the Harkins and Jura model is for molasses-stabilized clay concrete. These last two models augur well for future research.
文摘Cardiovascular Diseases (CVDs) pose a significant global health challenge, necessitating accurate risk prediction for effective preventive measures. This comprehensive comparative study explores the performance of traditional Machine Learning (ML) and Deep Learning (DL) models in predicting CVD risk, utilizing a meticulously curated dataset derived from health records. Rigorous preprocessing, including normalization and outlier removal, enhances model robustness. Diverse ML models (Logistic Regression, Random Forest, Support Vector Machine, K-Nearest Neighbor, Decision Tree, and Gradient Boosting) are compared with a Long Short-Term Memory (LSTM) neural network for DL. Evaluation metrics include accuracy, ROC AUC, computation time, and memory usage. Results identify the Gradient Boosting Classifier and LSTM as top performers, demonstrating high accuracy and ROC AUC scores. Comparative analyses highlight model strengths and limitations, contributing valuable insights for optimizing predictive strategies. This study advances predictive analytics for cardiovascular health, with implications for personalized medicine. The findings underscore the versatility of intelligent systems in addressing health challenges, emphasizing the broader applications of ML and DL in disease identification beyond cardiovascular health.
文摘We present an approach to classify medical text at a sentence level automatically.Given the inherent complexity of medical text classification,we employ adapters based on pre-trained language models to extract information from medical text,facilitating more accurate classification while minimizing the number of trainable parameters.Extensive experiments conducted on various datasets demonstrate the effectiveness of our approach.
基金The Science,Technology and Innovation Commission of Shenzhen,Grant/Award Number:JCYJ20220531093213030。
文摘Patient-derived tumor xenograft(PDX)models,a method involving the surgical extraction of tumor tissues from cancer patients and subsequent transplantation into immunodeficient mice,have emerged as a pivotal approach in translational research,particularly in advancing precision medicine.As the first stage of PDX development,the patient-derived orthotopic xenograft(PDOX)models implant tumor tissue in mice in the corresponding anatomical locations of the patient.The PDOX models have several advantages,including high fidelity to the original tumor,heightened drug sensitivity,and an elevated rate of successful transplantation.However,the PDOX models present significant challenges,requiring advanced surgical techniques and resourceintensive imaging technologies,which limit its application.And then,the humanized mouse models,as well as the zebrafish models,were developed.Humanized mouse models contain a human immune environment resembling the tumor and immune system interplay.The humanized mouse models are a hot topic in PDX model research.Regarding zebrafish patient-derived tumor xenografts(zPDX)and patient-derived organoids(PDO)as promising models for studying cancer and drug discovery,zPDX models are used to transplant tumors into zebrafish as novel personalized medical animal models with the advantage of reducing patient waiting time.PDO models provide a cost-effective approach for drug testing that replicates the in vivo environment and preserves important tumor-related information for patients.The present review highlights the functional characteristics of each new phase of PDX and provides insights into the challenges and prospective developments in this rapidly evolving field.
基金funding from the National Natural Science Foundation of China(42207228,51879036,51579032)the Liaoning Revitalization Talents Program(XLYC2002036)the Sichuan Science and Technology Program(2022NSFSC1060)。
文摘The geometry of a landslide dam plays a critical role in its stability and failure mode,and is influenced by the damming process.However,there is a lack of understanding of the factors that affect the 3D geometry of a landslide dam.To address this gap,we conducted a study using the smoothed particle hydrodynamics numerical method to investigate the evolution of landslide dams.Our study included 17 numerical simulations to examine the effects of several factors on the geometry of landslide dams,including valley inclination,sliding angle,landslide velocity,and landslide mass repose angle.Based on this,three rapid prediction models were established for calculating the maximum height,the minimum height,and the maximum width of a landslide dam.The results show that the downstream width of a landslide dam remarkably increases with the valley inclination.The position of the maximum dam height along the valley direction is independent of external factors and is always located in the middle of the landslide width area.In contrast,that position of the maximum dam height across the valley direction is significantly influenced by the sliding angle and landslide velocity.To validate our models,we applied them to three typical landslide dams and found that the calculated values of the landslide dam geometry were in good agreement with the actual values.The findings of the current study provide a better understanding of the evolution and geometry of landslide dams,giving crucial guidance for the prediction and early warning of landslide dam disasters.
基金funded by the National Natural Science Foundation of China (51879226)the Chinese Universities Scientific Fund (2452020018)。
文摘Maize (Zea mays L.) is one of the three major food crops and an important source of carbohydrates for maintaining food security around the world.Plant height (H),stem diameter (SD),leaf area index (LAI) and dry matter (DM) are important growth parameters that influence maize production.However,the combined effect of temperature and light on maize growth is rarely considered in crop growth models.Ten maize growth models based on the modified logistic growth equation (Mlog) and the Mitscherlich growth equation (Mit) were proposed to simulate the H,SD,LAI and DM of maize under different mulching practices based on experimental data from 2015–2018.Either the accumulative growing degree-days (AGDD),helio thermal units (HTU),photothermal units (PTU) or photoperiod thermal units (PPTU,first proposed here) was used as a single driving factor in the models;or AGDD was combined with either accumulative actual solar hours (ASS),accumulative photoperiod response (APR,first proposed here) or accumulative maximum possible sunshine hours (ADL) as the dual driving factors in the models.The model performances were evaluated using seven statistical indicators and a global performance index.The results showed that the three mulching practices significantly increased the maize growth rates and the maximum values of the growth curves compared with non-mulching.Among the four single factor-driven models,the overall performance of the Mlog_(PTU)Model was the best,followed by the Mlog_(AGDD)Model.The Mlog_(PPTU)Model was better than the Mlog_(AGDD)Model in simulating SD and LAI.Among the 10 models,the overall performance of the Mlog_(AGDD–APR)Model was the best,followed by the Mlog_(AGDD–ASS)Model.Specifically,the Mlog_(AGDD–APR)Model performed the best in simulating H and LAI,while the Mlog_(AGDD–ADL)and Mlog_(AGDD–ASS)models performed the best in simulating SD and DM,respectively.In conclusion,the modified logistic growth equations with AGDD and either APR,ASS or ADL as the dual driving factors outperformed the commonly used modified logistic growth model with AGDD as a single driving factor in simulating maize growth.
基金Financial support provided by the National Natural Science Foundation of China(Grant Nos.11702042 and 91952104)。
文摘Fully connected neural networks(FCNNs)have been developed for the closure of subgrid-scale(SGS)stress and SGS heat flux in large-eddy simulations of compressible turbulent channel flow.The FCNNbased SGS model trained using data with Mach number Ma=3.0 and Reynolds number Re=3000 was applied to situations with different Mach numbers and Reynolds numbers.The input variables of the neural network model were the filtered velocity gradients and temperature gradients at a single spatial grid point.The a priori test showed that the FCNN model had a correlation coefficient larger than 0.91 and a relative error smaller than 0.43,with much better reconstructions of SGS unclosed terms than the dynamic Smagorinsky model(DSM).In a posteriori test,the behavior of the FCNN model was marginally better than that of the DSM in predicting the mean velocity profiles,mean temperature profiles,turbulent intensities,total Reynolds stress,total Reynolds heat flux,and mean SGS flux of kinetic energy,and outperformed the Smagorinsky model.
文摘The proliferation of maliciously coded documents as file transfers increase has led to a rise in sophisticated attacks.Portable Document Format(PDF)files have emerged as a major attack vector for malware due to their adaptability and wide usage.Detecting malware in PDF files is challenging due to its ability to include various harmful elements such as embedded scripts,exploits,and malicious URLs.This paper presents a comparative analysis of machine learning(ML)techniques,including Naive Bayes(NB),K-Nearest Neighbor(KNN),Average One Dependency Estimator(A1DE),RandomForest(RF),and SupportVectorMachine(SVM)forPDFmalware detection.The study utilizes a dataset obtained from the Canadian Institute for Cyber-security and employs different testing criteria,namely percentage splitting and 10-fold cross-validation.The performance of the techniques is evaluated using F1-score,precision,recall,and accuracy measures.The results indicate that KNNoutperforms other models,achieving an accuracy of 99.8599%using 10-fold cross-validation.The findings highlight the effectiveness of ML models in accurately detecting PDF malware and provide insights for developing robust systems to protect against malicious activities.