Through combined applications of the transfer-matrix method and asymptotic expansion technique,we formulate a theory to predict the three-dimensional response of micropolar plates.No ad hoc assumptions regarding throu...Through combined applications of the transfer-matrix method and asymptotic expansion technique,we formulate a theory to predict the three-dimensional response of micropolar plates.No ad hoc assumptions regarding through-thickness assumptions of the field variables are made,and the governing equations are two-dimensional,with the displacements and microrotations of the mid-plane as the unknowns.Once the deformation of the mid-plane is solved,a three-dimensional micropolar elastic field within the plate is generated,which is exact up to the second order except in the boundary region close to the plate edge.As an illustrative example,the bending of a clamped infinitely long plate caused by a uniformly distributed transverse force is analyzed and discussed in detail.展开更多
Hypoxia is a typical feature of the tumor microenvironment,one of the most critical factors affecting cell behavior and tumor progression.However,the lack of tumor models able to precisely emulate natural brain tumor ...Hypoxia is a typical feature of the tumor microenvironment,one of the most critical factors affecting cell behavior and tumor progression.However,the lack of tumor models able to precisely emulate natural brain tumor tissue has impeded the study of the effects of hypoxia on the progression and growth of tumor cells.This study reports a three-dimensional(3D)brain tumor model obtained by encapsulating U87MG(U87)cells in a hydrogel containing type I collagen.It also documents the effect of various oxygen concentrations(1%,7%,and 21%)in the culture environment on U87 cell morphology,proliferation,viability,cell cycle,apoptosis rate,and migration.Finally,it compares two-dimensional(2D)and 3D cultures.For comparison purposes,cells cultured in flat culture dishes were used as the control(2D model).Cells cultured in the 3D model proliferated more slowly but had a higher apoptosis rate and proportion of cells in the resting phase(G0 phase)/gap I phase(G1 phase)than those cultured in the 2D model.Besides,the two models yielded significantly different cell morphologies.Finally,hypoxia(e.g.,1%O2)affected cell morphology,slowed cell growth,reduced cell viability,and increased the apoptosis rate in the 3D model.These results indicate that the constructed 3D model is effective for investigating the effects of biological and chemical factors on cell morphology and function,and can be more representative of the tumor microenvironment than 2D culture systems.The developed 3D glioblastoma tumor model is equally applicable to other studies in pharmacology and pathology.展开更多
Liver regeneration and the development of effective therapies for liver failure remain formidable challenges in modern medicine.In recent years,the utilization of 3D cell-based strategies has emerged as a promising ap...Liver regeneration and the development of effective therapies for liver failure remain formidable challenges in modern medicine.In recent years,the utilization of 3D cell-based strategies has emerged as a promising approach for addressing these urgent clinical requirements.This review provides a thorough analysis of the application of 3D cell-based approaches to liver regeneration and their potential impact on patients with end-stage liver failure.Here,we discuss various 3D culture models that incorporate hepatocytes and stem cells to restore liver function and ameliorate the consequences of liver failure.Furthermore,we explored the challenges in transitioning these innovative strategies from preclinical studies to clinical applications.The collective insights presented herein highlight the significance of 3D cell-based strategies as a transformative paradigm for liver regeneration and improved patient care.展开更多
The purpose of software defect prediction is to identify defect-prone code modules to assist software quality assurance teams with the appropriate allocation of resources and labor.In previous software defect predicti...The purpose of software defect prediction is to identify defect-prone code modules to assist software quality assurance teams with the appropriate allocation of resources and labor.In previous software defect prediction studies,transfer learning was effective in solving the problem of inconsistent project data distribution.However,target projects often lack sufficient data,which affects the performance of the transfer learning model.In addition,the presence of uncorrelated features between projects can decrease the prediction accuracy of the transfer learning model.To address these problems,this article propose a software defect prediction method based on stable learning(SDP-SL)that combines code visualization techniques and residual networks.This method first transforms code files into code images using code visualization techniques and then constructs a defect prediction model based on these code images.During the model training process,target project data are not required as prior knowledge.Following the principles of stable learning,this paper dynamically adjusted the weights of source project samples to eliminate dependencies between features,thereby capturing the“invariance mechanism”within the data.This approach explores the genuine relationship between code defect features and labels,thereby enhancing defect prediction performance.To evaluate the performance of SDP-SL,this article conducted comparative experiments on 10 open-source projects in the PROMISE dataset.The experimental results demonstrated that in terms of the F-measure,the proposed SDP-SL method outperformed other within-project defect prediction methods by 2.11%-44.03%.In cross-project defect prediction,the SDP-SL method provided an improvement of 5.89%-25.46% in prediction performance compared to other cross-project defect prediction methods.Therefore,SDP-SL can effectively enhance within-and cross-project defect predictions.展开更多
Spinal cord injury is considered one of the most difficult injuries to repair and has one of the worst prognoses for injuries to the nervous system.Following surgery,the poor regenerative capacity of nerve cells and t...Spinal cord injury is considered one of the most difficult injuries to repair and has one of the worst prognoses for injuries to the nervous system.Following surgery,the poor regenerative capacity of nerve cells and the generation of new scars can make it very difficult for the impaired nervous system to restore its neural functionality.Traditional treatments can only alleviate secondary injuries but cannot fundamentally repair the spinal cord.Consequently,there is a critical need to develop new treatments to promote functional repair after spinal cord injury.Over recent years,there have been seve ral developments in the use of stem cell therapy for the treatment of spinal cord injury.Alongside significant developments in the field of tissue engineering,three-dimensional bioprinting technology has become a hot research topic due to its ability to accurately print complex structures.This led to the loading of three-dimensional bioprinting scaffolds which provided precise cell localization.These three-dimensional bioprinting scaffolds co uld repair damaged neural circuits and had the potential to repair the damaged spinal cord.In this review,we discuss the mechanisms underlying simple stem cell therapy,the application of different types of stem cells for the treatment of spinal cord injury,and the different manufa cturing methods for three-dimensional bioprinting scaffolds.In particular,we focus on the development of three-dimensional bioprinting scaffolds for the treatment of spinal cord injury.展开更多
Software Defined Networking(SDN)is programmable by separation of forwarding control through the centralization of the controller.The controller plays the role of the‘brain’that dictates the intelligent part of SDN t...Software Defined Networking(SDN)is programmable by separation of forwarding control through the centralization of the controller.The controller plays the role of the‘brain’that dictates the intelligent part of SDN technology.Various versions of SDN controllers exist as a response to the diverse demands and functions expected of them.There are several SDN controllers available in the open market besides a large number of commercial controllers;some are developed tomeet carrier-grade service levels and one of the recent trends in open-source SDN controllers is the Open Network Operating System(ONOS).This paper presents a comparative study between open source SDN controllers,which are known as Network Controller Platform(NOX),Python-based Network Controller(POX),component-based SDN framework(Ryu),Java-based OpenFlow controller(Floodlight),OpenDayLight(ODL)and ONOS.The discussion is further extended into ONOS architecture,as well as,the evolution of ONOS controllers.This article will review use cases based on ONOS controllers in several application deployments.Moreover,the opportunities and challenges of open source SDN controllers will be discussed,exploring carriergrade ONOS for future real-world deployments,ONOS unique features and identifying the suitable choice of SDN controller for service providers.In addition,we attempt to provide answers to several critical questions relating to the implications of the open-source nature of SDN controllers regarding vendor lock-in,interoperability,and standards compliance,Similarly,real-world use cases of organizations using open-source SDN are highlighted and how the open-source community contributes to the development of SDN controllers.Furthermore,challenges faced by open-source projects,and considerations when choosing an open-source SDN controller are underscored.Then the role of Artificial Intelligence(AI)and Machine Learning(ML)in the evolution of open-source SDN controllers in light of recent research is indicated.In addition,the challenges and limitations associated with deploying open-source SDN controllers in production networks,how can they be mitigated,and finally how opensource SDN controllers handle network security and ensure that network configurations and policies are robust and resilient are presented.Potential opportunities and challenges for future Open SDN deployment are outlined to conclude the article.展开更多
The Message Passing Interface (MPI) is a widely accepted standard for parallel computing on distributed memorysystems.However, MPI implementations can contain defects that impact the reliability and performance of par...The Message Passing Interface (MPI) is a widely accepted standard for parallel computing on distributed memorysystems.However, MPI implementations can contain defects that impact the reliability and performance of parallelapplications. Detecting and correcting these defects is crucial, yet there is a lack of published models specificallydesigned for correctingMPI defects. To address this, we propose a model for detecting and correcting MPI defects(DC_MPI), which aims to detect and correct defects in various types of MPI communication, including blockingpoint-to-point (BPTP), nonblocking point-to-point (NBPTP), and collective communication (CC). The defectsaddressed by the DC_MPI model include illegal MPI calls, deadlocks (DL), race conditions (RC), and messagemismatches (MM). To assess the effectiveness of the DC_MPI model, we performed experiments on a datasetconsisting of 40 MPI codes. The results indicate that the model achieved a detection rate of 37 out of 40 codes,resulting in an overall detection accuracy of 92.5%. Additionally, the execution duration of the DC_MPI modelranged from 0.81 to 1.36 s. These findings show that the DC_MPI model is useful in detecting and correctingdefects in MPI implementations, thereby enhancing the reliability and performance of parallel applications. TheDC_MPImodel fills an important research gap and provides a valuable tool for improving the quality ofMPI-basedparallel computing systems.展开更多
Software testing is a critical phase due to misconceptions about ambiguities in the requirements during specification,which affect the testing process.Therefore,it is difficult to identify all faults in software.As re...Software testing is a critical phase due to misconceptions about ambiguities in the requirements during specification,which affect the testing process.Therefore,it is difficult to identify all faults in software.As requirement changes continuously,it increases the irrelevancy and redundancy during testing.Due to these challenges;fault detection capability decreases and there arises a need to improve the testing process,which is based on changes in requirements specification.In this research,we have developed a model to resolve testing challenges through requirement prioritization and prediction in an agile-based environment.The research objective is to identify the most relevant and meaningful requirements through semantic analysis for correct change analysis.Then compute the similarity of requirements through case-based reasoning,which predicted the requirements for reuse and restricted to error-based requirements.Afterward,the apriori algorithm mapped out requirement frequency to select relevant test cases based on frequently reused or not reused test cases to increase the fault detection rate.Furthermore,the proposed model was evaluated by conducting experiments.The results showed that requirement redundancy and irrelevancy improved due to semantic analysis,which correctly predicted the requirements,increasing the fault detection rate and resulting in high user satisfaction.The predicted requirements are mapped into test cases,increasing the fault detection rate after changes to achieve higher user satisfaction.Therefore,the model improves the redundancy and irrelevancy of requirements by more than 90%compared to other clustering methods and the analytical hierarchical process,achieving an 80%fault detection rate at an earlier stage.Hence,it provides guidelines for practitioners and researchers in the modern era.In the future,we will provide the working prototype of this model for proof of concept.展开更多
Software Development Life Cycle (SDLC) is one of the major ingredients for the development of efficient software systems within a time frame and low-cost involvement. From the literature, it is evident that there are ...Software Development Life Cycle (SDLC) is one of the major ingredients for the development of efficient software systems within a time frame and low-cost involvement. From the literature, it is evident that there are various kinds of process models that are used by the software industries for the development of small, medium and long-term software projects, but many of them do not cover risk management. It is quite obvious that the improper selection of the software development process model leads to failure of the software products as it is time bound activity. In the present work, a new software development process model is proposed which covers the risks at any stage of the development of the software product. The model is named a Hemant-Vipin (HV) process model and may be helpful for the software industries for development of the efficient software products and timely delivery at the end of the client. The efficiency of the HV process model is observed by considering various kinds of factors like requirement clarity, user feedback, change agility, predictability, risk identification, practical implementation, customer satisfaction, incremental development, use of ready-made components, quick design, resource organization and many more and found through a case study that the presented approach covers many of parameters in comparison of the existing process models. .展开更多
Agile Transformations are challenging processes for organizations that look to extend the benefits of Agile philosophy and methods beyond software engineering.Despite the impact of these transformations on orga-nizati...Agile Transformations are challenging processes for organizations that look to extend the benefits of Agile philosophy and methods beyond software engineering.Despite the impact of these transformations on orga-nizations,they have not been extensively studied in academia.We conducted a study grounded in workshops and interviews with 99 participants from 30 organizations,including organizations undergoing transformations(“final organizations”)and companies supporting these processes(“consultants”).The study aims to understand the motivations,objectives,and factors driving and challenging these transformations.Over 700 responses were collected to the question and categorized into 32 objectives.The findings show that organizations primarily aim to achieve customer centricity and adaptability,both with 8%of the mentions.Other primary important objectives,with above 4%of mentions,include alignment of goals,lean delivery,sustainable processes,and a flatter,more team-based organizational structure.We also detect discrepancies in perspectives between the objectives identified by the two kinds of organizations and the existing agile literature and models.This misalignment highlights the need for practitioners to understand with the practical realities the organizations face.展开更多
Accurate software cost estimation in Global Software Development(GSD)remains challenging due to reliance on historical data and expert judgments.Traditional models,such as the Constructive Cost Model(COCOMO II),rely h...Accurate software cost estimation in Global Software Development(GSD)remains challenging due to reliance on historical data and expert judgments.Traditional models,such as the Constructive Cost Model(COCOMO II),rely heavily on historical and accurate data.In addition,expert judgment is required to set many input parameters,which can introduce subjectivity and variability in the estimation process.Consequently,there is a need to improve the current GSD models to mitigate reliance on historical data,subjectivity in expert judgment,inadequate consideration of GSD-based cost drivers and limited integration of modern technologies with cost overruns.This study introduces a novel hybrid model that synergizes the COCOMO II with Artificial Neural Networks(ANN)to address these challenges.The proposed hybrid model integrates additional GSD-based cost drivers identified through a systematic literature review and further vetted by industry experts.This article compares the effectiveness of the proposedmodelwith state-of-the-artmachine learning-basedmodels for software cost estimation.Evaluating the NASA 93 dataset by adopting twenty-six GSD-based cost drivers reveals that our hybrid model achieves superior accuracy,outperforming existing state-of-the-artmodels.The findings indicate the potential of combining COCOMO II,ANN,and additional GSD-based cost drivers to transform cost estimation in GSD.展开更多
Purpose: To clarify the effectiveness of 3-D delivery animation software for the mother’s and husband’s satisfaction with delivery. Subjects and Method: We independently developed a software application used to disp...Purpose: To clarify the effectiveness of 3-D delivery animation software for the mother’s and husband’s satisfaction with delivery. Subjects and Method: We independently developed a software application used to display the pelvic region and explain the labor process. The study involved a collaboration with hospital staff who recruited 18 primiparous and 18 multiparous mothers who were hospitalized for delivery at Facility A. The midwife explained the process of delivery using the “Delivery Animation Software”. A self-administered, anonymous questionnaire was distributed and analyzed separately for primiparous and multiparous mothers and their husbands. Results: 1) For both primiparous and multiparous couples, both mothers and their husbands gained a significantly higher level of understanding after delivery than during pregnancy. 2) The Self-Evaluation Scale for Experience of Delivery results were as follows: “I did my best for the baby even if it was painful” was selected more often for “birth coping skills”;“reliable medical staff” was selected more often for “physiological birth process”;“the birth progressed as I expected” was selected frequently by primiparous mothers;and “the birth progressed smoothly” was selected often by multiparous mothers. 3) In terms of husbands’ satisfaction with the delivery, “I was satisfied with the delivery”, “I was given an easy-to-understand explanation”, and “They explained the process to me” was selected of primiparous and multiparous fathers. 4) All primiparous and multiparous mothers positively evaluated whether the delivery animation was helpful in understanding the process of delivery. Conclusion: The delivery animation was effective in improving the understanding and satisfaction of both the mothers and their husbands.展开更多
A toroidal soft x-ray imaging(T-SXRI)system has been developed to investigate threedimensional(3D)plasma physics on J-TEXT.This T-SXRI system consists of three sets of SXR arrays.Two sets are newly developed and locat...A toroidal soft x-ray imaging(T-SXRI)system has been developed to investigate threedimensional(3D)plasma physics on J-TEXT.This T-SXRI system consists of three sets of SXR arrays.Two sets are newly developed and located on the vacuum chamber wall at toroidal positionsφof 126.4°and 272.6°,respectively,while one set was established previously atφ=65.50.Each set of SXR arrays consists of three arrays viewing the plasma poloidally,and hence can be used separately to obtain SXR images via the tomographic method.The sawtooth precursor oscillations are measured by T-SXRI,and the corresponding images of perturbative SXR signals are successfully reconstructed at these three toroidal positions,hence providing measurement of the 3D structure of precursor oscillations.The observed 3D structure is consistent with the helical structure of the m/n=1/1 mode.The experimental observation confirms that the T-SXRI system is able to observe 3D structures in the J-TEXT plasma.展开更多
BACKGROUND Acetabular component positioning in total hip arthroplasty(THA)is of key importance to ensure satisfactory post-operative outcomes and to minimize the risk of complications.The majority of acetabular compon...BACKGROUND Acetabular component positioning in total hip arthroplasty(THA)is of key importance to ensure satisfactory post-operative outcomes and to minimize the risk of complications.The majority of acetabular components are aligned freehand,without the use of navigation methods.Patient specific instruments(PSI)and three-dimensional(3D)printing of THA placement guides are increasingly used in primary THA to ensure optimal positioning.AIM To summarize the literature on 3D printing in THA and how they improve acetabular component alignment.METHODS PubMed was used to identify and access scientific studies reporting on different 3D printing methods used in THA.Eight studies with 236 hips in 228 patients were included.The studies could be divided into two main categories;3D printed models and 3D printed guides.RESULTS 3D printing in THA helped improve preoperative cup size planning and post-operative Harris hip scores between intervention and control groups(P=0.019,P=0.009).Otherwise,outcome measures were heterogeneous and thus difficult to compare.The overarching consensus between the studies is that the use of 3D guidance tools can assist in improving THA cup positioning and reduce the need for revision THA and the associated costs.CONCLUSION The implementation of 3D printing and PSI for primary THA can significantly improve the positioning accuracy of the acetabular cup component and reduce the number of complications caused by malpositioning.展开更多
Software delivery is vital for modern organizations, driving innovation and competitiveness. Measuring an organization’s maturity in software delivery is crucial for efficiency and quality. The Capability Maturity Mo...Software delivery is vital for modern organizations, driving innovation and competitiveness. Measuring an organization’s maturity in software delivery is crucial for efficiency and quality. The Capability Maturity Model (CMM) framework provides a roadmap for improvement but assessing an organization’s CMM Level is challenging. This paper offers a quantitative approach tailored to the CMM framework, using Goal-Question-Metric (GQM) frame-works for each key process area (KPA). These frameworks include metrics and questions to compute maturity scores effectively. The study also refines practices into questions for a thorough assessment. The result is an Analysis Matrix that calculates weighted scores and an overall maturity score. This approach helps organizations assess and enhance their software delivery processes systematically, aiming for improved practices and growth.展开更多
BACKGROUND The management of hepatoblastoma(HB)becomes challenging when the tumor remains in close proximity to the major liver vasculature(PMV)even after a full course of neoadjuvant chemotherapy(NAC).In such cases,e...BACKGROUND The management of hepatoblastoma(HB)becomes challenging when the tumor remains in close proximity to the major liver vasculature(PMV)even after a full course of neoadjuvant chemotherapy(NAC).In such cases,extreme liver resection can be considered a potential option.AIM To explore whether computer-assisted three-dimensional individualized extreme liver resection is safe and feasible for children with HB who still have PMV after a full course of NAC.METHODS We retrospectively collected data from children with HB who underwent surgical resection at our center from June 2013 to June 2023.We then analyzed the detailed clinical and three-dimensional characteristics of children with HB who still had PMV after a full course of NAC.RESULTS Sixty-seven children diagnosed with HB underwent surgical resection.The age at diagnosis was 21.4±18.8 months,and 40 boys and 27 girls were included.Fifty-nine(88.1%)patients had a single tumor,39(58.2%)of which was located in the right lobe of the liver.A total of 47 patients(70.1%)had PRE-TEXT III or IV.Thirty-nine patients(58.2%)underwent delayed resection.After a full course of NAC,16 patients still had close PMV(within 1 cm in two patients,touching in 11 patients,compressing in four patients,and showing tumor thrombus in three patients).There were 6 patients of tumors in the middle lobe of the liver,and four of those patients exhibited liver anatomy variations.These 16 children underwent extreme liver resection after comprehensive preoperative evaluation.Intraoperative procedures were performed according to the preoperative plan,and the operations were successfully performed.Currently,the 3-year event-free survival of 67 children with HB is 88%.Among the 16 children who underwent extreme liver resection,three experienced recurrence,and one died due to multiple metastases.CONCLUSION Extreme liver resection for HB that is still in close PMV after a full course of NAC is both safe and feasible.This approach not only reduces the necessity for liver transplantation but also results in a favorable prognosis.Individualized three-dimensional surgical planning is beneficial for accurate and complete resection of HB,particularly for assessing vascular involvement,remnant liver volume and anatomical variations.展开更多
When data privacy is imposed as a necessity,Federated learning(FL)emerges as a relevant artificial intelligence field for developing machine learning(ML)models in a distributed and decentralized environment.FL allows ...When data privacy is imposed as a necessity,Federated learning(FL)emerges as a relevant artificial intelligence field for developing machine learning(ML)models in a distributed and decentralized environment.FL allows ML models to be trained on local devices without any need for centralized data transfer,thereby reducing both the exposure of sensitive data and the possibility of data interception by malicious third parties.This paradigm has gained momentum in the last few years,spurred by the plethora of real-world applications that have leveraged its ability to improve the efficiency of distributed learning and to accommodate numerous participants with their data sources.By virtue of FL,models can be learned from all such distributed data sources while preserving data privacy.The aim of this paper is to provide a practical tutorial on FL,including a short methodology and a systematic analysis of existing software frameworks.Furthermore,our tutorial provides exemplary cases of study from three complementary perspectives:i)Foundations of FL,describing the main components of FL,from key elements to FL categories;ii)Implementation guidelines and exemplary cases of study,by systematically examining the functionalities provided by existing software frameworks for FL deployment,devising a methodology to design a FL scenario,and providing exemplary cases of study with source code for different ML approaches;and iii)Trends,shortly reviewing a non-exhaustive list of research directions that are under active investigation in the current FL landscape.The ultimate purpose of this work is to establish itself as a referential work for researchers,developers,and data scientists willing to explore the capabilities of FL in practical applications.展开更多
In recent years,the rapid development of computer software has led to numerous security problems,particularly software vulnerabilities.These flaws can cause significant harm to users’privacy and property.Current secu...In recent years,the rapid development of computer software has led to numerous security problems,particularly software vulnerabilities.These flaws can cause significant harm to users’privacy and property.Current security defect detection technology relies on manual or professional reasoning,leading to missed detection and high false detection rates.Artificial intelligence technology has led to the development of neural network models based on machine learning or deep learning to intelligently mine holes,reducing missed alarms and false alarms.So,this project aims to study Java source code defect detection methods for defects like null pointer reference exception,XSS(Transform),and Structured Query Language(SQL)injection.Also,the project uses open-source Javalang to translate the Java source code,conducts a deep search on the AST to obtain the empty syntax feature library,and converts the Java source code into a dependency graph.The feature vector is then used as the learning target for the neural network.Four types of Convolutional Neural Networks(CNN),Long Short-Term Memory(LSTM),Bi-directional Long Short-Term Memory(BiLSTM),and Attention Mechanism+Bidirectional LSTM,are used to investigate various code defects,including blank pointer reference exception,XSS,and SQL injection defects.Experimental results show that the attention mechanism in two-dimensional BLSTM is the most effective for object recognition,verifying the correctness of the method.展开更多
Understanding the pore water pressure distribution in unsaturated soil is crucial in predicting shallow landslides triggered by rainfall,mainly when dealing with different temporal patterns of rainfall intensity.Howev...Understanding the pore water pressure distribution in unsaturated soil is crucial in predicting shallow landslides triggered by rainfall,mainly when dealing with different temporal patterns of rainfall intensity.However,the hydrological response of vegetated slopes,especially three-dimensional(3D)slopes covered with shrubs,under different rainfall patterns remains unclear and requires further investigation.To address this issue,this study adopts a novel 3D numerical model for simulating hydraulic interactions between the root system of the shrub and the surrounding soil.Three series of numerical parametric studies are conducted to investigate the influences of slope inclination,rainfall pattern and rainfall duration.Four rainfall patterns(advanced,bimodal,delayed,and uniform)and two rainfall durations(4-h intense and 168-h mild rainfall)are considered to study the hydrological response of the slope.The computed results show that 17%higher transpiration-induced suction is found for a steeper slope,which remains even after a short,intense rainfall with a 100-year return period.The extreme rainfalls with advanced(PA),bimodal(PB)and uniform(PU)rainfall patterns need to be considered for the short rainfall duration(4 h),while the delayed(PD)and uniform(PU)rainfall patterns are highly recommended for long rainfall durations(168 h).The presence of plants can improve slope stability markedly under extreme rainfall with a short duration(4 h).For the long duration(168 h),the benefit of the plant in preserving pore-water pressure(PWP)and slope stability may not be sufficient.展开更多
The Internet of Things(IoT)has characteristics such as node mobility,node heterogeneity,link heterogeneity,and topology heterogeneity.In the face of the IoT characteristics and the explosive growth of IoT nodes,which ...The Internet of Things(IoT)has characteristics such as node mobility,node heterogeneity,link heterogeneity,and topology heterogeneity.In the face of the IoT characteristics and the explosive growth of IoT nodes,which brings about large-scale data processing requirements,edge computing architecture has become an emerging network architecture to support IoT applications due to its ability to provide powerful computing capabilities and good service functions.However,the defense mechanism of Edge Computing-enabled IoT Nodes(ECIoTNs)is still weak due to their limited resources,so that they are susceptible to malicious software spread,which can compromise data confidentiality and network service availability.Facing this situation,we put forward an epidemiology-based susceptible-curb-infectious-removed-dead(SCIRD)model.Then,we analyze the dynamics of ECIoTNs with different infection levels under different initial conditions to obtain the dynamic differential equations.Additionally,we establish the presence of equilibrium states in the SCIRD model.Furthermore,we conduct an analysis of the model’s stability and examine the conditions under which malicious software will either spread or disappear within Edge Computing-enabled IoT(ECIoT)networks.Lastly,we validate the efficacy and superiority of the SCIRD model through MATLAB simulations.These research findings offer a theoretical foundation for suppressing the propagation of malicious software in ECIoT networks.The experimental results indicate that the theoretical SCIRD model has instructive significance,deeply revealing the principles of malicious software propagation in ECIoT networks.This study solves a challenging security problem of ECIoT networks by determining the malicious software propagation threshold,which lays the foundation for buildingmore secure and reliable ECIoT networks.展开更多
基金Project supported by the National Natural Science Foundation of China (No. 12072337)。
文摘Through combined applications of the transfer-matrix method and asymptotic expansion technique,we formulate a theory to predict the three-dimensional response of micropolar plates.No ad hoc assumptions regarding through-thickness assumptions of the field variables are made,and the governing equations are two-dimensional,with the displacements and microrotations of the mid-plane as the unknowns.Once the deformation of the mid-plane is solved,a three-dimensional micropolar elastic field within the plate is generated,which is exact up to the second order except in the boundary region close to the plate edge.As an illustrative example,the bending of a clamped infinitely long plate caused by a uniformly distributed transverse force is analyzed and discussed in detail.
基金supported by the National Natural Science Foundation of China (No. 52275291)the Fundamental Research Funds for the Central Universitiesthe Program for Innovation Team of Shaanxi Province,China (No. 2023-CX-TD-17)
文摘Hypoxia is a typical feature of the tumor microenvironment,one of the most critical factors affecting cell behavior and tumor progression.However,the lack of tumor models able to precisely emulate natural brain tumor tissue has impeded the study of the effects of hypoxia on the progression and growth of tumor cells.This study reports a three-dimensional(3D)brain tumor model obtained by encapsulating U87MG(U87)cells in a hydrogel containing type I collagen.It also documents the effect of various oxygen concentrations(1%,7%,and 21%)in the culture environment on U87 cell morphology,proliferation,viability,cell cycle,apoptosis rate,and migration.Finally,it compares two-dimensional(2D)and 3D cultures.For comparison purposes,cells cultured in flat culture dishes were used as the control(2D model).Cells cultured in the 3D model proliferated more slowly but had a higher apoptosis rate and proportion of cells in the resting phase(G0 phase)/gap I phase(G1 phase)than those cultured in the 2D model.Besides,the two models yielded significantly different cell morphologies.Finally,hypoxia(e.g.,1%O2)affected cell morphology,slowed cell growth,reduced cell viability,and increased the apoptosis rate in the 3D model.These results indicate that the constructed 3D model is effective for investigating the effects of biological and chemical factors on cell morphology and function,and can be more representative of the tumor microenvironment than 2D culture systems.The developed 3D glioblastoma tumor model is equally applicable to other studies in pharmacology and pathology.
基金This work was supported by grants fromthe Sichuan Science and Technology Program(2023NSFSC1877).
文摘Liver regeneration and the development of effective therapies for liver failure remain formidable challenges in modern medicine.In recent years,the utilization of 3D cell-based strategies has emerged as a promising approach for addressing these urgent clinical requirements.This review provides a thorough analysis of the application of 3D cell-based approaches to liver regeneration and their potential impact on patients with end-stage liver failure.Here,we discuss various 3D culture models that incorporate hepatocytes and stem cells to restore liver function and ameliorate the consequences of liver failure.Furthermore,we explored the challenges in transitioning these innovative strategies from preclinical studies to clinical applications.The collective insights presented herein highlight the significance of 3D cell-based strategies as a transformative paradigm for liver regeneration and improved patient care.
基金supported by the NationalNatural Science Foundation of China(Grant No.61867004)the Youth Fund of the National Natural Science Foundation of China(Grant No.41801288).
文摘The purpose of software defect prediction is to identify defect-prone code modules to assist software quality assurance teams with the appropriate allocation of resources and labor.In previous software defect prediction studies,transfer learning was effective in solving the problem of inconsistent project data distribution.However,target projects often lack sufficient data,which affects the performance of the transfer learning model.In addition,the presence of uncorrelated features between projects can decrease the prediction accuracy of the transfer learning model.To address these problems,this article propose a software defect prediction method based on stable learning(SDP-SL)that combines code visualization techniques and residual networks.This method first transforms code files into code images using code visualization techniques and then constructs a defect prediction model based on these code images.During the model training process,target project data are not required as prior knowledge.Following the principles of stable learning,this paper dynamically adjusted the weights of source project samples to eliminate dependencies between features,thereby capturing the“invariance mechanism”within the data.This approach explores the genuine relationship between code defect features and labels,thereby enhancing defect prediction performance.To evaluate the performance of SDP-SL,this article conducted comparative experiments on 10 open-source projects in the PROMISE dataset.The experimental results demonstrated that in terms of the F-measure,the proposed SDP-SL method outperformed other within-project defect prediction methods by 2.11%-44.03%.In cross-project defect prediction,the SDP-SL method provided an improvement of 5.89%-25.46% in prediction performance compared to other cross-project defect prediction methods.Therefore,SDP-SL can effectively enhance within-and cross-project defect predictions.
基金supported by the National Natural Science Foundation of China,No.82171380(to CD)Jiangsu Students’Platform for Innovation and Entrepreneurship Training Program,No.202110304098Y(to DJ)。
文摘Spinal cord injury is considered one of the most difficult injuries to repair and has one of the worst prognoses for injuries to the nervous system.Following surgery,the poor regenerative capacity of nerve cells and the generation of new scars can make it very difficult for the impaired nervous system to restore its neural functionality.Traditional treatments can only alleviate secondary injuries but cannot fundamentally repair the spinal cord.Consequently,there is a critical need to develop new treatments to promote functional repair after spinal cord injury.Over recent years,there have been seve ral developments in the use of stem cell therapy for the treatment of spinal cord injury.Alongside significant developments in the field of tissue engineering,three-dimensional bioprinting technology has become a hot research topic due to its ability to accurately print complex structures.This led to the loading of three-dimensional bioprinting scaffolds which provided precise cell localization.These three-dimensional bioprinting scaffolds co uld repair damaged neural circuits and had the potential to repair the damaged spinal cord.In this review,we discuss the mechanisms underlying simple stem cell therapy,the application of different types of stem cells for the treatment of spinal cord injury,and the different manufa cturing methods for three-dimensional bioprinting scaffolds.In particular,we focus on the development of three-dimensional bioprinting scaffolds for the treatment of spinal cord injury.
基金supported by UniversitiKebangsaan Malaysia,under Dana Impak Perdana 2.0.(Ref:DIP–2022–020).
文摘Software Defined Networking(SDN)is programmable by separation of forwarding control through the centralization of the controller.The controller plays the role of the‘brain’that dictates the intelligent part of SDN technology.Various versions of SDN controllers exist as a response to the diverse demands and functions expected of them.There are several SDN controllers available in the open market besides a large number of commercial controllers;some are developed tomeet carrier-grade service levels and one of the recent trends in open-source SDN controllers is the Open Network Operating System(ONOS).This paper presents a comparative study between open source SDN controllers,which are known as Network Controller Platform(NOX),Python-based Network Controller(POX),component-based SDN framework(Ryu),Java-based OpenFlow controller(Floodlight),OpenDayLight(ODL)and ONOS.The discussion is further extended into ONOS architecture,as well as,the evolution of ONOS controllers.This article will review use cases based on ONOS controllers in several application deployments.Moreover,the opportunities and challenges of open source SDN controllers will be discussed,exploring carriergrade ONOS for future real-world deployments,ONOS unique features and identifying the suitable choice of SDN controller for service providers.In addition,we attempt to provide answers to several critical questions relating to the implications of the open-source nature of SDN controllers regarding vendor lock-in,interoperability,and standards compliance,Similarly,real-world use cases of organizations using open-source SDN are highlighted and how the open-source community contributes to the development of SDN controllers.Furthermore,challenges faced by open-source projects,and considerations when choosing an open-source SDN controller are underscored.Then the role of Artificial Intelligence(AI)and Machine Learning(ML)in the evolution of open-source SDN controllers in light of recent research is indicated.In addition,the challenges and limitations associated with deploying open-source SDN controllers in production networks,how can they be mitigated,and finally how opensource SDN controllers handle network security and ensure that network configurations and policies are robust and resilient are presented.Potential opportunities and challenges for future Open SDN deployment are outlined to conclude the article.
基金the Deanship of Scientific Research at King Abdulaziz University,Jeddah,Saudi Arabia under the Grant No.RG-12-611-43.
文摘The Message Passing Interface (MPI) is a widely accepted standard for parallel computing on distributed memorysystems.However, MPI implementations can contain defects that impact the reliability and performance of parallelapplications. Detecting and correcting these defects is crucial, yet there is a lack of published models specificallydesigned for correctingMPI defects. To address this, we propose a model for detecting and correcting MPI defects(DC_MPI), which aims to detect and correct defects in various types of MPI communication, including blockingpoint-to-point (BPTP), nonblocking point-to-point (NBPTP), and collective communication (CC). The defectsaddressed by the DC_MPI model include illegal MPI calls, deadlocks (DL), race conditions (RC), and messagemismatches (MM). To assess the effectiveness of the DC_MPI model, we performed experiments on a datasetconsisting of 40 MPI codes. The results indicate that the model achieved a detection rate of 37 out of 40 codes,resulting in an overall detection accuracy of 92.5%. Additionally, the execution duration of the DC_MPI modelranged from 0.81 to 1.36 s. These findings show that the DC_MPI model is useful in detecting and correctingdefects in MPI implementations, thereby enhancing the reliability and performance of parallel applications. TheDC_MPImodel fills an important research gap and provides a valuable tool for improving the quality ofMPI-basedparallel computing systems.
文摘Software testing is a critical phase due to misconceptions about ambiguities in the requirements during specification,which affect the testing process.Therefore,it is difficult to identify all faults in software.As requirement changes continuously,it increases the irrelevancy and redundancy during testing.Due to these challenges;fault detection capability decreases and there arises a need to improve the testing process,which is based on changes in requirements specification.In this research,we have developed a model to resolve testing challenges through requirement prioritization and prediction in an agile-based environment.The research objective is to identify the most relevant and meaningful requirements through semantic analysis for correct change analysis.Then compute the similarity of requirements through case-based reasoning,which predicted the requirements for reuse and restricted to error-based requirements.Afterward,the apriori algorithm mapped out requirement frequency to select relevant test cases based on frequently reused or not reused test cases to increase the fault detection rate.Furthermore,the proposed model was evaluated by conducting experiments.The results showed that requirement redundancy and irrelevancy improved due to semantic analysis,which correctly predicted the requirements,increasing the fault detection rate and resulting in high user satisfaction.The predicted requirements are mapped into test cases,increasing the fault detection rate after changes to achieve higher user satisfaction.Therefore,the model improves the redundancy and irrelevancy of requirements by more than 90%compared to other clustering methods and the analytical hierarchical process,achieving an 80%fault detection rate at an earlier stage.Hence,it provides guidelines for practitioners and researchers in the modern era.In the future,we will provide the working prototype of this model for proof of concept.
文摘Software Development Life Cycle (SDLC) is one of the major ingredients for the development of efficient software systems within a time frame and low-cost involvement. From the literature, it is evident that there are various kinds of process models that are used by the software industries for the development of small, medium and long-term software projects, but many of them do not cover risk management. It is quite obvious that the improper selection of the software development process model leads to failure of the software products as it is time bound activity. In the present work, a new software development process model is proposed which covers the risks at any stage of the development of the software product. The model is named a Hemant-Vipin (HV) process model and may be helpful for the software industries for development of the efficient software products and timely delivery at the end of the client. The efficiency of the HV process model is observed by considering various kinds of factors like requirement clarity, user feedback, change agility, predictability, risk identification, practical implementation, customer satisfaction, incremental development, use of ready-made components, quick design, resource organization and many more and found through a case study that the presented approach covers many of parameters in comparison of the existing process models. .
基金funding from the European Commission for the Ruralities Project(grant agreement no.101060876).
文摘Agile Transformations are challenging processes for organizations that look to extend the benefits of Agile philosophy and methods beyond software engineering.Despite the impact of these transformations on orga-nizations,they have not been extensively studied in academia.We conducted a study grounded in workshops and interviews with 99 participants from 30 organizations,including organizations undergoing transformations(“final organizations”)and companies supporting these processes(“consultants”).The study aims to understand the motivations,objectives,and factors driving and challenging these transformations.Over 700 responses were collected to the question and categorized into 32 objectives.The findings show that organizations primarily aim to achieve customer centricity and adaptability,both with 8%of the mentions.Other primary important objectives,with above 4%of mentions,include alignment of goals,lean delivery,sustainable processes,and a flatter,more team-based organizational structure.We also detect discrepancies in perspectives between the objectives identified by the two kinds of organizations and the existing agile literature and models.This misalignment highlights the need for practitioners to understand with the practical realities the organizations face.
文摘Accurate software cost estimation in Global Software Development(GSD)remains challenging due to reliance on historical data and expert judgments.Traditional models,such as the Constructive Cost Model(COCOMO II),rely heavily on historical and accurate data.In addition,expert judgment is required to set many input parameters,which can introduce subjectivity and variability in the estimation process.Consequently,there is a need to improve the current GSD models to mitigate reliance on historical data,subjectivity in expert judgment,inadequate consideration of GSD-based cost drivers and limited integration of modern technologies with cost overruns.This study introduces a novel hybrid model that synergizes the COCOMO II with Artificial Neural Networks(ANN)to address these challenges.The proposed hybrid model integrates additional GSD-based cost drivers identified through a systematic literature review and further vetted by industry experts.This article compares the effectiveness of the proposedmodelwith state-of-the-artmachine learning-basedmodels for software cost estimation.Evaluating the NASA 93 dataset by adopting twenty-six GSD-based cost drivers reveals that our hybrid model achieves superior accuracy,outperforming existing state-of-the-artmodels.The findings indicate the potential of combining COCOMO II,ANN,and additional GSD-based cost drivers to transform cost estimation in GSD.
文摘Purpose: To clarify the effectiveness of 3-D delivery animation software for the mother’s and husband’s satisfaction with delivery. Subjects and Method: We independently developed a software application used to display the pelvic region and explain the labor process. The study involved a collaboration with hospital staff who recruited 18 primiparous and 18 multiparous mothers who were hospitalized for delivery at Facility A. The midwife explained the process of delivery using the “Delivery Animation Software”. A self-administered, anonymous questionnaire was distributed and analyzed separately for primiparous and multiparous mothers and their husbands. Results: 1) For both primiparous and multiparous couples, both mothers and their husbands gained a significantly higher level of understanding after delivery than during pregnancy. 2) The Self-Evaluation Scale for Experience of Delivery results were as follows: “I did my best for the baby even if it was painful” was selected more often for “birth coping skills”;“reliable medical staff” was selected more often for “physiological birth process”;“the birth progressed as I expected” was selected frequently by primiparous mothers;and “the birth progressed smoothly” was selected often by multiparous mothers. 3) In terms of husbands’ satisfaction with the delivery, “I was satisfied with the delivery”, “I was given an easy-to-understand explanation”, and “They explained the process to me” was selected of primiparous and multiparous fathers. 4) All primiparous and multiparous mothers positively evaluated whether the delivery animation was helpful in understanding the process of delivery. Conclusion: The delivery animation was effective in improving the understanding and satisfaction of both the mothers and their husbands.
基金supported by the National Magnetic Confinement Fusion Energy R&D Program of China(Nos.2018YFE0309100 and 2019YFE03010004)National Natural Science Foundation of China(No.51821005)。
文摘A toroidal soft x-ray imaging(T-SXRI)system has been developed to investigate threedimensional(3D)plasma physics on J-TEXT.This T-SXRI system consists of three sets of SXR arrays.Two sets are newly developed and located on the vacuum chamber wall at toroidal positionsφof 126.4°and 272.6°,respectively,while one set was established previously atφ=65.50.Each set of SXR arrays consists of three arrays viewing the plasma poloidally,and hence can be used separately to obtain SXR images via the tomographic method.The sawtooth precursor oscillations are measured by T-SXRI,and the corresponding images of perturbative SXR signals are successfully reconstructed at these three toroidal positions,hence providing measurement of the 3D structure of precursor oscillations.The observed 3D structure is consistent with the helical structure of the m/n=1/1 mode.The experimental observation confirms that the T-SXRI system is able to observe 3D structures in the J-TEXT plasma.
文摘BACKGROUND Acetabular component positioning in total hip arthroplasty(THA)is of key importance to ensure satisfactory post-operative outcomes and to minimize the risk of complications.The majority of acetabular components are aligned freehand,without the use of navigation methods.Patient specific instruments(PSI)and three-dimensional(3D)printing of THA placement guides are increasingly used in primary THA to ensure optimal positioning.AIM To summarize the literature on 3D printing in THA and how they improve acetabular component alignment.METHODS PubMed was used to identify and access scientific studies reporting on different 3D printing methods used in THA.Eight studies with 236 hips in 228 patients were included.The studies could be divided into two main categories;3D printed models and 3D printed guides.RESULTS 3D printing in THA helped improve preoperative cup size planning and post-operative Harris hip scores between intervention and control groups(P=0.019,P=0.009).Otherwise,outcome measures were heterogeneous and thus difficult to compare.The overarching consensus between the studies is that the use of 3D guidance tools can assist in improving THA cup positioning and reduce the need for revision THA and the associated costs.CONCLUSION The implementation of 3D printing and PSI for primary THA can significantly improve the positioning accuracy of the acetabular cup component and reduce the number of complications caused by malpositioning.
文摘Software delivery is vital for modern organizations, driving innovation and competitiveness. Measuring an organization’s maturity in software delivery is crucial for efficiency and quality. The Capability Maturity Model (CMM) framework provides a roadmap for improvement but assessing an organization’s CMM Level is challenging. This paper offers a quantitative approach tailored to the CMM framework, using Goal-Question-Metric (GQM) frame-works for each key process area (KPA). These frameworks include metrics and questions to compute maturity scores effectively. The study also refines practices into questions for a thorough assessment. The result is an Analysis Matrix that calculates weighted scores and an overall maturity score. This approach helps organizations assess and enhance their software delivery processes systematically, aiming for improved practices and growth.
基金Supported by National Natural Science Foundation of China,No.82293665Anhui Provincial Department of Education University Research Project,No.2023AH051763.
文摘BACKGROUND The management of hepatoblastoma(HB)becomes challenging when the tumor remains in close proximity to the major liver vasculature(PMV)even after a full course of neoadjuvant chemotherapy(NAC).In such cases,extreme liver resection can be considered a potential option.AIM To explore whether computer-assisted three-dimensional individualized extreme liver resection is safe and feasible for children with HB who still have PMV after a full course of NAC.METHODS We retrospectively collected data from children with HB who underwent surgical resection at our center from June 2013 to June 2023.We then analyzed the detailed clinical and three-dimensional characteristics of children with HB who still had PMV after a full course of NAC.RESULTS Sixty-seven children diagnosed with HB underwent surgical resection.The age at diagnosis was 21.4±18.8 months,and 40 boys and 27 girls were included.Fifty-nine(88.1%)patients had a single tumor,39(58.2%)of which was located in the right lobe of the liver.A total of 47 patients(70.1%)had PRE-TEXT III or IV.Thirty-nine patients(58.2%)underwent delayed resection.After a full course of NAC,16 patients still had close PMV(within 1 cm in two patients,touching in 11 patients,compressing in four patients,and showing tumor thrombus in three patients).There were 6 patients of tumors in the middle lobe of the liver,and four of those patients exhibited liver anatomy variations.These 16 children underwent extreme liver resection after comprehensive preoperative evaluation.Intraoperative procedures were performed according to the preoperative plan,and the operations were successfully performed.Currently,the 3-year event-free survival of 67 children with HB is 88%.Among the 16 children who underwent extreme liver resection,three experienced recurrence,and one died due to multiple metastases.CONCLUSION Extreme liver resection for HB that is still in close PMV after a full course of NAC is both safe and feasible.This approach not only reduces the necessity for liver transplantation but also results in a favorable prognosis.Individualized three-dimensional surgical planning is beneficial for accurate and complete resection of HB,particularly for assessing vascular involvement,remnant liver volume and anatomical variations.
基金the R&D&I,Spain grants PID2020-119478GB-I00 and,PID2020-115832GB-I00 funded by MCIN/AEI/10.13039/501100011033.N.Rodríguez-Barroso was supported by the grant FPU18/04475 funded by MCIN/AEI/10.13039/501100011033 and by“ESF Investing in your future”Spain.J.Moyano was supported by a postdoctoral Juan de la Cierva Formación grant FJC2020-043823-I funded by MCIN/AEI/10.13039/501100011033 and by European Union NextGenerationEU/PRTR.J.Del Ser acknowledges funding support from the Spanish Centro para el Desarrollo Tecnológico Industrial(CDTI)through the AI4ES projectthe Department of Education of the Basque Government(consolidated research group MATHMODE,IT1456-22)。
文摘When data privacy is imposed as a necessity,Federated learning(FL)emerges as a relevant artificial intelligence field for developing machine learning(ML)models in a distributed and decentralized environment.FL allows ML models to be trained on local devices without any need for centralized data transfer,thereby reducing both the exposure of sensitive data and the possibility of data interception by malicious third parties.This paradigm has gained momentum in the last few years,spurred by the plethora of real-world applications that have leveraged its ability to improve the efficiency of distributed learning and to accommodate numerous participants with their data sources.By virtue of FL,models can be learned from all such distributed data sources while preserving data privacy.The aim of this paper is to provide a practical tutorial on FL,including a short methodology and a systematic analysis of existing software frameworks.Furthermore,our tutorial provides exemplary cases of study from three complementary perspectives:i)Foundations of FL,describing the main components of FL,from key elements to FL categories;ii)Implementation guidelines and exemplary cases of study,by systematically examining the functionalities provided by existing software frameworks for FL deployment,devising a methodology to design a FL scenario,and providing exemplary cases of study with source code for different ML approaches;and iii)Trends,shortly reviewing a non-exhaustive list of research directions that are under active investigation in the current FL landscape.The ultimate purpose of this work is to establish itself as a referential work for researchers,developers,and data scientists willing to explore the capabilities of FL in practical applications.
基金This work is supported by the Provincial Key Science and Technology Special Project of Henan(No.221100240100)。
文摘In recent years,the rapid development of computer software has led to numerous security problems,particularly software vulnerabilities.These flaws can cause significant harm to users’privacy and property.Current security defect detection technology relies on manual or professional reasoning,leading to missed detection and high false detection rates.Artificial intelligence technology has led to the development of neural network models based on machine learning or deep learning to intelligently mine holes,reducing missed alarms and false alarms.So,this project aims to study Java source code defect detection methods for defects like null pointer reference exception,XSS(Transform),and Structured Query Language(SQL)injection.Also,the project uses open-source Javalang to translate the Java source code,conducts a deep search on the AST to obtain the empty syntax feature library,and converts the Java source code into a dependency graph.The feature vector is then used as the learning target for the neural network.Four types of Convolutional Neural Networks(CNN),Long Short-Term Memory(LSTM),Bi-directional Long Short-Term Memory(BiLSTM),and Attention Mechanism+Bidirectional LSTM,are used to investigate various code defects,including blank pointer reference exception,XSS,and SQL injection defects.Experimental results show that the attention mechanism in two-dimensional BLSTM is the most effective for object recognition,verifying the correctness of the method.
文摘Understanding the pore water pressure distribution in unsaturated soil is crucial in predicting shallow landslides triggered by rainfall,mainly when dealing with different temporal patterns of rainfall intensity.However,the hydrological response of vegetated slopes,especially three-dimensional(3D)slopes covered with shrubs,under different rainfall patterns remains unclear and requires further investigation.To address this issue,this study adopts a novel 3D numerical model for simulating hydraulic interactions between the root system of the shrub and the surrounding soil.Three series of numerical parametric studies are conducted to investigate the influences of slope inclination,rainfall pattern and rainfall duration.Four rainfall patterns(advanced,bimodal,delayed,and uniform)and two rainfall durations(4-h intense and 168-h mild rainfall)are considered to study the hydrological response of the slope.The computed results show that 17%higher transpiration-induced suction is found for a steeper slope,which remains even after a short,intense rainfall with a 100-year return period.The extreme rainfalls with advanced(PA),bimodal(PB)and uniform(PU)rainfall patterns need to be considered for the short rainfall duration(4 h),while the delayed(PD)and uniform(PU)rainfall patterns are highly recommended for long rainfall durations(168 h).The presence of plants can improve slope stability markedly under extreme rainfall with a short duration(4 h).For the long duration(168 h),the benefit of the plant in preserving pore-water pressure(PWP)and slope stability may not be sufficient.
基金in part by National Undergraduate Innovation and Entrepreneurship Training Program under Grant No.202310347039Zhejiang Provincial Natural Science Foundation of China under Grant No.LZ22F020002Huzhou Science and Technology Planning Foundation under Grant No.2023GZ04.
文摘The Internet of Things(IoT)has characteristics such as node mobility,node heterogeneity,link heterogeneity,and topology heterogeneity.In the face of the IoT characteristics and the explosive growth of IoT nodes,which brings about large-scale data processing requirements,edge computing architecture has become an emerging network architecture to support IoT applications due to its ability to provide powerful computing capabilities and good service functions.However,the defense mechanism of Edge Computing-enabled IoT Nodes(ECIoTNs)is still weak due to their limited resources,so that they are susceptible to malicious software spread,which can compromise data confidentiality and network service availability.Facing this situation,we put forward an epidemiology-based susceptible-curb-infectious-removed-dead(SCIRD)model.Then,we analyze the dynamics of ECIoTNs with different infection levels under different initial conditions to obtain the dynamic differential equations.Additionally,we establish the presence of equilibrium states in the SCIRD model.Furthermore,we conduct an analysis of the model’s stability and examine the conditions under which malicious software will either spread or disappear within Edge Computing-enabled IoT(ECIoT)networks.Lastly,we validate the efficacy and superiority of the SCIRD model through MATLAB simulations.These research findings offer a theoretical foundation for suppressing the propagation of malicious software in ECIoT networks.The experimental results indicate that the theoretical SCIRD model has instructive significance,deeply revealing the principles of malicious software propagation in ECIoT networks.This study solves a challenging security problem of ECIoT networks by determining the malicious software propagation threshold,which lays the foundation for buildingmore secure and reliable ECIoT networks.