Efforts were exerted to enhance the live virtual machines(VMs)migration,including performance improvements of the live migration of services to the cloud.The VMs empower the cloud users to store relevant data and reso...Efforts were exerted to enhance the live virtual machines(VMs)migration,including performance improvements of the live migration of services to the cloud.The VMs empower the cloud users to store relevant data and resources.However,the utilization of servers has increased significantly because of the virtualization of computer systems,leading to a rise in power consumption and storage requirements by data centers,and thereby the running costs.Data center migration technologies are used to reduce risk,minimize downtime,and streamline and accelerate the data center move process.Indeed,several parameters,such as non-network overheads and downtime adjustment,may impact the live migration time and server downtime to a large extent.By virtualizing the network resources,the infrastructure as a service(IaaS)can be used dynamically to allocate the bandwidth to services and monitor the network flow routing.Due to the large amount of filthy retransmission,existing live migration systems still suffer from extensive downtime and significant performance degradation in crossdata-center situations.This study aims to minimize the energy consumption by restricting the VMs migration and switching off the guests depending on a threshold,thereby boosting the residual network bandwidth in the data center with a minimal breach of the service level agreement(SLA).In this research,we analyzed and evaluated the findings observed through simulating different parameters,like availability,downtime,and outage of VMs in data center processes.This new paradigm is composed of two forms of detection strategies in the live migration approach from the source host to the destination source machine.展开更多
Understanding how evolutionary pressures related to climate change have shaped the current genetic background of domestic animals is a fundamental pursuit of biology. Here, we generated wholegenome sequencing data fro...Understanding how evolutionary pressures related to climate change have shaped the current genetic background of domestic animals is a fundamental pursuit of biology. Here, we generated wholegenome sequencing data from native goat populations in Iraq and Pakistan. Combined with previously published data on modern, ancient(Late Neolithic to Medieval periods), and wild Capra species worldwide, we explored the genetic population structure, ancestry components, and signatures of natural positive selection in native goat populations in Southwest Asia(SWA). Results revealed that the genetic structure of SWA goats was deeply influenced by gene flow from the eastern Mediterranean during the Chalcolithic period, which may reflect adaptation to gradual warming and aridity in the region. Furthermore, comparative genomic analysis revealed adaptive introgression of the KITLG locus from the Nubian ibex(C. nubiana) into African and SWA goats. The frequency of the selected allele at this locus was significantly higher among goat populations located near northeastern Africa. These results provide new insights into the genetic composition and history of goat populations in the SWA region.展开更多
The software engineering field has long focused on creating high-quality software despite limited resources.Detecting defects before the testing stage of software development can enable quality assurance engineers to ...The software engineering field has long focused on creating high-quality software despite limited resources.Detecting defects before the testing stage of software development can enable quality assurance engineers to con-centrate on problematic modules rather than all the modules.This approach can enhance the quality of the final product while lowering development costs.Identifying defective modules early on can allow for early corrections and ensure the timely delivery of a high-quality product that satisfies customers and instills greater confidence in the development team.This process is known as software defect prediction,and it can improve end-product quality while reducing the cost of testing and maintenance.This study proposes a software defect prediction system that utilizes data fusion,feature selection,and ensemble machine learning fusion techniques.A novel filter-based metric selection technique is proposed in the framework to select the optimum features.A three-step nested approach is presented for predicting defective modules to achieve high accuracy.In the first step,three supervised machine learning techniques,including Decision Tree,Support Vector Machines,and Naïve Bayes,are used to detect faulty modules.The second step involves integrating the predictive accuracy of these classification techniques through three ensemble machine-learning methods:Bagging,Voting,and Stacking.Finally,in the third step,a fuzzy logic technique is employed to integrate the predictive accuracy of the ensemble machine learning techniques.The experiments are performed on a fused software defect dataset to ensure that the developed fused ensemble model can perform effectively on diverse datasets.Five NASA datasets are integrated to create the fused dataset:MW1,PC1,PC3,PC4,and CM1.According to the results,the proposed system exhibited superior performance to other advanced techniques for predicting software defects,achieving a remarkable accuracy rate of 92.08%.展开更多
The Far North Region of Cameroon is home to a great diversity of bird species, which unfortunately remains very little explored. This work was initiated to establish an inventory of birds and the factors affecting the...The Far North Region of Cameroon is home to a great diversity of bird species, which unfortunately remains very little explored. This work was initiated to establish an inventory of birds and the factors affecting their diversity and distribution for sustainable management in the Kalfou Forest Reserve (KFR) and its periphery. Two methods were used for sampling, linear strip transects from which direct counts and indirect observations were made and the mist netting to complement the first. In total, 2525 birds were observed, including 149 species, belonging to 20 orders and 55 families. Accipitridae had the greatest number of species (11). The species richness was greater in the KFR (117 species) compared to the periphery (95 species). The specific richness was higher in wooded savannah compared to other habitats. Shannon index was significantly higher in the KFR (3.99) compared to that obtained in the periphery (3.80). The value of the Simpson index was higher on the outskirts of the KFR than on the periphery. The indices of species diversity were greater in the wooded savannah compared to other vegetation types. The seasons had no influence on bird diversity. Among the human activities encountered, the pressure indices were more important for grazing (7.3 contacts/km). Human activities have resulted in a significant decrease in specific richness. Six endangered species were encountered, four belonging to the Accipitridae family. The greater bird diversity in the reserve compared to the periphery shows that protected areas are a long-term solution for biodiversity conservation.展开更多
Artificial intelligence(AI)is expanding its roots in medical diagnostics.Various acute and chronic diseases can be identified accurately at the initial level by using AI methods to prevent the progression of health co...Artificial intelligence(AI)is expanding its roots in medical diagnostics.Various acute and chronic diseases can be identified accurately at the initial level by using AI methods to prevent the progression of health complications.Kidney diseases are producing a high impact on global health and medical practitioners are suggested that the diagnosis at earlier stages is one of the foremost approaches to avert chronic kidney disease and renal failure.High blood pressure,diabetes mellitus,and glomerulonephritis are the root causes of kidney disease.Therefore,the present study is proposed a set of multiple techniques such as simulation,modeling,and optimization of intelligent kidney disease prediction(SMOIKD)which is based on computational intelligence approaches.Initially,seven parameters were used for the fuzzy logic system(FLS),and then twenty-five different attributes of the kidney dataset were used for the artificial neural network(ANN)and deep extreme machine learning(DEML).The expert system was proposed with the assistance of medical experts.For the quick and accurate evaluation of the proposed system,Matlab version 2019 was used.The proposed SMOIKD-FLSANN-DEML expert system has shown 94.16%accuracy.Hence this study concluded that SMOIKD-FLS-ANN-DEML system is effective to accurately diagnose kidney disease at initial levels.展开更多
Internet of Things(IoT)network used for industrial management is vulnerable to different security threats due to its unstructured deployment,and dynamic communication behavior.In literature various mechanisms addresse...Internet of Things(IoT)network used for industrial management is vulnerable to different security threats due to its unstructured deployment,and dynamic communication behavior.In literature various mechanisms addressed the security issue of Industrial IoT networks,but proper maintenance of the performance reliability is among the common challenges.In this paper,we proposed an intelligent mutual authentication scheme leveraging authentication aware node(AAN)and base station(BS)to identify routing attacks in Industrial IoT networks.The AAN and BS uses the communication parameter such as a route request(RREQ),node-ID,received signal strength(RSS),and round-trip time(RTT)information to identify malicious devices and routes in the deployed network.The feasibility of the proposed model is validated in the simulation environment,where OMNeT++was used as a simulation tool.We compare the results of the proposed model with existing field-proven schemes in terms of routing attacks detection,communication cost,latency,computational cost,and throughput.The results show that our proposed scheme surpasses the previous schemes regarding these performance parameters with the attack detection rate of 97.7%.展开更多
The internet,particularly online social networking platforms have revolutionized the way extremist groups are influencing and radicalizing individuals.Recent research reveals that the process initiates by exposing vas...The internet,particularly online social networking platforms have revolutionized the way extremist groups are influencing and radicalizing individuals.Recent research reveals that the process initiates by exposing vast audiences to extremist content and then migrating potential victims to confined platforms for intensive radicalization.Consequently,social networks have evolved as a persuasive tool for extremism aiding as recruitment platform and psychological warfare.Thus,recognizing potential radical text or material is vital to restrict the circulation of the extremist chronicle.The aim of this research work is to identify radical text in social media.Our contributions are as follows:(i)A new dataset to be employed in radicalization detection;(ii)In depth analysis of new and previous datasets so that the variation in extremist group narrative could be identified;(iii)An approach to train classifier employing religious features along with radical features to detect radicalization;(iv)Observing the use of violent and bad words in radical,neutral and random groups by employing violent,terrorism and bad words dictionaries.Our research results clearly indicate that incorporating religious text in model training improves the accuracy,precision,recall,and F1-score of the classifiers.Secondly a variation in extremist narrative has been observed implying that usage of new dataset can have substantial effect on classifier performance.In addition to this,violence and bad words are creating a differentiating factor between radical and random users but for neutral(anti-ISIS)group it needs further investigation.展开更多
Lysine Lipoylation is a protective and conserved Post Translational Modification(PTM)in proteomics research like prokaryotes and eukaryotes.It is connected with many biological processes and closely linked with many m...Lysine Lipoylation is a protective and conserved Post Translational Modification(PTM)in proteomics research like prokaryotes and eukaryotes.It is connected with many biological processes and closely linked with many metabolic diseases.To develop a perfect and accurate classification model for identifying lipoylation sites at the protein level,the computational methods and several other factors play a key role in this purpose.Usually,most of the techniques and different traditional experimental models have a very high cost.They are time-consuming;so,it is required to construct a predictor model to extract lysine lipoylation sites.This study proposes a model that could predict lysine lipoylation sites with the help of a classification method known as Artificial Neural Network(ANN).The ANN algorithm deals with the noise problem and imbalance classification in lipoylation sites dataset samples.As the result shows in ten-fold cross-validation,a brilliant performance is achieved through the predictor model with an accuracy of 99.88%,and also achieved 0.9976 as the highest value of MCC.So,the predictor model is a very useful and helpful tool for lipoylation sites prediction.Some of the residues around lysine lipoylation sites play a vital part in prediction,as demonstrated during feature analysis.The wonderful results reported through the evaluation and prediction of this model can provide an informative and relative explanation for lipoylation and its molecular mechanisms.展开更多
Cloud computing is becoming popular technology due to its functional properties and variety of customer-oriented services over the Internet.The design of reliable and high-quality cloud applications requires a strong ...Cloud computing is becoming popular technology due to its functional properties and variety of customer-oriented services over the Internet.The design of reliable and high-quality cloud applications requires a strong Quality of Service QoS parameter metric.In a hyperconverged cloud ecosystem environment,building high-reliability cloud applications is a challenging job.The selection of cloud services is based on the QoS parameters that play essential roles in optimizing and improving cloud rankings.The emergence of cloud computing is significantly reshaping the digital ecosystem,and the numerous services offered by cloud service providers are playing a vital role in this transformation.Hyperconverged software-based unified utilities combine storage virtualization,compute virtualization,and network virtualization.The availability of the latter has also raised the demand for QoS.Due to the diversity of services,the respective quality parameters are also in abundance and need a carefully designed mechanism to compare and identify the critical,common,and impactful parameters.It is also necessary to reconsider the market needs in terms of service requirements and the QoS provided by various CSPs.This research provides a machine learning-based mechanism to monitor the QoS in a hyperconverged environment with three core service parameters:service quality,downtime of servers,and outage of cloud services.展开更多
Nowadays rapidly increasing technology is mobile phone technology in telecommunication sector. This mobile device technology has great effect on everyone’s life. This technology has reduced the burden of people in th...Nowadays rapidly increasing technology is mobile phone technology in telecommunication sector. This mobile device technology has great effect on everyone’s life. This technology has reduced the burden of people in their daily life. To manage the rising demand for such mobile devices, numerous operating systems came in the market as a platform upon which modern application can be produced. As a result, numbers of platforms and essential depository describe these platforms;customers may or may not be aware of these platforms that are appropriate for their needs. In order to solve this issue, we examine the most famous mobile phone operating systems to decide which operating system is most suitable for developers, business applications as well as casual use. In this paper we make assessment on the popular operating systems of mobile devices available in the business market, and on behalf of such assessment we distinguish that operating system OS is much useful of its particular characteristics compared with other systems.展开更多
The requirement for high-quality seafood is a global challenge in today’s world due to climate change and natural resource limitations.Internet of Things(IoT)based Modern fish farming systems can significantly optimi...The requirement for high-quality seafood is a global challenge in today’s world due to climate change and natural resource limitations.Internet of Things(IoT)based Modern fish farming systems can significantly optimize seafood production by minimizing resource utilization and improving healthy fish production.This objective requires intensive monitoring,prediction,and control by optimizing leading factors that impact fish growth,including temperature,the potential of hydrogen(pH),water level,and feeding rate.This paper proposes the IoT based predictive optimization approach for efficient control and energy utilization in smart fish farming.The proposed fish farm control mechanism has a predictive optimization to deal with water quality control and efficient energy consumption problems.Fish farm indoor and outdoor values are applied to predict the water quality parameters,whereas a novel objective function is proposed to achieve an optimal fish growth environment based on predicted parameters.Fuzzy logic control is utilized to calculate control parameters for IoT actuators based on predictive optimal water quality parameters by minimizing energy consumption.To evaluate the efficiency of the proposed system,the overall approach has been deployed to the fish tank as a case study,and a number of experiments have been carried out.The results show that the predictive optimization module allowed the water quality parameters to be maintained at the optimal level with nearly 30%of energy efficiency at the maximum actuator control rate compared with other control levels.展开更多
A well-managed financial market of stocks,commodities,derivatives,and bonds is crucial to a country’s economic growth.It provides confidence to investors,which encourages the inflow of cash to ensure good market liqu...A well-managed financial market of stocks,commodities,derivatives,and bonds is crucial to a country’s economic growth.It provides confidence to investors,which encourages the inflow of cash to ensure good market liquidity.However,there will always be a group of traders that aims to manipulate market pricing to negatively influence stock values in their favor.These illegal trading activities are surely prohibited according to the rules and regulations of every country’s stockmarket.It is the role of regulators to detect and prevent any manipulation cases in order to provide a trading platform that is fair and efficient.However,the complexity of manipulation cases has increased significantly,coupled with high trading volumes,which makes the manual observations of such cases by human operators no longer feasible.As a result,many intelligent systems have been developed by researchers all over the world to automatically detect various types of manipulation cases.Therefore,this review paper aims to comprehensively discuss the state-of-theart methods that have been developed to detect and recognize stock market manipulation cases.It also provides a concise definition of manipulation taxonomy,including manipulation types and categories,as well as some of the output of early experimental research.In summary,this paper provides a thorough review of the automated methods for detecting stock market manipulation cases.展开更多
The query optimizer uses cost-based optimization to create an execution plan with the least cost,which also consumes the least amount of resources.The challenge of query optimization for relational database systems is...The query optimizer uses cost-based optimization to create an execution plan with the least cost,which also consumes the least amount of resources.The challenge of query optimization for relational database systems is a combinatorial optimization problem,which renders exhaustive search impossible as query sizes rise.Increases in CPU performance have surpassed main memory,and disk access speeds in recent decades,allowing data compression to be used—strategies for improving database performance systems.For performance enhancement,compression and query optimization are the two most factors.Compression reduces the volume of data,whereas query optimization minimizes execution time.Compressing the database reduces memory requirement,data takes less time to load into memory,fewer buffer missing occur,and the size of intermediate results is more diminutive.This paper performed query optimization on the graph database in a cloud dew environment by considering,which requires less time to execute a query.The factors compression and query optimization improve the performance of the databases.This research compares the performance of MySQL and Neo4j databases in terms of memory usage and execution time running on cloud dew servers.展开更多
Kubernetes,a container orchestrator for cloud-deployed applications,allows the application provider to scale automatically to match thefluctuating intensity of processing demand.Container cluster technology is used to...Kubernetes,a container orchestrator for cloud-deployed applications,allows the application provider to scale automatically to match thefluctuating intensity of processing demand.Container cluster technology is used to encapsulate,isolate,and deploy applications,addressing the issue of low system reliability due to interlocking failures.Cloud-based platforms usually entail users define application resource supplies for eco container virtualization.There is a constant problem of over-service in data centers for cloud service providers.Higher operating costs and incompetent resource utilization can occur in a waste of resources.Kubernetes revolutionized the orchestration of the container in the cloud-native age.It can adaptively manage resources and schedule containers,which provide real-time status of the cluster at runtime without the user’s contribution.Kubernetes clusters face unpredictable traffic,and the cluster performs manual expansion configuration by the controller.Due to operational delays,the system will become unstable,and the service will be unavailable.This work proposed an RBACS that vigorously amended the distribution of containers operating in the entire Kubernetes cluster.RBACS allocation pattern is analyzed with the Kubernetes VPA.To estimate the overall cost of RBACS,we use several scientific benchmarks comparing the accomplishment of container to remote node migration and on-site relocation.The experiments ran on the simulations to show the method’s effectiveness yielded high precision in the real-time deployment of resources in eco containers.Compared to the default baseline,Kubernetes results in much fewer dropped requests with only slightly more supplied resources.展开更多
Leukemia is a kind of blood cancer that damages the cells in the blood and bone marrow of the human body.It produces cancerous blood cells that disturb the human’s immune system and significantly affect bone marrow’...Leukemia is a kind of blood cancer that damages the cells in the blood and bone marrow of the human body.It produces cancerous blood cells that disturb the human’s immune system and significantly affect bone marrow’s production ability to effectively create different types of blood cells like red blood cells(RBCs)and white blood cells(WBC),and platelets.Leukemia can be diagnosed manually by taking a complete blood count test of the patient’s blood,from which medical professionals can investigate the signs of leukemia cells.Furthermore,two other methods,microscopic inspection of blood smears and bone marrow aspiration,are also utilized while examining the patient for leukemia.However,all these methods are labor-intensive,slow,inaccurate,and require a lot of human experience and dedication.Different authors have proposed automated detection systems for leukemia diagnosis to overcome these limitations.They have deployed digital image processing and machine learning algorithms to classify the cells into normal and blast cells.However,these systems are more efficient,reliable,and fast than previous manual diagnosing methods.However,more work is required to classify leukemia-affected cells due to the complex characteristics of blood images and leukemia cells having much intra-class variability and inter-class similarity.In this paper,we have proposed a robust automated system to diagnose leukemia and its sub-types.We have classified ALL into its sub-types based on FAB classification,i.e.,L1,L2,and L3 types with better performance.We have achieved 96.06%accuracy for subtypes classification,which is better when compared with the state-of-the-art methodologies.展开更多
The deep learning advancements have greatly improved the performance of speech recognition systems,and most recent systems are based on the Recurrent Neural Network(RNN).Overall,the RNN works fine with the small seque...The deep learning advancements have greatly improved the performance of speech recognition systems,and most recent systems are based on the Recurrent Neural Network(RNN).Overall,the RNN works fine with the small sequence data,but suffers from the gradient vanishing problem in case of large sequence.The transformer networks have neutralized this issue and have shown state-of-the-art results on sequential or speech-related data.Generally,in speech recognition,the input audio is converted into an image using Mel-spectrogram to illustrate frequencies and intensities.The image is classified by the machine learning mechanism to generate a classification transcript.However,the audio frequency in the image has low resolution and causing inaccurate predictions.This paper presents a novel end-to-end binary view transformer-based architecture for speech recognition to cope with the frequency resolution problem.Firstly,the input audio signal is transformed into a 2D image using Mel-spectrogram.Secondly,the modified universal transformers utilize the multi-head attention to derive contextual information and derive different speech-related features.Moreover,a feedforward neural network is also deployed for classification.The proposed system has generated robust results on Google’s speech command dataset with an accuracy of 95.16%and with minimal loss.The binary-view transformer eradicates the eventuality of the over-fitting problem by deploying a multiview mechanism to diversify the input data,and multi-head attention captures multiple contexts from the data’s feature map.展开更多
Astrocytoma IV or glioblastoma is one of the fatal and dangerous types of brain tumors.Early detection of brain tumor increases the survival rate and helps in reducing the fatality rate.Various imaging modalities have...Astrocytoma IV or glioblastoma is one of the fatal and dangerous types of brain tumors.Early detection of brain tumor increases the survival rate and helps in reducing the fatality rate.Various imaging modalities have been used for diagnosing by expert radiologists,and Medical Resonance Image(MRI)is considered a better option for detecting brain tumors as MRI is a non-invasive technique and provides better visualization of the brain region.One of the challenging issues is to identify the tumorous region from the MRI scans correctly.Manual segmentation is performed by medical experts,which is a time-consuming task and got chances of errors.To overcome this issue,automatic segmentation is performed for quick and accurate results.The proposed approach is to capture inter-slice information and reduce the outliers.Deep learning-based brain tumor segmentation techniques proved best among available segmentation techniques.However,deep learning may miss some preliminary info while using MRI images during segmentation.As MRI volumes are volumetric,3D U-Net-based models are used but complex.Combinations of multiple 2D U-Net predictions in axial,sagittal,and coronal views help to capture inter-slice information.This approach may reduce the system complexity.Moreover,the Conditional Random Fields(CRF)reduce the predictions’false positives and improve the segmentation results.This model is applied to Brain Tumor Segmentation(BraTS)2019 dataset,and cross-validation is performed to check the accuracy of results.The proposed approach achieves Dice Similarity Score(DSC)of 0.77 on Enhancing Tumor(ET),0.90 on Whole Tumor(WT),and 0.84 on Tumor Core(TC)with reduced Hausdorff Distance(HD)of 3.05 on ET,5.12 on WT and 3.89 on TC.展开更多
The smart vehicles are one of critical enablers for automated services in smart cities to provide intelligent transportation means without human intervention.In order to fulfil requirements,Vehicle-to-Anything(V2X)com...The smart vehicles are one of critical enablers for automated services in smart cities to provide intelligent transportation means without human intervention.In order to fulfil requirements,Vehicle-to-Anything(V2X)communications aims to manage massive connectivity and high traffic load on base stations and extend the range over multiple hops in 5G networks.However,V2X networking faces several challenges from dynamic topology caused by high velocity of nodes and routing overhead that degrades the network performance and increases energy consumption.The existing routing scheme for V2X networking lacks energy efficiency and scalability for high velocity nodes with dense distribution.In order to handle the challenges,this article proposes a scalable and energy-efficient routing scheme called Dynamic proactive reactive routing for 5G(DPR5)for high mobility speed and dense environment.As compared to existing schemes it uses a single routing table and efficiently minimizes the energy consumption in dense environment,improves node‘s lifetime upto 42%,and optimizes network performance by reducing the packet loss ratio upto 46%in a high velocity dense environment.展开更多
Security is a significant issue for everyone due to new and creative ways to commit cybercrime.The Closed-Circuit Television(CCTV)systems are being installed in offices,houses,shopping malls,and on streets to protect ...Security is a significant issue for everyone due to new and creative ways to commit cybercrime.The Closed-Circuit Television(CCTV)systems are being installed in offices,houses,shopping malls,and on streets to protect lives.Operators monitor CCTV;however,it is difficult for a single person to monitor the actions of multiple people at one time.Consequently,there is a dire need for an automated monitoring system that detects a person with ammunition or any other harmful material Based on our research and findings of this study,we have designed a new Intelligent Ammunition Detection and Classification(IADC)system using Convolutional Neural Network(CNN).The proposed system is designed to identify persons carrying weapons and ammunition using CCTV cameras.When weapons are identified,the cameras sound an alarm.In the proposed IADC system,CNN was used to detect firearms and ammunition.The CNN model which is a Deep Learning technique consists of neural networks,most commonly applied to analyzing visual imagery has gained popularity for unstructured(images,videos)data classification.Additionally,this system generates an early warning through detection of ammunition before conditions become critical.Hence the faster and earlier the prediction,the lower the response time,loses and potential victims.The proposed IADC system provides better results than earlier published models like VGGNet,OverFeat-1,OverFeat-2,and OverFeat-3.展开更多
Cloud computing is an emerging domain that is capturing global users from all walks of life—the corporate sector,government sector,and social arena as well.Various cloud providers have offered multiple services and f...Cloud computing is an emerging domain that is capturing global users from all walks of life—the corporate sector,government sector,and social arena as well.Various cloud providers have offered multiple services and facilities to this audience and the number of providers is increasing very swiftly.This enormous pace is generating the requirement of a comprehensive ecosystem that shall provide a seamless and customized user environment not only to enhance the user experience but also to improve security,availability,accessibility,and latency.Emerging technology is providing robust solutions to many of our problems,the cloud platform is one of them.It is worth mentioning that these solutions are also amplifying the complexity and need of sustenance of these rapid solutions.As with cloud computing,new entrants as cloud service providers,resellers,tech-support,hardware manufacturers,and software developers appear on a daily basis.These actors playing their role in the growth and sustenance of the cloud ecosystem.Our objective is to use convergence for cloud services,software-defined networks,network function virtualization for infrastructure,cognition for pattern development,and knowledge repository.In order to gear up these processes,machine learning to induce intelligence to maintain ecosystem growth,to monitor performance,and to become able to make decisions for the sustenance of the ecosystem.Workloads may be programmed to“superficially”imitate most business applications and create large numbers using lightweight workload generators that merely stress the storage.In today’s current IT environment,when many enterprises use the cloud to service some of their application demands,a different performance testing technique that assesses more than the storage is necessary.Compute and storage are merged into a single building block with HCI(Hyper-converged infrastructure),resulting in a huge pool of compute and storage resources when clustered with other building blocks.The novelty of thiswork to design and test cloud storage using themeasurement of availability,downtime,and outage parameters.Results showed that the storage reliability in a hyper-converged system is above 92%.展开更多
文摘Efforts were exerted to enhance the live virtual machines(VMs)migration,including performance improvements of the live migration of services to the cloud.The VMs empower the cloud users to store relevant data and resources.However,the utilization of servers has increased significantly because of the virtualization of computer systems,leading to a rise in power consumption and storage requirements by data centers,and thereby the running costs.Data center migration technologies are used to reduce risk,minimize downtime,and streamline and accelerate the data center move process.Indeed,several parameters,such as non-network overheads and downtime adjustment,may impact the live migration time and server downtime to a large extent.By virtualizing the network resources,the infrastructure as a service(IaaS)can be used dynamically to allocate the bandwidth to services and monitor the network flow routing.Due to the large amount of filthy retransmission,existing live migration systems still suffer from extensive downtime and significant performance degradation in crossdata-center situations.This study aims to minimize the energy consumption by restricting the VMs migration and switching off the guests depending on a threshold,thereby boosting the residual network bandwidth in the data center with a minimal breach of the service level agreement(SLA).In this research,we analyzed and evaluated the findings observed through simulating different parameters,like availability,downtime,and outage of VMs in data center processes.This new paradigm is composed of two forms of detection strategies in the live migration approach from the source host to the destination source machine.
基金supported by the National Natural Science Foundation of China(32050410304,32002140,31822052,91431572381)National Thousand Youth Talents Plan to Y.J。
文摘Understanding how evolutionary pressures related to climate change have shaped the current genetic background of domestic animals is a fundamental pursuit of biology. Here, we generated wholegenome sequencing data from native goat populations in Iraq and Pakistan. Combined with previously published data on modern, ancient(Late Neolithic to Medieval periods), and wild Capra species worldwide, we explored the genetic population structure, ancestry components, and signatures of natural positive selection in native goat populations in Southwest Asia(SWA). Results revealed that the genetic structure of SWA goats was deeply influenced by gene flow from the eastern Mediterranean during the Chalcolithic period, which may reflect adaptation to gradual warming and aridity in the region. Furthermore, comparative genomic analysis revealed adaptive introgression of the KITLG locus from the Nubian ibex(C. nubiana) into African and SWA goats. The frequency of the selected allele at this locus was significantly higher among goat populations located near northeastern Africa. These results provide new insights into the genetic composition and history of goat populations in the SWA region.
基金supported by the Center for Cyber-Physical Systems,Khalifa University,under Grant 8474000137-RC1-C2PS-T5.
文摘The software engineering field has long focused on creating high-quality software despite limited resources.Detecting defects before the testing stage of software development can enable quality assurance engineers to con-centrate on problematic modules rather than all the modules.This approach can enhance the quality of the final product while lowering development costs.Identifying defective modules early on can allow for early corrections and ensure the timely delivery of a high-quality product that satisfies customers and instills greater confidence in the development team.This process is known as software defect prediction,and it can improve end-product quality while reducing the cost of testing and maintenance.This study proposes a software defect prediction system that utilizes data fusion,feature selection,and ensemble machine learning fusion techniques.A novel filter-based metric selection technique is proposed in the framework to select the optimum features.A three-step nested approach is presented for predicting defective modules to achieve high accuracy.In the first step,three supervised machine learning techniques,including Decision Tree,Support Vector Machines,and Naïve Bayes,are used to detect faulty modules.The second step involves integrating the predictive accuracy of these classification techniques through three ensemble machine-learning methods:Bagging,Voting,and Stacking.Finally,in the third step,a fuzzy logic technique is employed to integrate the predictive accuracy of the ensemble machine learning techniques.The experiments are performed on a fused software defect dataset to ensure that the developed fused ensemble model can perform effectively on diverse datasets.Five NASA datasets are integrated to create the fused dataset:MW1,PC1,PC3,PC4,and CM1.According to the results,the proposed system exhibited superior performance to other advanced techniques for predicting software defects,achieving a remarkable accuracy rate of 92.08%.
文摘The Far North Region of Cameroon is home to a great diversity of bird species, which unfortunately remains very little explored. This work was initiated to establish an inventory of birds and the factors affecting their diversity and distribution for sustainable management in the Kalfou Forest Reserve (KFR) and its periphery. Two methods were used for sampling, linear strip transects from which direct counts and indirect observations were made and the mist netting to complement the first. In total, 2525 birds were observed, including 149 species, belonging to 20 orders and 55 families. Accipitridae had the greatest number of species (11). The species richness was greater in the KFR (117 species) compared to the periphery (95 species). The specific richness was higher in wooded savannah compared to other habitats. Shannon index was significantly higher in the KFR (3.99) compared to that obtained in the periphery (3.80). The value of the Simpson index was higher on the outskirts of the KFR than on the periphery. The indices of species diversity were greater in the wooded savannah compared to other vegetation types. The seasons had no influence on bird diversity. Among the human activities encountered, the pressure indices were more important for grazing (7.3 contacts/km). Human activities have resulted in a significant decrease in specific richness. Six endangered species were encountered, four belonging to the Accipitridae family. The greater bird diversity in the reserve compared to the periphery shows that protected areas are a long-term solution for biodiversity conservation.
文摘Artificial intelligence(AI)is expanding its roots in medical diagnostics.Various acute and chronic diseases can be identified accurately at the initial level by using AI methods to prevent the progression of health complications.Kidney diseases are producing a high impact on global health and medical practitioners are suggested that the diagnosis at earlier stages is one of the foremost approaches to avert chronic kidney disease and renal failure.High blood pressure,diabetes mellitus,and glomerulonephritis are the root causes of kidney disease.Therefore,the present study is proposed a set of multiple techniques such as simulation,modeling,and optimization of intelligent kidney disease prediction(SMOIKD)which is based on computational intelligence approaches.Initially,seven parameters were used for the fuzzy logic system(FLS),and then twenty-five different attributes of the kidney dataset were used for the artificial neural network(ANN)and deep extreme machine learning(DEML).The expert system was proposed with the assistance of medical experts.For the quick and accurate evaluation of the proposed system,Matlab version 2019 was used.The proposed SMOIKD-FLSANN-DEML expert system has shown 94.16%accuracy.Hence this study concluded that SMOIKD-FLS-ANN-DEML system is effective to accurately diagnose kidney disease at initial levels.
基金supported by the MSIT(Ministry of Science and ICT),Korea under the ITRC(Information Technology Research Center)support program(IITP-2020-2018-0-01426)supervised by IITP(Institute for Information and Communication Technology Planning&Evaluation)+1 种基金in part by the National Research Foundation(NRF)funded by the Korea government(MSIT)(No.2019R1F1A1059125).
文摘Internet of Things(IoT)network used for industrial management is vulnerable to different security threats due to its unstructured deployment,and dynamic communication behavior.In literature various mechanisms addressed the security issue of Industrial IoT networks,but proper maintenance of the performance reliability is among the common challenges.In this paper,we proposed an intelligent mutual authentication scheme leveraging authentication aware node(AAN)and base station(BS)to identify routing attacks in Industrial IoT networks.The AAN and BS uses the communication parameter such as a route request(RREQ),node-ID,received signal strength(RSS),and round-trip time(RTT)information to identify malicious devices and routes in the deployed network.The feasibility of the proposed model is validated in the simulation environment,where OMNeT++was used as a simulation tool.We compare the results of the proposed model with existing field-proven schemes in terms of routing attacks detection,communication cost,latency,computational cost,and throughput.The results show that our proposed scheme surpasses the previous schemes regarding these performance parameters with the attack detection rate of 97.7%.
文摘The internet,particularly online social networking platforms have revolutionized the way extremist groups are influencing and radicalizing individuals.Recent research reveals that the process initiates by exposing vast audiences to extremist content and then migrating potential victims to confined platforms for intensive radicalization.Consequently,social networks have evolved as a persuasive tool for extremism aiding as recruitment platform and psychological warfare.Thus,recognizing potential radical text or material is vital to restrict the circulation of the extremist chronicle.The aim of this research work is to identify radical text in social media.Our contributions are as follows:(i)A new dataset to be employed in radicalization detection;(ii)In depth analysis of new and previous datasets so that the variation in extremist group narrative could be identified;(iii)An approach to train classifier employing religious features along with radical features to detect radicalization;(iv)Observing the use of violent and bad words in radical,neutral and random groups by employing violent,terrorism and bad words dictionaries.Our research results clearly indicate that incorporating religious text in model training improves the accuracy,precision,recall,and F1-score of the classifiers.Secondly a variation in extremist narrative has been observed implying that usage of new dataset can have substantial effect on classifier performance.In addition to this,violence and bad words are creating a differentiating factor between radical and random users but for neutral(anti-ISIS)group it needs further investigation.
文摘Lysine Lipoylation is a protective and conserved Post Translational Modification(PTM)in proteomics research like prokaryotes and eukaryotes.It is connected with many biological processes and closely linked with many metabolic diseases.To develop a perfect and accurate classification model for identifying lipoylation sites at the protein level,the computational methods and several other factors play a key role in this purpose.Usually,most of the techniques and different traditional experimental models have a very high cost.They are time-consuming;so,it is required to construct a predictor model to extract lysine lipoylation sites.This study proposes a model that could predict lysine lipoylation sites with the help of a classification method known as Artificial Neural Network(ANN).The ANN algorithm deals with the noise problem and imbalance classification in lipoylation sites dataset samples.As the result shows in ten-fold cross-validation,a brilliant performance is achieved through the predictor model with an accuracy of 99.88%,and also achieved 0.9976 as the highest value of MCC.So,the predictor model is a very useful and helpful tool for lipoylation sites prediction.Some of the residues around lysine lipoylation sites play a vital part in prediction,as demonstrated during feature analysis.The wonderful results reported through the evaluation and prediction of this model can provide an informative and relative explanation for lipoylation and its molecular mechanisms.
文摘Cloud computing is becoming popular technology due to its functional properties and variety of customer-oriented services over the Internet.The design of reliable and high-quality cloud applications requires a strong Quality of Service QoS parameter metric.In a hyperconverged cloud ecosystem environment,building high-reliability cloud applications is a challenging job.The selection of cloud services is based on the QoS parameters that play essential roles in optimizing and improving cloud rankings.The emergence of cloud computing is significantly reshaping the digital ecosystem,and the numerous services offered by cloud service providers are playing a vital role in this transformation.Hyperconverged software-based unified utilities combine storage virtualization,compute virtualization,and network virtualization.The availability of the latter has also raised the demand for QoS.Due to the diversity of services,the respective quality parameters are also in abundance and need a carefully designed mechanism to compare and identify the critical,common,and impactful parameters.It is also necessary to reconsider the market needs in terms of service requirements and the QoS provided by various CSPs.This research provides a machine learning-based mechanism to monitor the QoS in a hyperconverged environment with three core service parameters:service quality,downtime of servers,and outage of cloud services.
文摘Nowadays rapidly increasing technology is mobile phone technology in telecommunication sector. This mobile device technology has great effect on everyone’s life. This technology has reduced the burden of people in their daily life. To manage the rising demand for such mobile devices, numerous operating systems came in the market as a platform upon which modern application can be produced. As a result, numbers of platforms and essential depository describe these platforms;customers may or may not be aware of these platforms that are appropriate for their needs. In order to solve this issue, we examine the most famous mobile phone operating systems to decide which operating system is most suitable for developers, business applications as well as casual use. In this paper we make assessment on the popular operating systems of mobile devices available in the business market, and on behalf of such assessment we distinguish that operating system OS is much useful of its particular characteristics compared with other systems.
基金funded by the Ministry of Science,ICT CMC,202327(2019M3F2A1073387)this work was supported by the Institute for Information&communications Technology Promotion(IITP)(NO.2022-0-00980,Cooperative Intelligence Framework of Scene Perception for Autonomous IoT Device).
文摘The requirement for high-quality seafood is a global challenge in today’s world due to climate change and natural resource limitations.Internet of Things(IoT)based Modern fish farming systems can significantly optimize seafood production by minimizing resource utilization and improving healthy fish production.This objective requires intensive monitoring,prediction,and control by optimizing leading factors that impact fish growth,including temperature,the potential of hydrogen(pH),water level,and feeding rate.This paper proposes the IoT based predictive optimization approach for efficient control and energy utilization in smart fish farming.The proposed fish farm control mechanism has a predictive optimization to deal with water quality control and efficient energy consumption problems.Fish farm indoor and outdoor values are applied to predict the water quality parameters,whereas a novel objective function is proposed to achieve an optimal fish growth environment based on predicted parameters.Fuzzy logic control is utilized to calculate control parameters for IoT actuators based on predictive optimal water quality parameters by minimizing energy consumption.To evaluate the efficiency of the proposed system,the overall approach has been deployed to the fish tank as a case study,and a number of experiments have been carried out.The results show that the predictive optimization module allowed the water quality parameters to be maintained at the optimal level with nearly 30%of energy efficiency at the maximum actuator control rate compared with other control levels.
基金This work was supported in part by the RHB-UKM Endowment Fund through Dana Endowmen RHB-UKM under Grant RHB-UKM-2021-001in part by the Universiti Kebangsaan Malaysia through the Dana Padanan Kolaborasi under Grant DPK-2021-012.
文摘A well-managed financial market of stocks,commodities,derivatives,and bonds is crucial to a country’s economic growth.It provides confidence to investors,which encourages the inflow of cash to ensure good market liquidity.However,there will always be a group of traders that aims to manipulate market pricing to negatively influence stock values in their favor.These illegal trading activities are surely prohibited according to the rules and regulations of every country’s stockmarket.It is the role of regulators to detect and prevent any manipulation cases in order to provide a trading platform that is fair and efficient.However,the complexity of manipulation cases has increased significantly,coupled with high trading volumes,which makes the manual observations of such cases by human operators no longer feasible.As a result,many intelligent systems have been developed by researchers all over the world to automatically detect various types of manipulation cases.Therefore,this review paper aims to comprehensively discuss the state-of-theart methods that have been developed to detect and recognize stock market manipulation cases.It also provides a concise definition of manipulation taxonomy,including manipulation types and categories,as well as some of the output of early experimental research.In summary,this paper provides a thorough review of the automated methods for detecting stock market manipulation cases.
文摘The query optimizer uses cost-based optimization to create an execution plan with the least cost,which also consumes the least amount of resources.The challenge of query optimization for relational database systems is a combinatorial optimization problem,which renders exhaustive search impossible as query sizes rise.Increases in CPU performance have surpassed main memory,and disk access speeds in recent decades,allowing data compression to be used—strategies for improving database performance systems.For performance enhancement,compression and query optimization are the two most factors.Compression reduces the volume of data,whereas query optimization minimizes execution time.Compressing the database reduces memory requirement,data takes less time to load into memory,fewer buffer missing occur,and the size of intermediate results is more diminutive.This paper performed query optimization on the graph database in a cloud dew environment by considering,which requires less time to execute a query.The factors compression and query optimization improve the performance of the databases.This research compares the performance of MySQL and Neo4j databases in terms of memory usage and execution time running on cloud dew servers.
文摘Kubernetes,a container orchestrator for cloud-deployed applications,allows the application provider to scale automatically to match thefluctuating intensity of processing demand.Container cluster technology is used to encapsulate,isolate,and deploy applications,addressing the issue of low system reliability due to interlocking failures.Cloud-based platforms usually entail users define application resource supplies for eco container virtualization.There is a constant problem of over-service in data centers for cloud service providers.Higher operating costs and incompetent resource utilization can occur in a waste of resources.Kubernetes revolutionized the orchestration of the container in the cloud-native age.It can adaptively manage resources and schedule containers,which provide real-time status of the cluster at runtime without the user’s contribution.Kubernetes clusters face unpredictable traffic,and the cluster performs manual expansion configuration by the controller.Due to operational delays,the system will become unstable,and the service will be unavailable.This work proposed an RBACS that vigorously amended the distribution of containers operating in the entire Kubernetes cluster.RBACS allocation pattern is analyzed with the Kubernetes VPA.To estimate the overall cost of RBACS,we use several scientific benchmarks comparing the accomplishment of container to remote node migration and on-site relocation.The experiments ran on the simulations to show the method’s effectiveness yielded high precision in the real-time deployment of resources in eco containers.Compared to the default baseline,Kubernetes results in much fewer dropped requests with only slightly more supplied resources.
基金The authors acknowledge the support from the Deanship of Scientific Research,Najran University.Kingdom of Saudi Arabia,for funding this work under theResearch Collaboration funding program grant code number(NU/RC/SERC/11/7).
文摘Leukemia is a kind of blood cancer that damages the cells in the blood and bone marrow of the human body.It produces cancerous blood cells that disturb the human’s immune system and significantly affect bone marrow’s production ability to effectively create different types of blood cells like red blood cells(RBCs)and white blood cells(WBC),and platelets.Leukemia can be diagnosed manually by taking a complete blood count test of the patient’s blood,from which medical professionals can investigate the signs of leukemia cells.Furthermore,two other methods,microscopic inspection of blood smears and bone marrow aspiration,are also utilized while examining the patient for leukemia.However,all these methods are labor-intensive,slow,inaccurate,and require a lot of human experience and dedication.Different authors have proposed automated detection systems for leukemia diagnosis to overcome these limitations.They have deployed digital image processing and machine learning algorithms to classify the cells into normal and blast cells.However,these systems are more efficient,reliable,and fast than previous manual diagnosing methods.However,more work is required to classify leukemia-affected cells due to the complex characteristics of blood images and leukemia cells having much intra-class variability and inter-class similarity.In this paper,we have proposed a robust automated system to diagnose leukemia and its sub-types.We have classified ALL into its sub-types based on FAB classification,i.e.,L1,L2,and L3 types with better performance.We have achieved 96.06%accuracy for subtypes classification,which is better when compared with the state-of-the-art methodologies.
基金This research was supported by Suranaree University of Technology,Thailand,Grant Number:BRO7-709-62-12-03.
文摘The deep learning advancements have greatly improved the performance of speech recognition systems,and most recent systems are based on the Recurrent Neural Network(RNN).Overall,the RNN works fine with the small sequence data,but suffers from the gradient vanishing problem in case of large sequence.The transformer networks have neutralized this issue and have shown state-of-the-art results on sequential or speech-related data.Generally,in speech recognition,the input audio is converted into an image using Mel-spectrogram to illustrate frequencies and intensities.The image is classified by the machine learning mechanism to generate a classification transcript.However,the audio frequency in the image has low resolution and causing inaccurate predictions.This paper presents a novel end-to-end binary view transformer-based architecture for speech recognition to cope with the frequency resolution problem.Firstly,the input audio signal is transformed into a 2D image using Mel-spectrogram.Secondly,the modified universal transformers utilize the multi-head attention to derive contextual information and derive different speech-related features.Moreover,a feedforward neural network is also deployed for classification.The proposed system has generated robust results on Google’s speech command dataset with an accuracy of 95.16%and with minimal loss.The binary-view transformer eradicates the eventuality of the over-fitting problem by deploying a multiview mechanism to diversify the input data,and multi-head attention captures multiple contexts from the data’s feature map.
基金This research was supported by Suranaree University of Technology,Thailand,Grant Number:BRO7-709-62-12-03.
文摘Astrocytoma IV or glioblastoma is one of the fatal and dangerous types of brain tumors.Early detection of brain tumor increases the survival rate and helps in reducing the fatality rate.Various imaging modalities have been used for diagnosing by expert radiologists,and Medical Resonance Image(MRI)is considered a better option for detecting brain tumors as MRI is a non-invasive technique and provides better visualization of the brain region.One of the challenging issues is to identify the tumorous region from the MRI scans correctly.Manual segmentation is performed by medical experts,which is a time-consuming task and got chances of errors.To overcome this issue,automatic segmentation is performed for quick and accurate results.The proposed approach is to capture inter-slice information and reduce the outliers.Deep learning-based brain tumor segmentation techniques proved best among available segmentation techniques.However,deep learning may miss some preliminary info while using MRI images during segmentation.As MRI volumes are volumetric,3D U-Net-based models are used but complex.Combinations of multiple 2D U-Net predictions in axial,sagittal,and coronal views help to capture inter-slice information.This approach may reduce the system complexity.Moreover,the Conditional Random Fields(CRF)reduce the predictions’false positives and improve the segmentation results.This model is applied to Brain Tumor Segmentation(BraTS)2019 dataset,and cross-validation is performed to check the accuracy of results.The proposed approach achieves Dice Similarity Score(DSC)of 0.77 on Enhancing Tumor(ET),0.90 on Whole Tumor(WT),and 0.84 on Tumor Core(TC)with reduced Hausdorff Distance(HD)of 3.05 on ET,5.12 on WT and 3.89 on TC.
文摘The smart vehicles are one of critical enablers for automated services in smart cities to provide intelligent transportation means without human intervention.In order to fulfil requirements,Vehicle-to-Anything(V2X)communications aims to manage massive connectivity and high traffic load on base stations and extend the range over multiple hops in 5G networks.However,V2X networking faces several challenges from dynamic topology caused by high velocity of nodes and routing overhead that degrades the network performance and increases energy consumption.The existing routing scheme for V2X networking lacks energy efficiency and scalability for high velocity nodes with dense distribution.In order to handle the challenges,this article proposes a scalable and energy-efficient routing scheme called Dynamic proactive reactive routing for 5G(DPR5)for high mobility speed and dense environment.As compared to existing schemes it uses a single routing table and efficiently minimizes the energy consumption in dense environment,improves node‘s lifetime upto 42%,and optimizes network performance by reducing the packet loss ratio upto 46%in a high velocity dense environment.
文摘Security is a significant issue for everyone due to new and creative ways to commit cybercrime.The Closed-Circuit Television(CCTV)systems are being installed in offices,houses,shopping malls,and on streets to protect lives.Operators monitor CCTV;however,it is difficult for a single person to monitor the actions of multiple people at one time.Consequently,there is a dire need for an automated monitoring system that detects a person with ammunition or any other harmful material Based on our research and findings of this study,we have designed a new Intelligent Ammunition Detection and Classification(IADC)system using Convolutional Neural Network(CNN).The proposed system is designed to identify persons carrying weapons and ammunition using CCTV cameras.When weapons are identified,the cameras sound an alarm.In the proposed IADC system,CNN was used to detect firearms and ammunition.The CNN model which is a Deep Learning technique consists of neural networks,most commonly applied to analyzing visual imagery has gained popularity for unstructured(images,videos)data classification.Additionally,this system generates an early warning through detection of ammunition before conditions become critical.Hence the faster and earlier the prediction,the lower the response time,loses and potential victims.The proposed IADC system provides better results than earlier published models like VGGNet,OverFeat-1,OverFeat-2,and OverFeat-3.
文摘Cloud computing is an emerging domain that is capturing global users from all walks of life—the corporate sector,government sector,and social arena as well.Various cloud providers have offered multiple services and facilities to this audience and the number of providers is increasing very swiftly.This enormous pace is generating the requirement of a comprehensive ecosystem that shall provide a seamless and customized user environment not only to enhance the user experience but also to improve security,availability,accessibility,and latency.Emerging technology is providing robust solutions to many of our problems,the cloud platform is one of them.It is worth mentioning that these solutions are also amplifying the complexity and need of sustenance of these rapid solutions.As with cloud computing,new entrants as cloud service providers,resellers,tech-support,hardware manufacturers,and software developers appear on a daily basis.These actors playing their role in the growth and sustenance of the cloud ecosystem.Our objective is to use convergence for cloud services,software-defined networks,network function virtualization for infrastructure,cognition for pattern development,and knowledge repository.In order to gear up these processes,machine learning to induce intelligence to maintain ecosystem growth,to monitor performance,and to become able to make decisions for the sustenance of the ecosystem.Workloads may be programmed to“superficially”imitate most business applications and create large numbers using lightweight workload generators that merely stress the storage.In today’s current IT environment,when many enterprises use the cloud to service some of their application demands,a different performance testing technique that assesses more than the storage is necessary.Compute and storage are merged into a single building block with HCI(Hyper-converged infrastructure),resulting in a huge pool of compute and storage resources when clustered with other building blocks.The novelty of thiswork to design and test cloud storage using themeasurement of availability,downtime,and outage parameters.Results showed that the storage reliability in a hyper-converged system is above 92%.