As the extensive use of cloud computing raises questions about the security of any personal data stored there,cryptography is being used more frequently as a security tool to protect data confidentiality and privacy i...As the extensive use of cloud computing raises questions about the security of any personal data stored there,cryptography is being used more frequently as a security tool to protect data confidentiality and privacy in the cloud environment.A hypervisor is a virtualization software used in cloud hosting to divide and allocate resources on various pieces of hardware.The choice of hypervisor can significantly impact the performance of cryptographic operations in the cloud environment.An important issue that must be carefully examined is that no hypervisor is completely superior in terms of performance;Each hypervisor should be examined to meet specific needs.The main objective of this study is to provide accurate results to compare the performance of Hyper-V and Kernel-based Virtual Machine(KVM)while implementing different cryptographic algorithms to guide cloud service providers and end users in choosing the most suitable hypervisor for their cryptographic needs.This study evaluated the efficiency of two hypervisors,Hyper-V and KVM,in implementing six cryptographic algorithms:Rivest,Shamir,Adleman(RSA),Advanced Encryption Standard(AES),Triple Data Encryption Standard(TripleDES),Carlisle Adams and Stafford Tavares(CAST-128),BLOWFISH,and TwoFish.The study’s findings show that KVM outperforms Hyper-V,with 12.2%less Central Processing Unit(CPU)use and 12.95%less time overall for encryption and decryption operations with various file sizes.The study’s findings emphasize how crucial it is to pick a hypervisor that is appropriate for cryptographic needs in a cloud environment,which could assist both cloud service providers and end users.Future research may focus more on how various hypervisors perform while handling cryptographic workloads.展开更多
Cloud computing provides a diverse and adaptable resource pool over the internet,allowing users to tap into various resources as needed.It has been seen as a robust solution to relevant challenges.A significant delay ...Cloud computing provides a diverse and adaptable resource pool over the internet,allowing users to tap into various resources as needed.It has been seen as a robust solution to relevant challenges.A significant delay can hamper the performance of IoT-enabled cloud platforms.However,efficient task scheduling can lower the cloud infrastructure’s energy consumption,thus maximizing the service provider’s revenue by decreasing user job processing times.The proposed Modified Chimp-Whale Optimization Algorithm called Modified Chimp-Whale Optimization Algorithm(MCWOA),combines elements of the Chimp Optimization Algorithm(COA)and the Whale Optimization Algorithm(WOA).To enhance MCWOA’s identification precision,the Sobol sequence is used in the population initialization phase,ensuring an even distribution of the population across the solution space.Moreover,the traditional MCWOA’s local search capabilities are augmented by incorporating the whale optimization algorithm’s bubble-net hunting and random search mechanisms into MCWOA’s position-updating process.This study demonstrates the effectiveness of the proposed approach using a two-story rigid frame and a simply supported beam model.Simulated outcomes reveal that the new method outperforms the original MCWOA,especially in multi-damage detection scenarios.MCWOA excels in avoiding false positives and enhancing computational speed,making it an optimal choice for structural damage detection.The efficiency of the proposed MCWOA is assessed against metrics such as energy usage,computational expense,task duration,and delay.The simulated data indicates that the new MCWOA outpaces other methods across all metrics.The study also references the Whale Optimization Algorithm(WOA),Chimp Algorithm(CA),Ant Lion Optimizer(ALO),Genetic Algorithm(GA)and Grey Wolf Optimizer(GWO).展开更多
Task scheduling plays a key role in effectively managing and allocating computing resources to meet various computing tasks in a cloud computing environment.Short execution time and low load imbalance may be the chall...Task scheduling plays a key role in effectively managing and allocating computing resources to meet various computing tasks in a cloud computing environment.Short execution time and low load imbalance may be the challenges for some algorithms in resource scheduling scenarios.In this work,the Hierarchical Particle Swarm Optimization-Evolutionary Artificial Bee Colony Algorithm(HPSO-EABC)has been proposed,which hybrids our presented Evolutionary Artificial Bee Colony(EABC),and Hierarchical Particle Swarm Optimization(HPSO)algorithm.The HPSO-EABC algorithm incorporates both the advantages of the HPSO and the EABC algorithm.Comprehensive testing including evaluations of algorithm convergence speed,resource execution time,load balancing,and operational costs has been done.The results indicate that the EABC algorithm exhibits greater parallelism compared to the Artificial Bee Colony algorithm.Compared with the Particle Swarm Optimization algorithm,the HPSO algorithmnot only improves the global search capability but also effectively mitigates getting stuck in local optima.As a result,the hybrid HPSO-EABC algorithm demonstrates significant improvements in terms of stability and convergence speed.Moreover,it exhibits enhanced resource scheduling performance in both homogeneous and heterogeneous environments,effectively reducing execution time and cost,which also is verified by the ablation experimental.展开更多
Security issues in cloud networks and edge computing have become very common. This research focuses on analyzing such issues and developing the best solutions. A detailed literature review has been conducted in this r...Security issues in cloud networks and edge computing have become very common. This research focuses on analyzing such issues and developing the best solutions. A detailed literature review has been conducted in this regard. The findings have shown that many challenges are linked to edge computing, such as privacy concerns, security breaches, high costs, low efficiency, etc. Therefore, there is a need to implement proper security measures to overcome these issues. Using emerging trends, like machine learning, encryption, artificial intelligence, real-time monitoring, etc., can help mitigate security issues. They can also develop a secure and safe future in cloud computing. It was concluded that the security implications of edge computing can easily be covered with the help of new technologies and techniques.展开更多
With the rapid development of the Internet of Things(IoT),there are several challenges pertaining to security in IoT applications.Compared with the characteristics of the traditional Internet,the IoT has many problems...With the rapid development of the Internet of Things(IoT),there are several challenges pertaining to security in IoT applications.Compared with the characteristics of the traditional Internet,the IoT has many problems,such as large assets,complex and diverse structures,and lack of computing resources.Traditional network intrusion detection systems cannot meet the security needs of IoT applications.In view of this situation,this study applies cloud computing and machine learning to the intrusion detection system of IoT to improve detection performance.Usually,traditional intrusion detection algorithms require considerable time for training,and these intrusion detection algorithms are not suitable for cloud computing due to the limited computing power and storage capacity of cloud nodes;therefore,it is necessary to study intrusion detection algorithms with low weights,short training time,and high detection accuracy for deployment and application on cloud nodes.An appropriate classification algorithm is a primary factor for deploying cloud computing intrusion prevention systems and a prerequisite for the system to respond to intrusion and reduce intrusion threats.This paper discusses the problems related to IoT intrusion prevention in cloud computing environments.Based on the analysis of cloud computing security threats,this study extensively explores IoT intrusion detection,cloud node monitoring,and intrusion response in cloud computing environments by using cloud computing,an improved extreme learning machine,and other methods.We use the Multi-Feature Extraction Extreme Learning Machine(MFE-ELM)algorithm for cloud computing,which adds a multi-feature extraction process to cloud servers,and use the deployed MFE-ELM algorithm on cloud nodes to detect and discover network intrusions to cloud nodes.In our simulation experiments,a classical dataset for intrusion detection is selected as a test,and test steps such as data preprocessing,feature engineering,model training,and result analysis are performed.The experimental results show that the proposed algorithm can effectively detect and identify most network data packets with good model performance and achieve efficient intrusion detection for heterogeneous data of the IoT from cloud nodes.Furthermore,it can enable the cloud server to discover nodes with serious security threats in the cloud cluster in real time,so that further security protection measures can be taken to obtain the optimal intrusion response strategy for the cloud cluster.展开更多
Some of the significant new technologies researched in recent studies include BlockChain(BC),Software Defined Networking(SDN),and Smart Industrial Internet of Things(IIoT).All three technologies provide data integrity...Some of the significant new technologies researched in recent studies include BlockChain(BC),Software Defined Networking(SDN),and Smart Industrial Internet of Things(IIoT).All three technologies provide data integrity,confidentiality,and integrity in their respective use cases(especially in industrial fields).Additionally,cloud computing has been in use for several years now.Confidential information is exchanged with cloud infrastructure to provide clients with access to distant resources,such as computing and storage activities in the IIoT.There are also significant security risks,concerns,and difficulties associated with cloud computing.To address these challenges,we propose merging BC and SDN into a cloud computing platform for the IIoT.This paper introduces“DistB-SDCloud”,an architecture for enhanced cloud security for smart IIoT applications.The proposed architecture uses a distributed BC method to provide security,secrecy,privacy,and integrity while remaining flexible and scalable.Customers in the industrial sector benefit from the dispersed or decentralized,and efficient environment of BC.Additionally,we described an SDN method to improve the durability,stability,and load balancing of cloud infrastructure.The efficacy of our SDN and BC-based implementation was experimentally tested by using various parameters including throughput,packet analysis,response time,bandwidth,and latency analysis,as well as the monitoring of several attacks on the system itself.展开更多
To serve various tasks requested by various end devices with different requirements,end-edge-cloud(E2C)has attracted more and more attention from specialists in both academia and industry,by combining both benefits of...To serve various tasks requested by various end devices with different requirements,end-edge-cloud(E2C)has attracted more and more attention from specialists in both academia and industry,by combining both benefits of edge and cloud computing.But nowadays,E2C still suffers from low service quality and resource efficiency,due to the geographical distribution of edge resources and the high dynamic of network topology and user mobility.To address these issues,this paper focuses on task offloading,which makes decisions that which resources are allocated to tasks for their processing.This paper first formulates the problem into binary non-linear programming and then proposes a particle swarm optimization(PSO)-based algorithm to solve the problem.The proposed algorithm exploits an imbalance mutation operator and a task rescheduling approach to improve the performance of PSO.The proposed algorithm concerns the resource heterogeneity by correlating the probability that a computing node is decided to process a task with its capacity,by the imbalance mutation.The task rescheduling approach improves the acceptance ratio for a task offloading solution,by reassigning rejected tasks to computing nodes with available resources.Extensive simulated experiments are conducted.And the results show that the proposed offloading algorithm has an 8.93%–37.0%higher acceptance ratio than ten of the classical and up-to-date algorithms,and verify the effectiveness of the imbalanced mutation and the task rescheduling.展开更多
Star sensors are an important means of autonomous navigation and access to space information for satellites.They have been widely deployed in the aerospace field.To satisfy the requirements for high resolution,timelin...Star sensors are an important means of autonomous navigation and access to space information for satellites.They have been widely deployed in the aerospace field.To satisfy the requirements for high resolution,timeliness,and confidentiality of star images,we propose an edge computing algorithm based on the star sensor cloud.Multiple sensors cooperate with each other to forma sensor cloud,which in turn extends the performance of a single sensor.The research on the data obtained by the star sensor has very important research and application values.First,a star point extraction model is proposed based on the fuzzy set model by analyzing the star image composition,which can reduce the amount of data computation.Then,a mappingmodel between content and space is constructed to achieve low-rank image representation and efficient computation.Finally,the data collected by the wireless sensor is delivered to the edge server,and a differentmethod is used to achieve privacy protection.Only a small amount of core data is stored in edge servers and local servers,and other data is transmitted to the cloud.Experiments show that the proposed algorithm can effectively reduce the cost of communication and storage,and has strong privacy.展开更多
In recent years,statistics have indicated that the number of patients with malignant brain tumors has increased sharply.However,most surgeons still perform surgical training using the traditional autopsy and prosthesi...In recent years,statistics have indicated that the number of patients with malignant brain tumors has increased sharply.However,most surgeons still perform surgical training using the traditional autopsy and prosthesis model,which encounters many problems,such as insufficient corpse resources,low efficiency,and high cost.With the advent of the 5G era,a wide range of Industrial Internet of Things(IIOT)applications have been developed.Virtual Reality(VR)and Augmented Reality(AR)technologies that emerged with 5G are developing rapidly for intelligent medical training.To address the challenges encountered during neurosurgery training,and combining with cloud computing,in this paper,a highly immersive AR-based brain tumor neurosurgery remote collaborative virtual surgery training system is developed,in which a VR simulator is embedded.The system enables real-time remote surgery training interaction through 5G transmission.Six experts and 18 novices were invited to participate in the experiment to verify the system.Subsequently,the two simulators were evaluated using face and construction validation methods.The results obtained by training the novices 50 times were further analyzed using the Learning Curve-Cumulative Sum(LC-CUSUM)evaluation method to validate the effectiveness of the two simulators.The results of the face and content validation demonstrated that the AR simulator in the system was superior to the VR simulator in terms of vision and scene authenticity,and had a better effect on the improvement of surgical skills.Moreover,the surgical training scheme proposed in this paper is effective,and the remote collaborative training effect of the system is ideal.展开更多
Due to its decentralized,tamper-proof,and trust-free characteristics,blockchain is used in the Internet of Things(IoT)to guarantee the reliability of data.However,some technical flaws in blockchain itself prevent the ...Due to its decentralized,tamper-proof,and trust-free characteristics,blockchain is used in the Internet of Things(IoT)to guarantee the reliability of data.However,some technical flaws in blockchain itself prevent the development of these applications,such as the issue with linearly growing storage capacity of blockchain systems.On the other hand,there is a lack of storage resources for sensor devices in IoT,and numerous sensor devices will generate massive data at ultra-high speed,which makes the storage problem of the IoT enabled by blockchain more prominent.There are various solutions to reduce the storage burden by modifying the blockchain’s storage policy,but most of them do not consider the willingness of peers.In attempt to make the blockchain more compatible with the IoT,this paper proposes a storage optimization scheme that revisits the system data storage problem from amore practically oriented standpoint.Peers will only store transactional data that they are directly involved in.In addition,a transaction verification model is developed to enable peers to undertake transaction verification with the aid of cloud computing,and an incentive mechanism is premised on the storage optimization scheme to assure data integrity.The results of the simulation experiments demonstrate the proposed scheme’s advantage in terms of storage and throughput.展开更多
The solution strategy of the heuristic algorithm is pre-set and has good performance in the conventional cloud resource scheduling process.However,for complex and dynamic cloud service scheduling tasks,due to the diff...The solution strategy of the heuristic algorithm is pre-set and has good performance in the conventional cloud resource scheduling process.However,for complex and dynamic cloud service scheduling tasks,due to the difference in service attributes,the solution efficiency of a single strategy is low for such problems.In this paper,we presents a hyper-heuristic algorithm based on reinforcement learning(HHRL)to optimize the completion time of the task sequence.Firstly,In the reward table setting stage of HHRL,we introduce population diversity and integrate maximum time to comprehensively deter-mine the task scheduling and the selection of low-level heuristic strategies.Secondly,a task computational complexity estimation method integrated with linear regression is proposed to influence task scheduling priorities.Besides,we propose a high-quality candidate solution migration method to ensure the continuity and diversity of the solving process.Compared with HHSA,ACO,GA,F-PSO,etc,HHRL can quickly obtain task complexity,select appropriate heuristic strategies for task scheduling,search for the the best makspan and have stronger disturbance detection ability for population diversity.展开更多
In the present scenario of rapid growth in cloud computing models,several companies and users started to share their data on cloud servers.However,when the model is not completely trusted,the data owners face several ...In the present scenario of rapid growth in cloud computing models,several companies and users started to share their data on cloud servers.However,when the model is not completely trusted,the data owners face several security-related problems,such as user privacy breaches,data disclosure,data corruption,and so on,during the process of data outsourcing.For addressing and handling the security-related issues on Cloud,several models were proposed.With that concern,this paper develops a Privacy-Preserved Data Security Approach(PP-DSA)to provide the data security and data integrity for the out-sourcing data in Cloud Environment.Privacy preservation is ensured in this work with the Efficient Authentication Technique(EAT)using the Group Signature method that is applied with Third-Party Auditor(TPA).The role of the auditor is to secure the data and guarantee shared data integrity.Additionally,the Cloud Service Provider(CSP)and Data User(DU)can also be the attackers that are to be handled with the EAT.Here,the major objective of the work is to enhance cloud security and thereby,increase Quality of Service(QoS).The results are evaluated based on the model effectiveness,security,and reliability and show that the proposed model provides better results than existing works.展开更多
Cloud computing(CC)is developing as a powerful and flexible computational structure for providing ubiquitous service to users.It receives interrelated software and hardware resources in an integrated manner distinct f...Cloud computing(CC)is developing as a powerful and flexible computational structure for providing ubiquitous service to users.It receives interrelated software and hardware resources in an integrated manner distinct from the classical computational environment.The variation of software and hardware resources were combined and composed as a resource pool.The software no more resided in the single hardware environment,it can be executed on the schedule of resource pools to optimize resource consumption.Optimizing energy consumption in CC environments is the question that allows utilizing several energy conservation approaches for effective resource allocation.This study introduces a Battle Royale Optimization-based Resource Scheduling Scheme for Cloud Computing Environment(BRORSS-CCE)technique.The presented BRORSS-CCE technique majorly schedules the available resources for maximum utilization and effectual makespan.In the BRORSS-CCE technique,the BRO is a population-based algorithm where all the individuals are denoted by a soldier/player who likes to go towards the optimal place and ultimate survival.The BRORSS-CCE technique can be employed to balance the load,distribute resources based on demand and assure services to all requests.The experimental validation of the BRORSS-CCE technique is tested under distinct aspects.The experimental outcomes indicated the enhancements of the BRORSS-CCE technique over other models.展开更多
In many IIoT architectures,various devices connect to the edge cloud via gateway systems.For data processing,numerous data are delivered to the edge cloud.Delivering data to an appropriate edge cloud is critical to im...In many IIoT architectures,various devices connect to the edge cloud via gateway systems.For data processing,numerous data are delivered to the edge cloud.Delivering data to an appropriate edge cloud is critical to improve IIoT service efficiency.There are two types of costs for this kind of IoT network:a communication cost and a computing cost.For service efficiency,the communication cost of data transmission should be minimized,and the computing cost in the edge cloud should be also minimized.Therefore,in this paper,the communication cost for data transmission is defined as the delay factor,and the computing cost in the edge cloud is defined as the waiting time of the computing intensity.The proposed method selects an edge cloud that minimizes the total cost of the communication and computing costs.That is,a device chooses a routing path to the selected edge cloud based on the costs.The proposed method controls the data flows in a mesh-structured network and appropriately distributes the data processing load.The performance of the proposed method is validated through extensive computer simulation.When the transition probability from good to bad is 0.3 and the transition probability from bad to good is 0.7 in wireless and edge cloud states,the proposed method reduced both the average delay and the service pause counts to about 25%of the existing method.展开更多
Serverless computing is a promising paradigm in cloud computing that greatly simplifies cloud programming.With serverless computing,developers only provide function code to serverless platform,and these functions are ...Serverless computing is a promising paradigm in cloud computing that greatly simplifies cloud programming.With serverless computing,developers only provide function code to serverless platform,and these functions are invoked by its driven events.Nonetheless,security threats in serverless computing such as vulnerability-based security threats have become the pain point hindering its wide adoption.The ideas in proactive defense such as redundancy,diversity and dynamic provide promising approaches to protect against cyberattacks.However,these security technologies are mostly applied to serverless platform based on“stacked”mode,as they are designed independent with serverless computing.The lack of security consideration in the initial design makes it especially challenging to achieve the all life cycle protection for serverless application with limited cost.In this paper,we present ATSSC,a proactive defense enabled attack tolerant serverless platform.ATSSC integrates the characteristic of redundancy,diversity and dynamic into serverless seamless to achieve high-level security and efficiency.Specifically,ATSSC constructs multiple diverse function replicas to process the driven events and performs cross-validation to verify the results.In order to create diverse function replicas,both software diversity and environment diversity are adopted.Furthermore,a dynamic function refresh strategy is proposed to keep the clean state of serverless functions.We implement ATSSC based on Kubernetes and Knative.Analysis and experimental results demonstrate that ATSSC can effectively protect serverless computing against cyberattacks with acceptable costs.展开更多
This paper presents a novel approach to proxy blind signatures in the realm of quantum circuits,aiming to enhance security while safeguarding sensitive information.The main objective of this research is to introduce a...This paper presents a novel approach to proxy blind signatures in the realm of quantum circuits,aiming to enhance security while safeguarding sensitive information.The main objective of this research is to introduce a quantum proxy blind signature(QPBS)protocol that utilizes quantum logical gates and quantum measurement techniques.The QPBS protocol is constructed by the initial phase,proximal blinding message phase,remote authorization and signature phase,remote validation,and de-blinding phase.This innovative design ensures a secure mechanism for signing documents without revealing the content to the proxy signer,providing practical security authentication in a quantum environment under the assumption that the CNOT gates are securely implemented.Unlike existing approaches,our proposed QPBS protocol eliminates the need for quantum entanglement preparation,thus simplifying the implementation process.To assess the effectiveness and robustness of the QPBS protocol,we conduct comprehensive simulation studies in both ideal and noisy quantum environments on the IBM quantum cloud platform.The results demonstrate the superior performance of the QPBS algorithm,highlighting its resilience against repudiation and forgeability,which are key security concerns in the realm of proxy blind signatures.Furthermore,we have established authentic security thresholds(82.102%)in the presence of real noise,thereby emphasizing the practicality of our proposed solution.展开更多
Cloud computing has taken over the high-performance distributed computing area,and it currently provides on-demand services and resource polling over the web.As a result of constantly changing user service demand,the ...Cloud computing has taken over the high-performance distributed computing area,and it currently provides on-demand services and resource polling over the web.As a result of constantly changing user service demand,the task scheduling problem has emerged as a critical analytical topic in cloud computing.The primary goal of scheduling tasks is to distribute tasks to available processors to construct the shortest possible schedule without breaching precedence restrictions.Assignments and schedules of tasks substantially influence system operation in a heterogeneous multiprocessor system.The diverse processes inside the heuristic-based task scheduling method will result in varying makespan in the heterogeneous computing system.As a result,an intelligent scheduling algorithm should efficiently determine the priority of every subtask based on the resources necessary to lower the makespan.This research introduced a novel efficient scheduling task method in cloud computing systems based on the cooperation search algorithm to tackle an essential task and schedule a heterogeneous cloud computing problem.The basic idea of thismethod is to use the advantages of meta-heuristic algorithms to get the optimal solution.We assess our algorithm’s performance by running it through three scenarios with varying numbers of tasks.The findings demonstrate that the suggested technique beats existingmethods NewGenetic Algorithm(NGA),Genetic Algorithm(GA),Whale Optimization Algorithm(WOA),Gravitational Search Algorithm(GSA),and Hybrid Heuristic and Genetic(HHG)by 7.9%,2.1%,8.8%,7.7%,3.4%respectively according to makespan.展开更多
Mobile Edge Computing(MEC)assists clouds to handle enormous tasks from mobile devices in close proximity.The edge servers are not allocated efficiently according to the dynamic nature of the network.It leads to process...Mobile Edge Computing(MEC)assists clouds to handle enormous tasks from mobile devices in close proximity.The edge servers are not allocated efficiently according to the dynamic nature of the network.It leads to processing delay,and the tasks are dropped due to time limitations.The researchersfind it difficult and complex to determine the offloading decision because of uncertain load dynamic condition over the edge nodes.The challenge relies on the offload-ing decision on selection of edge nodes for offloading in a centralized manner.This study focuses on minimizing task-processing time while simultaneously increasing the success rate of service provided by edge servers.Initially,a task-offloading problem needs to be formulated based on the communication and pro-cessing.Then offloading decision problem is solved by deep analysis on taskflow in the network and feedback from the devices on edge services.The significance of the model is improved with the modelling of Deep Mobile-X architecture and bi-directional Long Short Term Memory(b-LSTM).The simulation is done in the Edgecloudsim environment,and the outcomes show the significance of the proposed idea.The processing time of the anticipated model is 6.6 s.The following perfor-mance metrics,improved server utilization,the ratio of the dropped task,and number of offloading tasks are evaluated and compared with existing learning approaches.The proposed model shows a better trade-off compared to existing approaches.展开更多
This paper presents a novel fuzzy firefly-based intelligent algorithm for load balancing in mobile cloud computing while reducing makespan.The proposed technique implicitly acts intelligently by using inherent traits ...This paper presents a novel fuzzy firefly-based intelligent algorithm for load balancing in mobile cloud computing while reducing makespan.The proposed technique implicitly acts intelligently by using inherent traits of fuzzy and firefly.It automatically adjusts its behavior or converges depending on the information gathered during the search process and objective function.It works for 3-tier architecture,including cloudlet and public cloud.As cloudlets have limited resources,fuzzy logic is used for cloudlet selection using capacity and waiting time as input.Fuzzy provides human-like decisions without using any mathematical model.Firefly is a powerful meta-heuristic optimization technique to balance diversification and solution speed.It balances the load on cloud and cloudlet while minimizing makespan and execution time.However,it may trap in local optimum;levy flight can handle it.Hybridization of fuzzy fireflywith levy flight is a novel technique that provides reduced makespan,execution time,and Degree of imbalance while balancing the load.Simulation has been carried out on the Cloud Analyst platform with National Aeronautics and Space Administration(NASA)and Clarknet datasets.Results show that the proposed algorithm outperforms Ant Colony Optimization Queue Decision Maker(ACOQDM),Distributed Scheduling Optimization Algorithm(DSOA),andUtility-based Firefly Algorithm(UFA)when compared in terms of makespan,Degree of imbalance,and Figure of Merit.展开更多
Healthcare is a fundamental part of every individual’s life.The healthcare industry is developing very rapidly with the help of advanced technologies.Many researchers are trying to build cloud-based healthcare applic...Healthcare is a fundamental part of every individual’s life.The healthcare industry is developing very rapidly with the help of advanced technologies.Many researchers are trying to build cloud-based healthcare applications that can be accessed by healthcare professionals from their premises,as well as by patients from their mobile devices through communication interfaces.These systems promote reliable and remote interactions between patients and healthcare professionals.However,there are several limitations to these innovative cloud computing-based systems,namely network availability,latency,battery life and resource availability.We propose a hybrid mobile cloud computing(HMCC)architecture to address these challenges.Furthermore,we also evaluate the performance of heuristic and dynamic machine learning based task scheduling and load balancing algorithms on our proposed architecture.We compare them,to identify the strengths and weaknesses of each algorithm;and provide their comparative results,to show latency and energy consumption performance.Challenging issues for cloudbased healthcare systems are discussed in detail.展开更多
文摘As the extensive use of cloud computing raises questions about the security of any personal data stored there,cryptography is being used more frequently as a security tool to protect data confidentiality and privacy in the cloud environment.A hypervisor is a virtualization software used in cloud hosting to divide and allocate resources on various pieces of hardware.The choice of hypervisor can significantly impact the performance of cryptographic operations in the cloud environment.An important issue that must be carefully examined is that no hypervisor is completely superior in terms of performance;Each hypervisor should be examined to meet specific needs.The main objective of this study is to provide accurate results to compare the performance of Hyper-V and Kernel-based Virtual Machine(KVM)while implementing different cryptographic algorithms to guide cloud service providers and end users in choosing the most suitable hypervisor for their cryptographic needs.This study evaluated the efficiency of two hypervisors,Hyper-V and KVM,in implementing six cryptographic algorithms:Rivest,Shamir,Adleman(RSA),Advanced Encryption Standard(AES),Triple Data Encryption Standard(TripleDES),Carlisle Adams and Stafford Tavares(CAST-128),BLOWFISH,and TwoFish.The study’s findings show that KVM outperforms Hyper-V,with 12.2%less Central Processing Unit(CPU)use and 12.95%less time overall for encryption and decryption operations with various file sizes.The study’s findings emphasize how crucial it is to pick a hypervisor that is appropriate for cryptographic needs in a cloud environment,which could assist both cloud service providers and end users.Future research may focus more on how various hypervisors perform while handling cryptographic workloads.
文摘Cloud computing provides a diverse and adaptable resource pool over the internet,allowing users to tap into various resources as needed.It has been seen as a robust solution to relevant challenges.A significant delay can hamper the performance of IoT-enabled cloud platforms.However,efficient task scheduling can lower the cloud infrastructure’s energy consumption,thus maximizing the service provider’s revenue by decreasing user job processing times.The proposed Modified Chimp-Whale Optimization Algorithm called Modified Chimp-Whale Optimization Algorithm(MCWOA),combines elements of the Chimp Optimization Algorithm(COA)and the Whale Optimization Algorithm(WOA).To enhance MCWOA’s identification precision,the Sobol sequence is used in the population initialization phase,ensuring an even distribution of the population across the solution space.Moreover,the traditional MCWOA’s local search capabilities are augmented by incorporating the whale optimization algorithm’s bubble-net hunting and random search mechanisms into MCWOA’s position-updating process.This study demonstrates the effectiveness of the proposed approach using a two-story rigid frame and a simply supported beam model.Simulated outcomes reveal that the new method outperforms the original MCWOA,especially in multi-damage detection scenarios.MCWOA excels in avoiding false positives and enhancing computational speed,making it an optimal choice for structural damage detection.The efficiency of the proposed MCWOA is assessed against metrics such as energy usage,computational expense,task duration,and delay.The simulated data indicates that the new MCWOA outpaces other methods across all metrics.The study also references the Whale Optimization Algorithm(WOA),Chimp Algorithm(CA),Ant Lion Optimizer(ALO),Genetic Algorithm(GA)and Grey Wolf Optimizer(GWO).
基金jointly supported by the Jiangsu Postgraduate Research and Practice Innovation Project under Grant KYCX22_1030,SJCX22_0283 and SJCX23_0293the NUPTSF under Grant NY220201.
文摘Task scheduling plays a key role in effectively managing and allocating computing resources to meet various computing tasks in a cloud computing environment.Short execution time and low load imbalance may be the challenges for some algorithms in resource scheduling scenarios.In this work,the Hierarchical Particle Swarm Optimization-Evolutionary Artificial Bee Colony Algorithm(HPSO-EABC)has been proposed,which hybrids our presented Evolutionary Artificial Bee Colony(EABC),and Hierarchical Particle Swarm Optimization(HPSO)algorithm.The HPSO-EABC algorithm incorporates both the advantages of the HPSO and the EABC algorithm.Comprehensive testing including evaluations of algorithm convergence speed,resource execution time,load balancing,and operational costs has been done.The results indicate that the EABC algorithm exhibits greater parallelism compared to the Artificial Bee Colony algorithm.Compared with the Particle Swarm Optimization algorithm,the HPSO algorithmnot only improves the global search capability but also effectively mitigates getting stuck in local optima.As a result,the hybrid HPSO-EABC algorithm demonstrates significant improvements in terms of stability and convergence speed.Moreover,it exhibits enhanced resource scheduling performance in both homogeneous and heterogeneous environments,effectively reducing execution time and cost,which also is verified by the ablation experimental.
文摘Security issues in cloud networks and edge computing have become very common. This research focuses on analyzing such issues and developing the best solutions. A detailed literature review has been conducted in this regard. The findings have shown that many challenges are linked to edge computing, such as privacy concerns, security breaches, high costs, low efficiency, etc. Therefore, there is a need to implement proper security measures to overcome these issues. Using emerging trends, like machine learning, encryption, artificial intelligence, real-time monitoring, etc., can help mitigate security issues. They can also develop a secure and safe future in cloud computing. It was concluded that the security implications of edge computing can easily be covered with the help of new technologies and techniques.
基金funded by the Key Research and Development plan of Jiangsu Province (Social Development)No.BE20217162Jiangsu Modern Agricultural Machinery Equipment and Technology Demonstration and Promotion Project No.NJ2021-19.
文摘With the rapid development of the Internet of Things(IoT),there are several challenges pertaining to security in IoT applications.Compared with the characteristics of the traditional Internet,the IoT has many problems,such as large assets,complex and diverse structures,and lack of computing resources.Traditional network intrusion detection systems cannot meet the security needs of IoT applications.In view of this situation,this study applies cloud computing and machine learning to the intrusion detection system of IoT to improve detection performance.Usually,traditional intrusion detection algorithms require considerable time for training,and these intrusion detection algorithms are not suitable for cloud computing due to the limited computing power and storage capacity of cloud nodes;therefore,it is necessary to study intrusion detection algorithms with low weights,short training time,and high detection accuracy for deployment and application on cloud nodes.An appropriate classification algorithm is a primary factor for deploying cloud computing intrusion prevention systems and a prerequisite for the system to respond to intrusion and reduce intrusion threats.This paper discusses the problems related to IoT intrusion prevention in cloud computing environments.Based on the analysis of cloud computing security threats,this study extensively explores IoT intrusion detection,cloud node monitoring,and intrusion response in cloud computing environments by using cloud computing,an improved extreme learning machine,and other methods.We use the Multi-Feature Extraction Extreme Learning Machine(MFE-ELM)algorithm for cloud computing,which adds a multi-feature extraction process to cloud servers,and use the deployed MFE-ELM algorithm on cloud nodes to detect and discover network intrusions to cloud nodes.In our simulation experiments,a classical dataset for intrusion detection is selected as a test,and test steps such as data preprocessing,feature engineering,model training,and result analysis are performed.The experimental results show that the proposed algorithm can effectively detect and identify most network data packets with good model performance and achieve efficient intrusion detection for heterogeneous data of the IoT from cloud nodes.Furthermore,it can enable the cloud server to discover nodes with serious security threats in the cloud cluster in real time,so that further security protection measures can be taken to obtain the optimal intrusion response strategy for the cloud cluster.
基金Supporting Project number(RSP2023R34)King Saud University,Riyadh,Saudi Arabia.
文摘Some of the significant new technologies researched in recent studies include BlockChain(BC),Software Defined Networking(SDN),and Smart Industrial Internet of Things(IIoT).All three technologies provide data integrity,confidentiality,and integrity in their respective use cases(especially in industrial fields).Additionally,cloud computing has been in use for several years now.Confidential information is exchanged with cloud infrastructure to provide clients with access to distant resources,such as computing and storage activities in the IIoT.There are also significant security risks,concerns,and difficulties associated with cloud computing.To address these challenges,we propose merging BC and SDN into a cloud computing platform for the IIoT.This paper introduces“DistB-SDCloud”,an architecture for enhanced cloud security for smart IIoT applications.The proposed architecture uses a distributed BC method to provide security,secrecy,privacy,and integrity while remaining flexible and scalable.Customers in the industrial sector benefit from the dispersed or decentralized,and efficient environment of BC.Additionally,we described an SDN method to improve the durability,stability,and load balancing of cloud infrastructure.The efficacy of our SDN and BC-based implementation was experimentally tested by using various parameters including throughput,packet analysis,response time,bandwidth,and latency analysis,as well as the monitoring of several attacks on the system itself.
基金supported by the key scientific and technological projects of Henan Province with Grant No.232102211084the Natural Science Foundation of Henan with Grant No.222300420582+2 种基金the Key Scientific Research Projects of Henan Higher School with Grant No.22A520033Zhengzhou Basic Research and Applied Research Project with Grant No.ZZSZX202107China Logistics Society with Grant No.2022CSLKT3-334.
文摘To serve various tasks requested by various end devices with different requirements,end-edge-cloud(E2C)has attracted more and more attention from specialists in both academia and industry,by combining both benefits of edge and cloud computing.But nowadays,E2C still suffers from low service quality and resource efficiency,due to the geographical distribution of edge resources and the high dynamic of network topology and user mobility.To address these issues,this paper focuses on task offloading,which makes decisions that which resources are allocated to tasks for their processing.This paper first formulates the problem into binary non-linear programming and then proposes a particle swarm optimization(PSO)-based algorithm to solve the problem.The proposed algorithm exploits an imbalance mutation operator and a task rescheduling approach to improve the performance of PSO.The proposed algorithm concerns the resource heterogeneity by correlating the probability that a computing node is decided to process a task with its capacity,by the imbalance mutation.The task rescheduling approach improves the acceptance ratio for a task offloading solution,by reassigning rejected tasks to computing nodes with available resources.Extensive simulated experiments are conducted.And the results show that the proposed offloading algorithm has an 8.93%–37.0%higher acceptance ratio than ten of the classical and up-to-date algorithms,and verify the effectiveness of the imbalanced mutation and the task rescheduling.
基金supported by Science and Technology Rising Star of Shaanxi Youth (No.2021KJXX-61)The Open Project Program of the State Key Lab of CAD&CG,Zhejiang University (No.A2206).
文摘Star sensors are an important means of autonomous navigation and access to space information for satellites.They have been widely deployed in the aerospace field.To satisfy the requirements for high resolution,timeliness,and confidentiality of star images,we propose an edge computing algorithm based on the star sensor cloud.Multiple sensors cooperate with each other to forma sensor cloud,which in turn extends the performance of a single sensor.The research on the data obtained by the star sensor has very important research and application values.First,a star point extraction model is proposed based on the fuzzy set model by analyzing the star image composition,which can reduce the amount of data computation.Then,a mappingmodel between content and space is constructed to achieve low-rank image representation and efficient computation.Finally,the data collected by the wireless sensor is delivered to the edge server,and a differentmethod is used to achieve privacy protection.Only a small amount of core data is stored in edge servers and local servers,and other data is transmitted to the cloud.Experiments show that the proposed algorithm can effectively reduce the cost of communication and storage,and has strong privacy.
基金supported by the Yunnan Key Laboratory of Optoelectronic Information Technology,and grant funded by the National Natural Science Foundation of China(62062069,62062070,and 62005235)Taif University Researchers Supporting Project(TURSP-2020/126)Taif University,Taif,Saudi Arabia.Jun Liu and Kai Qian contributed equally to this paper。
文摘In recent years,statistics have indicated that the number of patients with malignant brain tumors has increased sharply.However,most surgeons still perform surgical training using the traditional autopsy and prosthesis model,which encounters many problems,such as insufficient corpse resources,low efficiency,and high cost.With the advent of the 5G era,a wide range of Industrial Internet of Things(IIOT)applications have been developed.Virtual Reality(VR)and Augmented Reality(AR)technologies that emerged with 5G are developing rapidly for intelligent medical training.To address the challenges encountered during neurosurgery training,and combining with cloud computing,in this paper,a highly immersive AR-based brain tumor neurosurgery remote collaborative virtual surgery training system is developed,in which a VR simulator is embedded.The system enables real-time remote surgery training interaction through 5G transmission.Six experts and 18 novices were invited to participate in the experiment to verify the system.Subsequently,the two simulators were evaluated using face and construction validation methods.The results obtained by training the novices 50 times were further analyzed using the Learning Curve-Cumulative Sum(LC-CUSUM)evaluation method to validate the effectiveness of the two simulators.The results of the face and content validation demonstrated that the AR simulator in the system was superior to the VR simulator in terms of vision and scene authenticity,and had a better effect on the improvement of surgical skills.Moreover,the surgical training scheme proposed in this paper is effective,and the remote collaborative training effect of the system is ideal.
基金We would also thank the support from the National Natural Science Foundation of China(Nos.62172182,62202118,61962009)the Top Technology Talent Project from Guizhou Education Department(Qian jiao ji[2022]073)The Opening Foundation of Key Laboratory of Intelligent Control Technology for Wuling-Mountain Ecological Agriculture in Hunan Province(Grant No.ZNKZN2021-07).
文摘Due to its decentralized,tamper-proof,and trust-free characteristics,blockchain is used in the Internet of Things(IoT)to guarantee the reliability of data.However,some technical flaws in blockchain itself prevent the development of these applications,such as the issue with linearly growing storage capacity of blockchain systems.On the other hand,there is a lack of storage resources for sensor devices in IoT,and numerous sensor devices will generate massive data at ultra-high speed,which makes the storage problem of the IoT enabled by blockchain more prominent.There are various solutions to reduce the storage burden by modifying the blockchain’s storage policy,but most of them do not consider the willingness of peers.In attempt to make the blockchain more compatible with the IoT,this paper proposes a storage optimization scheme that revisits the system data storage problem from amore practically oriented standpoint.Peers will only store transactional data that they are directly involved in.In addition,a transaction verification model is developed to enable peers to undertake transaction verification with the aid of cloud computing,and an incentive mechanism is premised on the storage optimization scheme to assure data integrity.The results of the simulation experiments demonstrate the proposed scheme’s advantage in terms of storage and throughput.
基金supported in part by the National Key R&D Program of China under Grant 2017YFB1302400the Jinan“20 New Colleges and Universities”Funded Scientific Research Leader Studio under Grant 2021GXRC079+2 种基金the Major Agricultural Applied Technological Innovation Projects of Shandong Province underGrant SD2019NJ014the Shandong Natural Science Foundation under Grant ZR2019MF064the Beijing Advanced Innovation Center for Intelligent Robots and Systems under Grant 2019IRS19.
文摘The solution strategy of the heuristic algorithm is pre-set and has good performance in the conventional cloud resource scheduling process.However,for complex and dynamic cloud service scheduling tasks,due to the difference in service attributes,the solution efficiency of a single strategy is low for such problems.In this paper,we presents a hyper-heuristic algorithm based on reinforcement learning(HHRL)to optimize the completion time of the task sequence.Firstly,In the reward table setting stage of HHRL,we introduce population diversity and integrate maximum time to comprehensively deter-mine the task scheduling and the selection of low-level heuristic strategies.Secondly,a task computational complexity estimation method integrated with linear regression is proposed to influence task scheduling priorities.Besides,we propose a high-quality candidate solution migration method to ensure the continuity and diversity of the solving process.Compared with HHSA,ACO,GA,F-PSO,etc,HHRL can quickly obtain task complexity,select appropriate heuristic strategies for task scheduling,search for the the best makspan and have stronger disturbance detection ability for population diversity.
文摘In the present scenario of rapid growth in cloud computing models,several companies and users started to share their data on cloud servers.However,when the model is not completely trusted,the data owners face several security-related problems,such as user privacy breaches,data disclosure,data corruption,and so on,during the process of data outsourcing.For addressing and handling the security-related issues on Cloud,several models were proposed.With that concern,this paper develops a Privacy-Preserved Data Security Approach(PP-DSA)to provide the data security and data integrity for the out-sourcing data in Cloud Environment.Privacy preservation is ensured in this work with the Efficient Authentication Technique(EAT)using the Group Signature method that is applied with Third-Party Auditor(TPA).The role of the auditor is to secure the data and guarantee shared data integrity.Additionally,the Cloud Service Provider(CSP)and Data User(DU)can also be the attackers that are to be handled with the EAT.Here,the major objective of the work is to enhance cloud security and thereby,increase Quality of Service(QoS).The results are evaluated based on the model effectiveness,security,and reliability and show that the proposed model provides better results than existing works.
文摘Cloud computing(CC)is developing as a powerful and flexible computational structure for providing ubiquitous service to users.It receives interrelated software and hardware resources in an integrated manner distinct from the classical computational environment.The variation of software and hardware resources were combined and composed as a resource pool.The software no more resided in the single hardware environment,it can be executed on the schedule of resource pools to optimize resource consumption.Optimizing energy consumption in CC environments is the question that allows utilizing several energy conservation approaches for effective resource allocation.This study introduces a Battle Royale Optimization-based Resource Scheduling Scheme for Cloud Computing Environment(BRORSS-CCE)technique.The presented BRORSS-CCE technique majorly schedules the available resources for maximum utilization and effectual makespan.In the BRORSS-CCE technique,the BRO is a population-based algorithm where all the individuals are denoted by a soldier/player who likes to go towards the optimal place and ultimate survival.The BRORSS-CCE technique can be employed to balance the load,distribute resources based on demand and assure services to all requests.The experimental validation of the BRORSS-CCE technique is tested under distinct aspects.The experimental outcomes indicated the enhancements of the BRORSS-CCE technique over other models.
基金supported by the National Research Foundation of Korea (NRF) grant funded by the Korea Government (MSIT) (No.2021R1C1C1013133)supported by the Institute of Information and Communications Technology Planning and Evaluation (IITP)grant funded by the Korea Government (MSIT) (RS-2022-00167197,Development of Intelligent 5G/6G Infrastructure Technology for The Smart City)supported by the Soonchunhyang University Research Fund.
文摘In many IIoT architectures,various devices connect to the edge cloud via gateway systems.For data processing,numerous data are delivered to the edge cloud.Delivering data to an appropriate edge cloud is critical to improve IIoT service efficiency.There are two types of costs for this kind of IoT network:a communication cost and a computing cost.For service efficiency,the communication cost of data transmission should be minimized,and the computing cost in the edge cloud should be also minimized.Therefore,in this paper,the communication cost for data transmission is defined as the delay factor,and the computing cost in the edge cloud is defined as the waiting time of the computing intensity.The proposed method selects an edge cloud that minimizes the total cost of the communication and computing costs.That is,a device chooses a routing path to the selected edge cloud based on the costs.The proposed method controls the data flows in a mesh-structured network and appropriately distributes the data processing load.The performance of the proposed method is validated through extensive computer simulation.When the transition probability from good to bad is 0.3 and the transition probability from bad to good is 0.7 in wireless and edge cloud states,the proposed method reduced both the average delay and the service pause counts to about 25%of the existing method.
基金supported by the Foundation for Innovative Research Groups of the National Natural Science Foundation of China under Grant No.61521003the National Natural Science Foundation of China under Grant No.62072467 and 62002383.
文摘Serverless computing is a promising paradigm in cloud computing that greatly simplifies cloud programming.With serverless computing,developers only provide function code to serverless platform,and these functions are invoked by its driven events.Nonetheless,security threats in serverless computing such as vulnerability-based security threats have become the pain point hindering its wide adoption.The ideas in proactive defense such as redundancy,diversity and dynamic provide promising approaches to protect against cyberattacks.However,these security technologies are mostly applied to serverless platform based on“stacked”mode,as they are designed independent with serverless computing.The lack of security consideration in the initial design makes it especially challenging to achieve the all life cycle protection for serverless application with limited cost.In this paper,we present ATSSC,a proactive defense enabled attack tolerant serverless platform.ATSSC integrates the characteristic of redundancy,diversity and dynamic into serverless seamless to achieve high-level security and efficiency.Specifically,ATSSC constructs multiple diverse function replicas to process the driven events and performs cross-validation to verify the results.In order to create diverse function replicas,both software diversity and environment diversity are adopted.Furthermore,a dynamic function refresh strategy is proposed to keep the clean state of serverless functions.We implement ATSSC based on Kubernetes and Knative.Analysis and experimental results demonstrate that ATSSC can effectively protect serverless computing against cyberattacks with acceptable costs.
基金Project supported by the General Project of Natural Science Foundation of Hunan Province(Grant Nos.2024JJ5273 and 2023JJ50328)the Scientific Research Project of Education Department of Hunan Province(Grant Nos.22A0049 and 22B0699)。
文摘This paper presents a novel approach to proxy blind signatures in the realm of quantum circuits,aiming to enhance security while safeguarding sensitive information.The main objective of this research is to introduce a quantum proxy blind signature(QPBS)protocol that utilizes quantum logical gates and quantum measurement techniques.The QPBS protocol is constructed by the initial phase,proximal blinding message phase,remote authorization and signature phase,remote validation,and de-blinding phase.This innovative design ensures a secure mechanism for signing documents without revealing the content to the proxy signer,providing practical security authentication in a quantum environment under the assumption that the CNOT gates are securely implemented.Unlike existing approaches,our proposed QPBS protocol eliminates the need for quantum entanglement preparation,thus simplifying the implementation process.To assess the effectiveness and robustness of the QPBS protocol,we conduct comprehensive simulation studies in both ideal and noisy quantum environments on the IBM quantum cloud platform.The results demonstrate the superior performance of the QPBS algorithm,highlighting its resilience against repudiation and forgeability,which are key security concerns in the realm of proxy blind signatures.Furthermore,we have established authentic security thresholds(82.102%)in the presence of real noise,thereby emphasizing the practicality of our proposed solution.
文摘Cloud computing has taken over the high-performance distributed computing area,and it currently provides on-demand services and resource polling over the web.As a result of constantly changing user service demand,the task scheduling problem has emerged as a critical analytical topic in cloud computing.The primary goal of scheduling tasks is to distribute tasks to available processors to construct the shortest possible schedule without breaching precedence restrictions.Assignments and schedules of tasks substantially influence system operation in a heterogeneous multiprocessor system.The diverse processes inside the heuristic-based task scheduling method will result in varying makespan in the heterogeneous computing system.As a result,an intelligent scheduling algorithm should efficiently determine the priority of every subtask based on the resources necessary to lower the makespan.This research introduced a novel efficient scheduling task method in cloud computing systems based on the cooperation search algorithm to tackle an essential task and schedule a heterogeneous cloud computing problem.The basic idea of thismethod is to use the advantages of meta-heuristic algorithms to get the optimal solution.We assess our algorithm’s performance by running it through three scenarios with varying numbers of tasks.The findings demonstrate that the suggested technique beats existingmethods NewGenetic Algorithm(NGA),Genetic Algorithm(GA),Whale Optimization Algorithm(WOA),Gravitational Search Algorithm(GSA),and Hybrid Heuristic and Genetic(HHG)by 7.9%,2.1%,8.8%,7.7%,3.4%respectively according to makespan.
文摘Mobile Edge Computing(MEC)assists clouds to handle enormous tasks from mobile devices in close proximity.The edge servers are not allocated efficiently according to the dynamic nature of the network.It leads to processing delay,and the tasks are dropped due to time limitations.The researchersfind it difficult and complex to determine the offloading decision because of uncertain load dynamic condition over the edge nodes.The challenge relies on the offload-ing decision on selection of edge nodes for offloading in a centralized manner.This study focuses on minimizing task-processing time while simultaneously increasing the success rate of service provided by edge servers.Initially,a task-offloading problem needs to be formulated based on the communication and pro-cessing.Then offloading decision problem is solved by deep analysis on taskflow in the network and feedback from the devices on edge services.The significance of the model is improved with the modelling of Deep Mobile-X architecture and bi-directional Long Short Term Memory(b-LSTM).The simulation is done in the Edgecloudsim environment,and the outcomes show the significance of the proposed idea.The processing time of the anticipated model is 6.6 s.The following perfor-mance metrics,improved server utilization,the ratio of the dropped task,and number of offloading tasks are evaluated and compared with existing learning approaches.The proposed model shows a better trade-off compared to existing approaches.
基金funded by University Grant Commission with UGC-Ref.No.:3364/(NET-JUNE 2015).
文摘This paper presents a novel fuzzy firefly-based intelligent algorithm for load balancing in mobile cloud computing while reducing makespan.The proposed technique implicitly acts intelligently by using inherent traits of fuzzy and firefly.It automatically adjusts its behavior or converges depending on the information gathered during the search process and objective function.It works for 3-tier architecture,including cloudlet and public cloud.As cloudlets have limited resources,fuzzy logic is used for cloudlet selection using capacity and waiting time as input.Fuzzy provides human-like decisions without using any mathematical model.Firefly is a powerful meta-heuristic optimization technique to balance diversification and solution speed.It balances the load on cloud and cloudlet while minimizing makespan and execution time.However,it may trap in local optimum;levy flight can handle it.Hybridization of fuzzy fireflywith levy flight is a novel technique that provides reduced makespan,execution time,and Degree of imbalance while balancing the load.Simulation has been carried out on the Cloud Analyst platform with National Aeronautics and Space Administration(NASA)and Clarknet datasets.Results show that the proposed algorithm outperforms Ant Colony Optimization Queue Decision Maker(ACOQDM),Distributed Scheduling Optimization Algorithm(DSOA),andUtility-based Firefly Algorithm(UFA)when compared in terms of makespan,Degree of imbalance,and Figure of Merit.
基金supported by the Bio and Medical Technology Development Program of the National Research Foundation(NRF)funded by the Korean government(MSIT)(No.NRF-2019M3E5D1A02069073)supported by the Soonchunhyang University Research Fund.
文摘Healthcare is a fundamental part of every individual’s life.The healthcare industry is developing very rapidly with the help of advanced technologies.Many researchers are trying to build cloud-based healthcare applications that can be accessed by healthcare professionals from their premises,as well as by patients from their mobile devices through communication interfaces.These systems promote reliable and remote interactions between patients and healthcare professionals.However,there are several limitations to these innovative cloud computing-based systems,namely network availability,latency,battery life and resource availability.We propose a hybrid mobile cloud computing(HMCC)architecture to address these challenges.Furthermore,we also evaluate the performance of heuristic and dynamic machine learning based task scheduling and load balancing algorithms on our proposed architecture.We compare them,to identify the strengths and weaknesses of each algorithm;and provide their comparative results,to show latency and energy consumption performance.Challenging issues for cloudbased healthcare systems are discussed in detail.