This paper reviews task scheduling frameworks,methods,and evaluation metrics of central processing unit-graphics processing unit(CPU-GPU)heterogeneous clusters.Task scheduling of CPU-GPU heterogeneous clusters can be ...This paper reviews task scheduling frameworks,methods,and evaluation metrics of central processing unit-graphics processing unit(CPU-GPU)heterogeneous clusters.Task scheduling of CPU-GPU heterogeneous clusters can be carried out on the system level,nodelevel,and device level.Most task-scheduling technologies are heuristic based on the experts’experience,while some technologies are based on statistic methods using machine learning,deep learning,or reinforcement learning.Many metrics have been adopted to evaluate and compare different task scheduling technologies that try to optimize different goals of task scheduling.Although statistic task scheduling has reached fewer research achievements than heuristic task scheduling,the statistic task scheduling still has significant research potential.展开更多
More devices in the Intelligent Internet of Things(AIoT)result in an increased number of tasks that require low latency and real-time responsiveness,leading to an increased demand for computational resources.Cloud com...More devices in the Intelligent Internet of Things(AIoT)result in an increased number of tasks that require low latency and real-time responsiveness,leading to an increased demand for computational resources.Cloud computing’s low-latency performance issues in AIoT scenarios have led researchers to explore fog computing as a complementary extension.However,the effective allocation of resources for task execution within fog environments,characterized by limitations and heterogeneity in computational resources,remains a formidable challenge.To tackle this challenge,in this study,we integrate fog computing and cloud computing.We begin by establishing a fog-cloud environment framework,followed by the formulation of a mathematical model for task scheduling.Lastly,we introduce an enhanced hybrid Equilibrium Optimizer(EHEO)tailored for AIoT task scheduling.The overarching objective is to decrease both the makespan and energy consumption of the fog-cloud system while accounting for task deadlines.The proposed EHEO method undergoes a thorough evaluation against multiple benchmark algorithms,encompassing metrics likemakespan,total energy consumption,success rate,and average waiting time.Comprehensive experimental results unequivocally demonstrate the superior performance of EHEO across all assessed metrics.Notably,in the most favorable conditions,EHEO significantly diminishes both the makespan and energy consumption by approximately 50%and 35.5%,respectively,compared to the secondbest performing approach,which affirms its efficacy in advancing the efficiency of AIoT task scheduling within fog-cloud networks.展开更多
A dynamic multi-beam resource allocation algorithm for large low Earth orbit(LEO)constellation based on on-board distributed computing is proposed in this paper.The allocation is a combinatorial optimization process u...A dynamic multi-beam resource allocation algorithm for large low Earth orbit(LEO)constellation based on on-board distributed computing is proposed in this paper.The allocation is a combinatorial optimization process under a series of complex constraints,which is important for enhancing the matching between resources and requirements.A complex algorithm is not available because that the LEO on-board resources is limi-ted.The proposed genetic algorithm(GA)based on two-dimen-sional individual model and uncorrelated single paternal inheri-tance method is designed to support distributed computation to enhance the feasibility of on-board application.A distributed system composed of eight embedded devices is built to verify the algorithm.A typical scenario is built in the system to evalu-ate the resource allocation process,algorithm mathematical model,trigger strategy,and distributed computation architec-ture.According to the simulation and measurement results,the proposed algorithm can provide an allocation result for more than 1500 tasks in 14 s and the success rate is more than 91%in a typical scene.The response time is decreased by 40%com-pared with the conditional GA.展开更多
Deploying service nodes hierarchically at the edge of the network can effectively improve the service quality of offloaded task requests and increase the utilization of resources.In this paper,we study the task schedu...Deploying service nodes hierarchically at the edge of the network can effectively improve the service quality of offloaded task requests and increase the utilization of resources.In this paper,we study the task scheduling problem in the hierarchically deployed edge cloud.We first formulate the minimization of the service time of scheduled tasks in edge cloud as a combinatorial optimization problem,blue and then prove the NP-hardness of the problem.Different from the existing work that mostly designs heuristic approximation-based algorithms or policies to make scheduling decision,we propose a newly designed scheduling policy,named Joint Neural Network and Heuristic Scheduling(JNNHSP),which combines a neural network-based method with a heuristic based solution.JNNHSP takes the Sequence-to-Sequence(Seq2Seq)model trained by Reinforcement Learning(RL)as the primary policy and adopts the heuristic algorithm as the auxiliary policy to obtain the scheduling solution,thereby achieving a good balance between the quality and the efficiency of the scheduling solution.In-depth experiments show that compared with a variety of related policies and optimization solvers,JNNHSP can achieve better performance in terms of scheduling error ratio,the degree to which the policy is affected by re-sources limitations,average service latency,and execution efficiency in a typical hierarchical edge cloud.展开更多
Cloud computing has taken over the high-performance distributed computing area,and it currently provides on-demand services and resource polling over the web.As a result of constantly changing user service demand,the ...Cloud computing has taken over the high-performance distributed computing area,and it currently provides on-demand services and resource polling over the web.As a result of constantly changing user service demand,the task scheduling problem has emerged as a critical analytical topic in cloud computing.The primary goal of scheduling tasks is to distribute tasks to available processors to construct the shortest possible schedule without breaching precedence restrictions.Assignments and schedules of tasks substantially influence system operation in a heterogeneous multiprocessor system.The diverse processes inside the heuristic-based task scheduling method will result in varying makespan in the heterogeneous computing system.As a result,an intelligent scheduling algorithm should efficiently determine the priority of every subtask based on the resources necessary to lower the makespan.This research introduced a novel efficient scheduling task method in cloud computing systems based on the cooperation search algorithm to tackle an essential task and schedule a heterogeneous cloud computing problem.The basic idea of thismethod is to use the advantages of meta-heuristic algorithms to get the optimal solution.We assess our algorithm’s performance by running it through three scenarios with varying numbers of tasks.The findings demonstrate that the suggested technique beats existingmethods NewGenetic Algorithm(NGA),Genetic Algorithm(GA),Whale Optimization Algorithm(WOA),Gravitational Search Algorithm(GSA),and Hybrid Heuristic and Genetic(HHG)by 7.9%,2.1%,8.8%,7.7%,3.4%respectively according to makespan.展开更多
Cloud computing technology is favored by users because of its strong computing power and convenient services.At the same time,scheduling performance has an extremely efficient impact on promoting carbon neutrality.Cur...Cloud computing technology is favored by users because of its strong computing power and convenient services.At the same time,scheduling performance has an extremely efficient impact on promoting carbon neutrality.Currently,scheduling research in the multi-cloud environment aims to address the challenges brought by business demands to cloud data centers during peak hours.Therefore,the scheduling problem has promising application prospects under themulti-cloud environment.This paper points out that the currently studied scheduling problems in the multi-cloud environment mainly include independent task scheduling and workflow task scheduling based on the dependencies between tasks.This paper reviews the concepts,types,objectives,advantages,challenges,and research status of task scheduling in the multi-cloud environment.Task scheduling strategies proposed in the existing related references are analyzed,discussed,and summarized,including research motivation,optimization algorithm,and related objectives.Finally,the research status of the two kinds of task scheduling is compared,and several future important research directions of multi-cloud task scheduling are proposed.展开更多
Deploying task caching at edge servers has become an effectiveway to handle compute-intensive and latency-sensitive tasks on the industrialinternet. However, how to select the task scheduling location to reduce taskde...Deploying task caching at edge servers has become an effectiveway to handle compute-intensive and latency-sensitive tasks on the industrialinternet. However, how to select the task scheduling location to reduce taskdelay and cost while ensuring the data security and reliable communicationof edge computing remains a challenge. To solve this problem, this paperestablishes a task scheduling model with joint blockchain and task cachingin the industrial internet and designs a novel blockchain-assisted cachingmechanism to enhance system security. In this paper, the task schedulingproblem, which couples the task scheduling decision, task caching decision,and blockchain reward, is formulated as the minimum weighted cost problemunder delay constraints. This is a mixed integer nonlinear problem, which isproved to be nonconvex and NP-hard. To solve the optimal solution, thispaper proposes a task scheduling strategy algorithm based on an improvedgenetic algorithm (IGA-TSPA) by improving the genetic algorithm initializationand mutation operations to reduce the size of the initial solutionspace and enhance the optimal solution convergence speed. In addition,an Improved Least Frequently Used algorithm is proposed to improve thecontent hit rate. Simulation results show that IGA-TSPA has a faster optimalsolution-solving ability and shorter running time compared with the existingedge computing scheduling algorithms. The established task scheduling modelnot only saves 62.19% of system overhead consumption in comparison withlocal computing but also has great significance in protecting data security,reducing task processing delay, and reducing system cost.展开更多
Due to the security and scalability features of hybrid cloud architecture,it can bettermeet the diverse requirements of users for cloud services.And a reasonable resource allocation solution is the key to adequately u...Due to the security and scalability features of hybrid cloud architecture,it can bettermeet the diverse requirements of users for cloud services.And a reasonable resource allocation solution is the key to adequately utilize the hybrid cloud.However,most previous studies have not comprehensively optimized the performance of hybrid cloud task scheduling,even ignoring the conflicts between its security privacy features and other requirements.Based on the above problems,a many-objective hybrid cloud task scheduling optimization model(HCTSO)is constructed combining risk rate,resource utilization,total cost,and task completion time.Meanwhile,an opposition-based learning knee point-driven many-objective evolutionary algorithm(OBL-KnEA)is proposed to improve the performance of model solving.The algorithm uses opposition-based learning to generate initial populations for faster convergence.Furthermore,a perturbation-based multipoint crossover operator and a dynamic range mutation operator are designed to extend the search range.By comparing the experiments with other excellent algorithms on HCTSO,OBL-KnEA achieves excellent results in terms of evaluation metrics,initial populations,and model optimization effects.展开更多
Owing to massive technological developments in Internet of Things(IoT)and cloud environment,cloud computing(CC)offers a highly flexible heterogeneous resource pool over the network,and clients could exploit various re...Owing to massive technological developments in Internet of Things(IoT)and cloud environment,cloud computing(CC)offers a highly flexible heterogeneous resource pool over the network,and clients could exploit various resources on demand.Since IoT-enabled models are restricted to resources and require crisp response,minimum latency,and maximum bandwidth,which are outside the capabilities.CC was handled as a resource-rich solution to aforementioned challenge.As high delay reduces the performance of the IoT enabled cloud platform,efficient utilization of task scheduling(TS)reduces the energy usage of the cloud infrastructure and increases the income of service provider via minimizing processing time of user job.Therefore,this article concentration on the design of an oppositional red fox optimization based task scheduling scheme(ORFOTSS)for IoT enabled cloud environment.The presented ORFO-TSS model resolves the problem of allocating resources from the IoT based cloud platform.It achieves the makespan by performing optimum TS procedures with various aspects of incoming task.The designing of ORFO-TSS method includes the idea of oppositional based learning(OBL)as to traditional RFO approach in enhancing their efficiency.A wide-ranging experimental analysis was applied on the CloudSim platform.The experimental outcome highlighted the efficacy of the ORFO-TSS technique over existing approaches.展开更多
The solution strategy of the heuristic algorithm is pre-set and has good performance in the conventional cloud resource scheduling process.However,for complex and dynamic cloud service scheduling tasks,due to the diff...The solution strategy of the heuristic algorithm is pre-set and has good performance in the conventional cloud resource scheduling process.However,for complex and dynamic cloud service scheduling tasks,due to the difference in service attributes,the solution efficiency of a single strategy is low for such problems.In this paper,we presents a hyper-heuristic algorithm based on reinforcement learning(HHRL)to optimize the completion time of the task sequence.Firstly,In the reward table setting stage of HHRL,we introduce population diversity and integrate maximum time to comprehensively deter-mine the task scheduling and the selection of low-level heuristic strategies.Secondly,a task computational complexity estimation method integrated with linear regression is proposed to influence task scheduling priorities.Besides,we propose a high-quality candidate solution migration method to ensure the continuity and diversity of the solving process.Compared with HHSA,ACO,GA,F-PSO,etc,HHRL can quickly obtain task complexity,select appropriate heuristic strategies for task scheduling,search for the the best makspan and have stronger disturbance detection ability for population diversity.展开更多
In a cloud environment,consumers search for the best service provider that accomplishes the required tasks based on a set of criteria such as completion time and cost.On the other hand,Cloud Service Providers(CSPs)see...In a cloud environment,consumers search for the best service provider that accomplishes the required tasks based on a set of criteria such as completion time and cost.On the other hand,Cloud Service Providers(CSPs)seek to maximize their profits by attracting and serving more consumers based on their resource capabilities.The literature has discussed the problem by considering either consumers’needs or CSPs’capabilities.A problem resides in the lack of explicit models that combine preferences of consumers with the capabilities of CSPs to provide a unified process for resource allocation and task scheduling in a more efficient way.The paper proposes a model that adopts a Multi-Criteria Decision Making(MCDM)method,called Analytic Hierarchy Process(AHP),to acquire the information of consumers’preferences and service providers’capabilities to prioritize both tasks and resources.The model also provides a matching technique to assign each task to the best resource of a CSP while preserves the fairness of scheduling more tasks for resources with higher capabilities.Our experimental results prove the feasibility of the proposed model for prioritizing hundreds of tasks/services and CSPs based on a defined set of criteria,and matching each set of tasks/services to the best CSPS.展开更多
To solve the deadlock problem of tasks that the interdependence between tasks fails to consider during the course of resource assignment and task scheduling based on the heuristics algorithm, an improved ant colony sy...To solve the deadlock problem of tasks that the interdependence between tasks fails to consider during the course of resource assignment and task scheduling based on the heuristics algorithm, an improved ant colony system (ACS) based algorithm is proposed. First, how to map the resource assignment and task scheduling (RATS) problem into the optimization selection problem of task resource assignment graph (TRAG) and to add the semaphore mechanism in the optimal TRAG to solve deadlocks are explained. Secondly, how to utilize the grid pheromone system model to realize the algorithm based on ACS is explicated. This refers to the construction of TRAG by the random selection of appropriate resources for each task by the user agent and the optimization of TRAG through the positive feedback and distributed parallel computing mechanism of the ACS. Simulation results show that the proposed algorithm is effective and efficient in solving the deadlock problem.展开更多
In this paper combined with the advantages of genetic algorithm and simulated annealing, brings forward a parallel genetic simulated annealing hybrid algorithm (PGSAHA) and applied to solve task scheduling problem i...In this paper combined with the advantages of genetic algorithm and simulated annealing, brings forward a parallel genetic simulated annealing hybrid algorithm (PGSAHA) and applied to solve task scheduling problem in grid computing. It first generates a new group of individuals through genetic operation such as reproduction, crossover, mutation, etc, and than simulated anneals independently all the generated individuals respectively. When the temperature in the process of cooling no longer falls, the result is the optimal solution on the whole. From the analysis and experiment result, it is concluded that this algorithm is superior to genetic algorithm and simulated annealing.展开更多
The rise of multi-cloud systems has been spurred.For safety-critical missions,it is important to guarantee their security and reliability.To address trust constraints in a heterogeneous multi-cloud environment,this wo...The rise of multi-cloud systems has been spurred.For safety-critical missions,it is important to guarantee their security and reliability.To address trust constraints in a heterogeneous multi-cloud environment,this work proposes a novel scheduling method called matching and multi-round allocation(MMA)to optimize the makespan and total cost for all submitted tasks subject to security and reliability constraints.The method is divided into two phases for task scheduling.The first phase is to find the best matching candidate resources for the tasks to meet their preferential demands including performance,security,and reliability in a multi-cloud environment;the second one iteratively performs multiple rounds of re-allocating to optimize tasks execution time and cost by minimizing the variance of the estimated completion time.The proposed algorithm,the modified cuckoo search(MCS),hybrid chaotic particle search(HCPS),modified artificial bee colony(MABC),max-min,and min-min algorithms are implemented in CloudSim to create simulations.The simulations and experimental results show that our proposed method achieves shorter makespan,lower cost,higher resource utilization,and better trade-off between time and economic cost.It is more stable and efficient.展开更多
An increasing number of enterprises have adopted cloud computing to manage their important business applications in distributed green cloud(DGC)systems for low response time and high cost-effectiveness in recent years...An increasing number of enterprises have adopted cloud computing to manage their important business applications in distributed green cloud(DGC)systems for low response time and high cost-effectiveness in recent years.Task scheduling and resource allocation in DGCs have gained more attention in both academia and industry as they are costly to manage because of high energy consumption.Many factors in DGCs,e.g.,prices of power grid,and the amount of green energy express strong spatial variations.The dramatic increase of arriving tasks brings a big challenge to minimize the energy cost of a DGC provider in a market where above factors all possess spatial variations.This work adopts a G/G/1 queuing system to analyze the performance of servers in DGCs.Based on it,a single-objective constrained optimization problem is formulated and solved by a proposed simulated-annealing-based bees algorithm(SBA)to find SBA can minimize the energy cost of a DGC provider by optimally allocating tasks of heterogeneous applications among multiple DGCs,and specifying the running speed of each server and the number of powered-on servers in each GC while strictly meeting response time limits of tasks of all applications.Realistic databased experimental results prove that SBA achieves lower energy cost than several benchmark scheduling methods do.展开更多
In this paper,we consider a multi-UAV surveillance scenario where a team of unmanned aerial vehicles(UAVs)synchronously covers an area for monitoring the ground conditions.In this scenario,we adopt the leader-follower...In this paper,we consider a multi-UAV surveillance scenario where a team of unmanned aerial vehicles(UAVs)synchronously covers an area for monitoring the ground conditions.In this scenario,we adopt the leader-follower control mode and propose a modified Lyapunov guidance vector field(LGVF)approach for improving the precision of surveillance trajectory tracking.Then,in order to adopt to poor communication conditions,we propose a prediction-based synchronization method for keeping the formation consistently.Moreover,in order to adapt the multi-UAV system to dynamic and uncertain environment,this paper proposes a hierarchical dynamic task scheduling architecture.In this architecture,we firstly classify all the algorithms that perform tasks according to their functions,and then modularize the algorithms based on plugin technology.Afterwards,integrating the behavior model and plugin technique,this paper designs a three-layer control flow,which can efficiently achieve dynamic task scheduling.In order to verify the effectiveness of our architecture,we consider a multi-UAV traffic monitoring scenario and design several cases to demonstrate the online adjustment from three levels,respectively.展开更多
How to effectively reduce the energy consumption of large-scale data centers is a key issue in cloud computing. This paper presents a novel low-power task scheduling algorithm (L3SA) for large-scale cloud data cente...How to effectively reduce the energy consumption of large-scale data centers is a key issue in cloud computing. This paper presents a novel low-power task scheduling algorithm (L3SA) for large-scale cloud data centers. The winner tree is introduced to make the data nodes as the leaf nodes of the tree and the final winner on the purpose of reducing energy consumption is selected. The complexity of large-scale cloud data centers is fully consider, and the task comparson coefficient is defined to make task scheduling strategy more reasonable. Experiments and performance analysis show that the proposed algorithm can effectively improve the node utilization, and reduce the overall power consumption of the cloud data center.展开更多
A scheduling algorithm is presented aiming at the task scheduling problem in the phased array radar. Rather than assuming the scheduling interval(SI) time, which is the update interval of the radar invoking the schedu...A scheduling algorithm is presented aiming at the task scheduling problem in the phased array radar. Rather than assuming the scheduling interval(SI) time, which is the update interval of the radar invoking the scheduling algorithm, to be a fixed value,it is modeled as a fuzzy set to improve the scheduling flexibility.The scheduling algorithm exploits the fuzzy set model in order to intelligently adjust the SI time. The idle time in other SIs is provided for SIs which will be overload. Thereby more request tasks can be accommodated. The simulation results show that the proposed algorithm improves the successful scheduling ratio by 16%,the threat ratio of execution by 16% and the time utilization ratio by 15% compared with the highest task mode priority first(HPF)algorithm.展开更多
Cloud computing represents a novel computing model in the contemporary technology world. In a cloud system, the com- puting power of virtual machines (VMs) and network status can greatly affect the completion time o...Cloud computing represents a novel computing model in the contemporary technology world. In a cloud system, the com- puting power of virtual machines (VMs) and network status can greatly affect the completion time of data intensive tasks. How- ever, most of the current resource allocation policies focus only on network conditions and physical hosts. And the computing power of VMs is largely ignored. This paper proposes a comprehensive resource allocation policy which consists of a data intensive task scheduling algorithm that takes account of computing power of VMs and a VM allocation policy that considers bandwidth between storage nodes and hosts. The VM allocation policy includes VM placement and VM migration algorithms. Related simulations show that the proposed algorithms can greatly reduce the task comple- tion time and keep good load balance of physical hosts at the same time.展开更多
Cloud computing infrastructure has been evolving as a cost-effective platform for providing computational resources in the form of high-performance computing as a service(HPCaaS)to users for executing HPC applications...Cloud computing infrastructure has been evolving as a cost-effective platform for providing computational resources in the form of high-performance computing as a service(HPCaaS)to users for executing HPC applications.However,the broader use of the Cloud services,the rapid increase in the size,and the capacity of Cloud data centers bring a remarkable rise in energy consumption leading to a significant rise in the system provider expenses and carbon emissions in the environment.Besides this,users have become more demanding in terms of Quality-of-service(QoS)expectations in terms of execution time,budget cost,utilization,and makespan.This situation calls for the design of task scheduling policy,which ensures efficient task sequencing and allocation of computing resources to tasks to meet the trade-off between QoS promises and service provider requirements.Moreover,the task scheduling in the Cloud is a prevalent NP-Hard problem.Motivated by these concerns,this paper introduces and implements a QoS-aware Energy-Efficient Scheduling policy called as CSPSO,for scheduling tasks in Cloud systems to reduce the energy consumption of cloud resources and minimize the makespan of workload.The proposed multi-objective CSPSO policy hybridizes the search qualities of two robust metaheuristics viz.cuckoo search(CS)and particle swarm optimization(PSO)to overcome the slow convergence and lack of diversity of standard CS algorithm.A fitness-aware resource allocation(FARA)heuristic was developed and used by the proposed policy to allocate resources to tasks efficiently.A velocity update mechanism for cuckoo individuals is designed and incorporated in the proposed CSPSO policy.Further,the proposed scheduling policy has been implemented in the CloudSim simulator and tested with real supercomputing workload traces.The comparative analysis validated that the proposed scheduling policy can produce efficient schedules with better performance over other well-known heuristics and meta-heuristics scheduling policies.展开更多
基金supported by ZTE‑University‑Institute Fund Project under Grant No.IA20230629009.
文摘This paper reviews task scheduling frameworks,methods,and evaluation metrics of central processing unit-graphics processing unit(CPU-GPU)heterogeneous clusters.Task scheduling of CPU-GPU heterogeneous clusters can be carried out on the system level,nodelevel,and device level.Most task-scheduling technologies are heuristic based on the experts’experience,while some technologies are based on statistic methods using machine learning,deep learning,or reinforcement learning.Many metrics have been adopted to evaluate and compare different task scheduling technologies that try to optimize different goals of task scheduling.Although statistic task scheduling has reached fewer research achievements than heuristic task scheduling,the statistic task scheduling still has significant research potential.
基金in part by the Hubei Natural Science and Research Project under Grant 2020418in part by the 2021 Light of Taihu Science and Technology Projectin part by the 2022 Wuxi Science and Technology Innovation and Entrepreneurship Program.
文摘More devices in the Intelligent Internet of Things(AIoT)result in an increased number of tasks that require low latency and real-time responsiveness,leading to an increased demand for computational resources.Cloud computing’s low-latency performance issues in AIoT scenarios have led researchers to explore fog computing as a complementary extension.However,the effective allocation of resources for task execution within fog environments,characterized by limitations and heterogeneity in computational resources,remains a formidable challenge.To tackle this challenge,in this study,we integrate fog computing and cloud computing.We begin by establishing a fog-cloud environment framework,followed by the formulation of a mathematical model for task scheduling.Lastly,we introduce an enhanced hybrid Equilibrium Optimizer(EHEO)tailored for AIoT task scheduling.The overarching objective is to decrease both the makespan and energy consumption of the fog-cloud system while accounting for task deadlines.The proposed EHEO method undergoes a thorough evaluation against multiple benchmark algorithms,encompassing metrics likemakespan,total energy consumption,success rate,and average waiting time.Comprehensive experimental results unequivocally demonstrate the superior performance of EHEO across all assessed metrics.Notably,in the most favorable conditions,EHEO significantly diminishes both the makespan and energy consumption by approximately 50%and 35.5%,respectively,compared to the secondbest performing approach,which affirms its efficacy in advancing the efficiency of AIoT task scheduling within fog-cloud networks.
基金This work was supported by the National Key Research and Development Program of China(2021YFB2900603)the National Natural Science Foundation of China(61831008).
文摘A dynamic multi-beam resource allocation algorithm for large low Earth orbit(LEO)constellation based on on-board distributed computing is proposed in this paper.The allocation is a combinatorial optimization process under a series of complex constraints,which is important for enhancing the matching between resources and requirements.A complex algorithm is not available because that the LEO on-board resources is limi-ted.The proposed genetic algorithm(GA)based on two-dimen-sional individual model and uncorrelated single paternal inheri-tance method is designed to support distributed computation to enhance the feasibility of on-board application.A distributed system composed of eight embedded devices is built to verify the algorithm.A typical scenario is built in the system to evalu-ate the resource allocation process,algorithm mathematical model,trigger strategy,and distributed computation architec-ture.According to the simulation and measurement results,the proposed algorithm can provide an allocation result for more than 1500 tasks in 14 s and the success rate is more than 91%in a typical scene.The response time is decreased by 40%com-pared with the conditional GA.
基金Supported by Scientific and Technological Innovation Project of Chongqing(No.cstc2021jxjl20010)The Graduate Student Innovation Program of Chongqing University of Technology(No.clgycx-20203166,No.gzlcx20222061,No.gzlcx20223229)。
文摘Deploying service nodes hierarchically at the edge of the network can effectively improve the service quality of offloaded task requests and increase the utilization of resources.In this paper,we study the task scheduling problem in the hierarchically deployed edge cloud.We first formulate the minimization of the service time of scheduled tasks in edge cloud as a combinatorial optimization problem,blue and then prove the NP-hardness of the problem.Different from the existing work that mostly designs heuristic approximation-based algorithms or policies to make scheduling decision,we propose a newly designed scheduling policy,named Joint Neural Network and Heuristic Scheduling(JNNHSP),which combines a neural network-based method with a heuristic based solution.JNNHSP takes the Sequence-to-Sequence(Seq2Seq)model trained by Reinforcement Learning(RL)as the primary policy and adopts the heuristic algorithm as the auxiliary policy to obtain the scheduling solution,thereby achieving a good balance between the quality and the efficiency of the scheduling solution.In-depth experiments show that compared with a variety of related policies and optimization solvers,JNNHSP can achieve better performance in terms of scheduling error ratio,the degree to which the policy is affected by re-sources limitations,average service latency,and execution efficiency in a typical hierarchical edge cloud.
文摘Cloud computing has taken over the high-performance distributed computing area,and it currently provides on-demand services and resource polling over the web.As a result of constantly changing user service demand,the task scheduling problem has emerged as a critical analytical topic in cloud computing.The primary goal of scheduling tasks is to distribute tasks to available processors to construct the shortest possible schedule without breaching precedence restrictions.Assignments and schedules of tasks substantially influence system operation in a heterogeneous multiprocessor system.The diverse processes inside the heuristic-based task scheduling method will result in varying makespan in the heterogeneous computing system.As a result,an intelligent scheduling algorithm should efficiently determine the priority of every subtask based on the resources necessary to lower the makespan.This research introduced a novel efficient scheduling task method in cloud computing systems based on the cooperation search algorithm to tackle an essential task and schedule a heterogeneous cloud computing problem.The basic idea of thismethod is to use the advantages of meta-heuristic algorithms to get the optimal solution.We assess our algorithm’s performance by running it through three scenarios with varying numbers of tasks.The findings demonstrate that the suggested technique beats existingmethods NewGenetic Algorithm(NGA),Genetic Algorithm(GA),Whale Optimization Algorithm(WOA),Gravitational Search Algorithm(GSA),and Hybrid Heuristic and Genetic(HHG)by 7.9%,2.1%,8.8%,7.7%,3.4%respectively according to makespan.
基金supported by Science and Technology Development Foundation of the Central Guiding Local under Grant No.YDZJSX2021A038the National Natural Science Foundation of China under Grant No.61806138China University Industry-University-Research Collaborative Innovation Fund (Future Network Innovation Research and Application Project)under Grant No.2021FNA04014.
文摘Cloud computing technology is favored by users because of its strong computing power and convenient services.At the same time,scheduling performance has an extremely efficient impact on promoting carbon neutrality.Currently,scheduling research in the multi-cloud environment aims to address the challenges brought by business demands to cloud data centers during peak hours.Therefore,the scheduling problem has promising application prospects under themulti-cloud environment.This paper points out that the currently studied scheduling problems in the multi-cloud environment mainly include independent task scheduling and workflow task scheduling based on the dependencies between tasks.This paper reviews the concepts,types,objectives,advantages,challenges,and research status of task scheduling in the multi-cloud environment.Task scheduling strategies proposed in the existing related references are analyzed,discussed,and summarized,including research motivation,optimization algorithm,and related objectives.Finally,the research status of the two kinds of task scheduling is compared,and several future important research directions of multi-cloud task scheduling are proposed.
基金supported by theCommunication Soft Science Program of Ministry of Industry and Information Technology of China (No.2022-R-43)the Natural Science Basic Research Program of Shaanxi (No.2021JQ-719)Graduate Innovation Fund of Xi’an University of Posts and Telecommunications (No.CXJJZL2021014).
文摘Deploying task caching at edge servers has become an effectiveway to handle compute-intensive and latency-sensitive tasks on the industrialinternet. However, how to select the task scheduling location to reduce taskdelay and cost while ensuring the data security and reliable communicationof edge computing remains a challenge. To solve this problem, this paperestablishes a task scheduling model with joint blockchain and task cachingin the industrial internet and designs a novel blockchain-assisted cachingmechanism to enhance system security. In this paper, the task schedulingproblem, which couples the task scheduling decision, task caching decision,and blockchain reward, is formulated as the minimum weighted cost problemunder delay constraints. This is a mixed integer nonlinear problem, which isproved to be nonconvex and NP-hard. To solve the optimal solution, thispaper proposes a task scheduling strategy algorithm based on an improvedgenetic algorithm (IGA-TSPA) by improving the genetic algorithm initializationand mutation operations to reduce the size of the initial solutionspace and enhance the optimal solution convergence speed. In addition,an Improved Least Frequently Used algorithm is proposed to improve thecontent hit rate. Simulation results show that IGA-TSPA has a faster optimalsolution-solving ability and shorter running time compared with the existingedge computing scheduling algorithms. The established task scheduling modelnot only saves 62.19% of system overhead consumption in comparison withlocal computing but also has great significance in protecting data security,reducing task processing delay, and reducing system cost.
基金supported by National Natural Science Foundation of China(Grant No.61806138)the Central Government Guides Local Science and Technology Development Funds(Grant No.YDZJSX2021A038)+2 种基金Key RD Program of Shanxi Province(International Cooperation)under Grant No.201903D421048Outstanding Innovation Project for Graduate Students of Taiyuan University of Science and Technology(Project No.XCX211004)China University Industry-University-Research Collaborative Innovation Fund(Future Network Innovation Research and Application Project)(Grant 2021FNA04014).
文摘Due to the security and scalability features of hybrid cloud architecture,it can bettermeet the diverse requirements of users for cloud services.And a reasonable resource allocation solution is the key to adequately utilize the hybrid cloud.However,most previous studies have not comprehensively optimized the performance of hybrid cloud task scheduling,even ignoring the conflicts between its security privacy features and other requirements.Based on the above problems,a many-objective hybrid cloud task scheduling optimization model(HCTSO)is constructed combining risk rate,resource utilization,total cost,and task completion time.Meanwhile,an opposition-based learning knee point-driven many-objective evolutionary algorithm(OBL-KnEA)is proposed to improve the performance of model solving.The algorithm uses opposition-based learning to generate initial populations for faster convergence.Furthermore,a perturbation-based multipoint crossover operator and a dynamic range mutation operator are designed to extend the search range.By comparing the experiments with other excellent algorithms on HCTSO,OBL-KnEA achieves excellent results in terms of evaluation metrics,initial populations,and model optimization effects.
文摘Owing to massive technological developments in Internet of Things(IoT)and cloud environment,cloud computing(CC)offers a highly flexible heterogeneous resource pool over the network,and clients could exploit various resources on demand.Since IoT-enabled models are restricted to resources and require crisp response,minimum latency,and maximum bandwidth,which are outside the capabilities.CC was handled as a resource-rich solution to aforementioned challenge.As high delay reduces the performance of the IoT enabled cloud platform,efficient utilization of task scheduling(TS)reduces the energy usage of the cloud infrastructure and increases the income of service provider via minimizing processing time of user job.Therefore,this article concentration on the design of an oppositional red fox optimization based task scheduling scheme(ORFOTSS)for IoT enabled cloud environment.The presented ORFO-TSS model resolves the problem of allocating resources from the IoT based cloud platform.It achieves the makespan by performing optimum TS procedures with various aspects of incoming task.The designing of ORFO-TSS method includes the idea of oppositional based learning(OBL)as to traditional RFO approach in enhancing their efficiency.A wide-ranging experimental analysis was applied on the CloudSim platform.The experimental outcome highlighted the efficacy of the ORFO-TSS technique over existing approaches.
基金supported in part by the National Key R&D Program of China under Grant 2017YFB1302400the Jinan“20 New Colleges and Universities”Funded Scientific Research Leader Studio under Grant 2021GXRC079+2 种基金the Major Agricultural Applied Technological Innovation Projects of Shandong Province underGrant SD2019NJ014the Shandong Natural Science Foundation under Grant ZR2019MF064the Beijing Advanced Innovation Center for Intelligent Robots and Systems under Grant 2019IRS19.
文摘The solution strategy of the heuristic algorithm is pre-set and has good performance in the conventional cloud resource scheduling process.However,for complex and dynamic cloud service scheduling tasks,due to the difference in service attributes,the solution efficiency of a single strategy is low for such problems.In this paper,we presents a hyper-heuristic algorithm based on reinforcement learning(HHRL)to optimize the completion time of the task sequence.Firstly,In the reward table setting stage of HHRL,we introduce population diversity and integrate maximum time to comprehensively deter-mine the task scheduling and the selection of low-level heuristic strategies.Secondly,a task computational complexity estimation method integrated with linear regression is proposed to influence task scheduling priorities.Besides,we propose a high-quality candidate solution migration method to ensure the continuity and diversity of the solving process.Compared with HHSA,ACO,GA,F-PSO,etc,HHRL can quickly obtain task complexity,select appropriate heuristic strategies for task scheduling,search for the the best makspan and have stronger disturbance detection ability for population diversity.
文摘In a cloud environment,consumers search for the best service provider that accomplishes the required tasks based on a set of criteria such as completion time and cost.On the other hand,Cloud Service Providers(CSPs)seek to maximize their profits by attracting and serving more consumers based on their resource capabilities.The literature has discussed the problem by considering either consumers’needs or CSPs’capabilities.A problem resides in the lack of explicit models that combine preferences of consumers with the capabilities of CSPs to provide a unified process for resource allocation and task scheduling in a more efficient way.The paper proposes a model that adopts a Multi-Criteria Decision Making(MCDM)method,called Analytic Hierarchy Process(AHP),to acquire the information of consumers’preferences and service providers’capabilities to prioritize both tasks and resources.The model also provides a matching technique to assign each task to the best resource of a CSP while preserves the fairness of scheduling more tasks for resources with higher capabilities.Our experimental results prove the feasibility of the proposed model for prioritizing hundreds of tasks/services and CSPs based on a defined set of criteria,and matching each set of tasks/services to the best CSPS.
文摘To solve the deadlock problem of tasks that the interdependence between tasks fails to consider during the course of resource assignment and task scheduling based on the heuristics algorithm, an improved ant colony system (ACS) based algorithm is proposed. First, how to map the resource assignment and task scheduling (RATS) problem into the optimization selection problem of task resource assignment graph (TRAG) and to add the semaphore mechanism in the optimal TRAG to solve deadlocks are explained. Secondly, how to utilize the grid pheromone system model to realize the algorithm based on ACS is explicated. This refers to the construction of TRAG by the random selection of appropriate resources for each task by the user agent and the optimization of TRAG through the positive feedback and distributed parallel computing mechanism of the ACS. Simulation results show that the proposed algorithm is effective and efficient in solving the deadlock problem.
基金Supported by the National Basic ResearchProgramof China (973 Program2003CB314804)
文摘In this paper combined with the advantages of genetic algorithm and simulated annealing, brings forward a parallel genetic simulated annealing hybrid algorithm (PGSAHA) and applied to solve task scheduling problem in grid computing. It first generates a new group of individuals through genetic operation such as reproduction, crossover, mutation, etc, and than simulated anneals independently all the generated individuals respectively. When the temperature in the process of cooling no longer falls, the result is the optimal solution on the whole. From the analysis and experiment result, it is concluded that this algorithm is superior to genetic algorithm and simulated annealing.
基金supported in part by the National Natural Science Foundation of China(61673123,61603100)in part by the Natural Science Foundation of Guangdong Province,China(2020A151501482)。
文摘The rise of multi-cloud systems has been spurred.For safety-critical missions,it is important to guarantee their security and reliability.To address trust constraints in a heterogeneous multi-cloud environment,this work proposes a novel scheduling method called matching and multi-round allocation(MMA)to optimize the makespan and total cost for all submitted tasks subject to security and reliability constraints.The method is divided into two phases for task scheduling.The first phase is to find the best matching candidate resources for the tasks to meet their preferential demands including performance,security,and reliability in a multi-cloud environment;the second one iteratively performs multiple rounds of re-allocating to optimize tasks execution time and cost by minimizing the variance of the estimated completion time.The proposed algorithm,the modified cuckoo search(MCS),hybrid chaotic particle search(HCPS),modified artificial bee colony(MABC),max-min,and min-min algorithms are implemented in CloudSim to create simulations.The simulations and experimental results show that our proposed method achieves shorter makespan,lower cost,higher resource utilization,and better trade-off between time and economic cost.It is more stable and efficient.
基金supported in part by the National Natural Science Foundation of China(61802015,61703011)the Major Science and Technology Program for Water Pollution Control and Treatment of China(2018ZX07111005)+1 种基金the National Defense Pre-Research Foundation of China(41401020401,41401050102)the Deanship of Scientific Research(DSR)at King Abdulaziz University,Jeddah(D-422-135-1441)。
文摘An increasing number of enterprises have adopted cloud computing to manage their important business applications in distributed green cloud(DGC)systems for low response time and high cost-effectiveness in recent years.Task scheduling and resource allocation in DGCs have gained more attention in both academia and industry as they are costly to manage because of high energy consumption.Many factors in DGCs,e.g.,prices of power grid,and the amount of green energy express strong spatial variations.The dramatic increase of arriving tasks brings a big challenge to minimize the energy cost of a DGC provider in a market where above factors all possess spatial variations.This work adopts a G/G/1 queuing system to analyze the performance of servers in DGCs.Based on it,a single-objective constrained optimization problem is formulated and solved by a proposed simulated-annealing-based bees algorithm(SBA)to find SBA can minimize the energy cost of a DGC provider by optimally allocating tasks of heterogeneous applications among multiple DGCs,and specifying the running speed of each server and the number of powered-on servers in each GC while strictly meeting response time limits of tasks of all applications.Realistic databased experimental results prove that SBA achieves lower energy cost than several benchmark scheduling methods do.
基金Project(2017YFB1301104)supported by the National Key Research and Development Program of ChinaProjects(61906212,61802426)supported by the National Natural Science Foundation of China。
文摘In this paper,we consider a multi-UAV surveillance scenario where a team of unmanned aerial vehicles(UAVs)synchronously covers an area for monitoring the ground conditions.In this scenario,we adopt the leader-follower control mode and propose a modified Lyapunov guidance vector field(LGVF)approach for improving the precision of surveillance trajectory tracking.Then,in order to adopt to poor communication conditions,we propose a prediction-based synchronization method for keeping the formation consistently.Moreover,in order to adapt the multi-UAV system to dynamic and uncertain environment,this paper proposes a hierarchical dynamic task scheduling architecture.In this architecture,we firstly classify all the algorithms that perform tasks according to their functions,and then modularize the algorithms based on plugin technology.Afterwards,integrating the behavior model and plugin technique,this paper designs a three-layer control flow,which can efficiently achieve dynamic task scheduling.In order to verify the effectiveness of our architecture,we consider a multi-UAV traffic monitoring scenario and design several cases to demonstrate the online adjustment from three levels,respectively.
基金supported by the National Natural Science Foundation of China(6120200461272084)+9 种基金the National Key Basic Research Program of China(973 Program)(2011CB302903)the Specialized Research Fund for the Doctoral Program of Higher Education(2009322312000120113223110003)the China Postdoctoral Science Foundation Funded Project(2011M5000952012T50514)the Natural Science Foundation of Jiangsu Province(BK2011754BK2009426)the Jiangsu Postdoctoral Science Foundation Funded Project(1102103C)the Natural Science Fund of Higher Education of Jiangsu Province(12KJB520007)the Project Funded by the Priority Academic Program Development of Jiangsu Higher Education Institutions(yx002001)
文摘How to effectively reduce the energy consumption of large-scale data centers is a key issue in cloud computing. This paper presents a novel low-power task scheduling algorithm (L3SA) for large-scale cloud data centers. The winner tree is introduced to make the data nodes as the leaf nodes of the tree and the final winner on the purpose of reducing energy consumption is selected. The complexity of large-scale cloud data centers is fully consider, and the task comparson coefficient is defined to make task scheduling strategy more reasonable. Experiments and performance analysis show that the proposed algorithm can effectively improve the node utilization, and reduce the overall power consumption of the cloud data center.
基金supported by the National Youth Foundation(61503408)
文摘A scheduling algorithm is presented aiming at the task scheduling problem in the phased array radar. Rather than assuming the scheduling interval(SI) time, which is the update interval of the radar invoking the scheduling algorithm, to be a fixed value,it is modeled as a fuzzy set to improve the scheduling flexibility.The scheduling algorithm exploits the fuzzy set model in order to intelligently adjust the SI time. The idle time in other SIs is provided for SIs which will be overload. Thereby more request tasks can be accommodated. The simulation results show that the proposed algorithm improves the successful scheduling ratio by 16%,the threat ratio of execution by 16% and the time utilization ratio by 15% compared with the highest task mode priority first(HPF)algorithm.
基金supported by the National Natural Science Foundation of China(6120235461272422)the Scientific and Technological Support Project(Industry)of Jiangsu Province(BE2011189)
文摘Cloud computing represents a novel computing model in the contemporary technology world. In a cloud system, the com- puting power of virtual machines (VMs) and network status can greatly affect the completion time of data intensive tasks. How- ever, most of the current resource allocation policies focus only on network conditions and physical hosts. And the computing power of VMs is largely ignored. This paper proposes a comprehensive resource allocation policy which consists of a data intensive task scheduling algorithm that takes account of computing power of VMs and a VM allocation policy that considers bandwidth between storage nodes and hosts. The VM allocation policy includes VM placement and VM migration algorithms. Related simulations show that the proposed algorithms can greatly reduce the task comple- tion time and keep good load balance of physical hosts at the same time.
文摘Cloud computing infrastructure has been evolving as a cost-effective platform for providing computational resources in the form of high-performance computing as a service(HPCaaS)to users for executing HPC applications.However,the broader use of the Cloud services,the rapid increase in the size,and the capacity of Cloud data centers bring a remarkable rise in energy consumption leading to a significant rise in the system provider expenses and carbon emissions in the environment.Besides this,users have become more demanding in terms of Quality-of-service(QoS)expectations in terms of execution time,budget cost,utilization,and makespan.This situation calls for the design of task scheduling policy,which ensures efficient task sequencing and allocation of computing resources to tasks to meet the trade-off between QoS promises and service provider requirements.Moreover,the task scheduling in the Cloud is a prevalent NP-Hard problem.Motivated by these concerns,this paper introduces and implements a QoS-aware Energy-Efficient Scheduling policy called as CSPSO,for scheduling tasks in Cloud systems to reduce the energy consumption of cloud resources and minimize the makespan of workload.The proposed multi-objective CSPSO policy hybridizes the search qualities of two robust metaheuristics viz.cuckoo search(CS)and particle swarm optimization(PSO)to overcome the slow convergence and lack of diversity of standard CS algorithm.A fitness-aware resource allocation(FARA)heuristic was developed and used by the proposed policy to allocate resources to tasks efficiently.A velocity update mechanism for cuckoo individuals is designed and incorporated in the proposed CSPSO policy.Further,the proposed scheduling policy has been implemented in the CloudSim simulator and tested with real supercomputing workload traces.The comparative analysis validated that the proposed scheduling policy can produce efficient schedules with better performance over other well-known heuristics and meta-heuristics scheduling policies.