When the communication time is relatively shorter than the computation time for every task, the task duplication based scheduling (TDS) algorithm proposed by Darbha and Agrawal generates an optimal schedule. Park and ...When the communication time is relatively shorter than the computation time for every task, the task duplication based scheduling (TDS) algorithm proposed by Darbha and Agrawal generates an optimal schedule. Park and Choe also proposed an extended TDS algorithm whose optimality condition is less restricted than that of TDS algorithm, but the condition is very complex and is difficult to satisfy when the number of tasks is large. An efficient algorithm is proposed whose optimality condition is less restricted and simpler than both of the algorithms, and the schedule length is also shorter than both of the algorithms. The time complexity of the proposed algorithm is O(v2), where v represents the number of tasks.展开更多
Well organized datacentres with interconnected servers constitute the cloud computing infrastructure.User requests are submitted through an interface to these servers that provide service to them in an on-demand basis...Well organized datacentres with interconnected servers constitute the cloud computing infrastructure.User requests are submitted through an interface to these servers that provide service to them in an on-demand basis.The scientific applications that get executed at cloud by making use of the heterogeneous resources being allocated to them in a dynamic manner are grouped under NP hard problem category.Task scheduling in cloud poses numerous challenges impacting the cloud performance.If not handled properly,user satisfaction becomes questionable.More recently researchers had come up with meta-heuristic type of solutions for enriching the task scheduling activity in the cloud environment.The prime aim of task scheduling is to utilize the resources available in an optimal manner and reduce the time span of task execution.An improvised seagull optimization algorithm which combines the features of the Cuckoo search(CS)and seagull optimization algorithm(SOA)had been proposed in this work to enhance the performance of the scheduling activity inside the cloud computing environment.The proposed algorithm aims to minimize the cost and time parameters that are spent during task scheduling in the heterogeneous cloud environment.Performance evaluation of the proposed algorithm had been performed using the Cloudsim 3.0 toolkit by comparing it with Multi objective-Ant Colony Optimization(MO-ACO),ACO and Min-Min algorithms.The proposed SOA-CS technique had produced an improvement of 1.06%,4.2%,and 2.4%for makespan and had reduced the overall cost to the extent of 1.74%,3.93%and 2.77%when compared with PSO,ACO,IDEA algorithms respectively when 300 vms are considered.The comparative simulation results obtained had shown that the proposed improvised seagull optimization algorithm fares better than other contemporaries.展开更多
The solution strategy of the heuristic algorithm is pre-set and has good performance in the conventional cloud resource scheduling process.However,for complex and dynamic cloud service scheduling tasks,due to the diff...The solution strategy of the heuristic algorithm is pre-set and has good performance in the conventional cloud resource scheduling process.However,for complex and dynamic cloud service scheduling tasks,due to the difference in service attributes,the solution efficiency of a single strategy is low for such problems.In this paper,we presents a hyper-heuristic algorithm based on reinforcement learning(HHRL)to optimize the completion time of the task sequence.Firstly,In the reward table setting stage of HHRL,we introduce population diversity and integrate maximum time to comprehensively deter-mine the task scheduling and the selection of low-level heuristic strategies.Secondly,a task computational complexity estimation method integrated with linear regression is proposed to influence task scheduling priorities.Besides,we propose a high-quality candidate solution migration method to ensure the continuity and diversity of the solving process.Compared with HHSA,ACO,GA,F-PSO,etc,HHRL can quickly obtain task complexity,select appropriate heuristic strategies for task scheduling,search for the the best makspan and have stronger disturbance detection ability for population diversity.展开更多
Due to the security and scalability features of hybrid cloud architecture,it can bettermeet the diverse requirements of users for cloud services.And a reasonable resource allocation solution is the key to adequately u...Due to the security and scalability features of hybrid cloud architecture,it can bettermeet the diverse requirements of users for cloud services.And a reasonable resource allocation solution is the key to adequately utilize the hybrid cloud.However,most previous studies have not comprehensively optimized the performance of hybrid cloud task scheduling,even ignoring the conflicts between its security privacy features and other requirements.Based on the above problems,a many-objective hybrid cloud task scheduling optimization model(HCTSO)is constructed combining risk rate,resource utilization,total cost,and task completion time.Meanwhile,an opposition-based learning knee point-driven many-objective evolutionary algorithm(OBL-KnEA)is proposed to improve the performance of model solving.The algorithm uses opposition-based learning to generate initial populations for faster convergence.Furthermore,a perturbation-based multipoint crossover operator and a dynamic range mutation operator are designed to extend the search range.By comparing the experiments with other excellent algorithms on HCTSO,OBL-KnEA achieves excellent results in terms of evaluation metrics,initial populations,and model optimization effects.展开更多
Task scheduling for electro-magnetic detection satellite is a typical combinatorial optimization problem. The count of constraints that need to be taken into account is of large scale. An algorithm combined integer pr...Task scheduling for electro-magnetic detection satellite is a typical combinatorial optimization problem. The count of constraints that need to be taken into account is of large scale. An algorithm combined integer programming with constraint programming is presented. This algorithm is deployed in this problem through two steps. The first step is to decompose the original problem into master and sub-problem using the logic-based Benders decomposition; then a circus combines master and sub-problem solving process together, and the connection between them is general Benders cut. This hybrid algorithm is tested by a set of derived experiments. The result is compared with corresponding outcomes generated by the strength Pareto evolutionary algorithm and the pure constraint programming solver GECODE, which is an open source software. These tests and comparisons yield promising effect.展开更多
Task scheduling is the main problem in cloud computing that reduces system performance;it is an important way to arrange user needs and perform multiple goals.Cloud computing is the most popular technology nowadays an...Task scheduling is the main problem in cloud computing that reduces system performance;it is an important way to arrange user needs and perform multiple goals.Cloud computing is the most popular technology nowadays and has many research potential in various areas like resource allocation,task scheduling,security,privacy,etc.To improve system performance,an efficient task-scheduling algorithm is required.Existing task-scheduling algorithms focus on task-resource requirements,CPU memory,execution time,and execution cost.In this paper,a task scheduling algorithm based on a Genetic Algorithm(GA)has been presented for assigning and executing different tasks.The proposed algorithm aims to minimize both the completion time and execution cost of tasks and maximize resource utilization.We evaluate our algorithm’s performance by applying it to two examples with a different number of tasks and processors.The first example contains ten tasks and four processors;the computation costs are generated randomly.The last example has eight processors,and the number of tasks ranges from twenty to seventy;the computation cost of each task on different processors is generated randomly.The achieved results show that the proposed approach significantly succeeded in finding the optimal solutions for the three objectives;completion time,execution cost,and resource utilization.展开更多
To cope with the task scheduling problem under multi-task and transportation consideration in large-scale service oriented manufacturing systems(SOMS), a service allocation optimization mathematical model was establis...To cope with the task scheduling problem under multi-task and transportation consideration in large-scale service oriented manufacturing systems(SOMS), a service allocation optimization mathematical model was established, and then a hybrid discrete particle swarm optimization-genetic algorithm(HDPSOGA) was proposed. In SOMS, each resource involved in the whole life cycle of a product, whether it is provided by a piece of software or a hardware device, is encapsulated into a service. So, the transportation during production of a task should be taken into account because the hard-services selected are possibly provided by various providers in different areas. In the service allocation optimization mathematical model, multi-task and transportation were considered simultaneously. In the proposed HDPSOGA algorithm, integer coding method was applied to establish the mapping between the particle location matrix and the service allocation scheme. The position updating process was performed according to the cognition part, the social part, and the previous velocity and position while introducing the crossover and mutation idea of genetic algorithm to fit the discrete space. Finally, related simulation experiments were carried out to compare with other two previous algorithms. The results indicate the effectiveness and efficiency of the proposed hybrid algorithm.展开更多
An increasing number of enterprises have adopted cloud computing to manage their important business applications in distributed green cloud(DGC)systems for low response time and high cost-effectiveness in recent years...An increasing number of enterprises have adopted cloud computing to manage their important business applications in distributed green cloud(DGC)systems for low response time and high cost-effectiveness in recent years.Task scheduling and resource allocation in DGCs have gained more attention in both academia and industry as they are costly to manage because of high energy consumption.Many factors in DGCs,e.g.,prices of power grid,and the amount of green energy express strong spatial variations.The dramatic increase of arriving tasks brings a big challenge to minimize the energy cost of a DGC provider in a market where above factors all possess spatial variations.This work adopts a G/G/1 queuing system to analyze the performance of servers in DGCs.Based on it,a single-objective constrained optimization problem is formulated and solved by a proposed simulated-annealing-based bees algorithm(SBA)to find SBA can minimize the energy cost of a DGC provider by optimally allocating tasks of heterogeneous applications among multiple DGCs,and specifying the running speed of each server and the number of powered-on servers in each GC while strictly meeting response time limits of tasks of all applications.Realistic databased experimental results prove that SBA achieves lower energy cost than several benchmark scheduling methods do.展开更多
AI(Artificial Intelligence)workloads are proliferating in modernreal-time systems.As the tasks of AI workloads fluctuate over time,resourceplanning policies used for traditional fixed real-time tasks should be reexami...AI(Artificial Intelligence)workloads are proliferating in modernreal-time systems.As the tasks of AI workloads fluctuate over time,resourceplanning policies used for traditional fixed real-time tasks should be reexamined.In particular,it is difficult to immediately handle changes inreal-time tasks without violating the deadline constraints.To cope with thissituation,this paper analyzes the task situations of AI workloads and findsthe following two observations.First,resource planning for AI workloadsis a complicated search problem that requires much time for optimization.Second,although the task set of an AI workload may change over time,thepossible combinations of the task sets are known in advance.Based on theseobservations,this paper proposes a new resource planning scheme for AIworkloads that supports the re-planning of resources.Instead of generatingresource plans on the fly,the proposed scheme pre-determines resourceplans for various combinations of tasks.Thus,in any case,the workload isimmediately executed according to the resource plan maintained.Specifically,the proposed scheme maintains an optimized CPU(Central Processing Unit)and memory resource plan using genetic algorithms and applies it as soonas the workload changes.The proposed scheme is implemented in the opensourcesimulator SimRTS for the validation of its effectiveness.Simulationexperiments show that the proposed scheme reduces the energy consumptionof CPU and memory by 45.5%on average without deadline misses.展开更多
More devices in the Intelligent Internet of Things(AIoT)result in an increased number of tasks that require low latency and real-time responsiveness,leading to an increased demand for computational resources.Cloud com...More devices in the Intelligent Internet of Things(AIoT)result in an increased number of tasks that require low latency and real-time responsiveness,leading to an increased demand for computational resources.Cloud computing’s low-latency performance issues in AIoT scenarios have led researchers to explore fog computing as a complementary extension.However,the effective allocation of resources for task execution within fog environments,characterized by limitations and heterogeneity in computational resources,remains a formidable challenge.To tackle this challenge,in this study,we integrate fog computing and cloud computing.We begin by establishing a fog-cloud environment framework,followed by the formulation of a mathematical model for task scheduling.Lastly,we introduce an enhanced hybrid Equilibrium Optimizer(EHEO)tailored for AIoT task scheduling.The overarching objective is to decrease both the makespan and energy consumption of the fog-cloud system while accounting for task deadlines.The proposed EHEO method undergoes a thorough evaluation against multiple benchmark algorithms,encompassing metrics likemakespan,total energy consumption,success rate,and average waiting time.Comprehensive experimental results unequivocally demonstrate the superior performance of EHEO across all assessed metrics.Notably,in the most favorable conditions,EHEO significantly diminishes both the makespan and energy consumption by approximately 50%and 35.5%,respectively,compared to the secondbest performing approach,which affirms its efficacy in advancing the efficiency of AIoT task scheduling within fog-cloud networks.展开更多
In cloud computing,a lot of challenges like the server failures,loss of confidentiality,improper workloads,etc.are still bounding the efficiency of cloud systems in real-world scenarios.For this reason,many research w...In cloud computing,a lot of challenges like the server failures,loss of confidentiality,improper workloads,etc.are still bounding the efficiency of cloud systems in real-world scenarios.For this reason,many research works are being performed to overcome the shortcoming of existing systems.Among them,load balancing seems to be the most critical issue that worsen the performance of the cloud sector,and hence there necessitates the optimal load balancing with optimal task scheduling.With the intention of accomplishing optimal load balancing by effectual task deployment,this paper plans to develop an advanced load balancing model with the assistance acquired from the metaheuristic algorithms.Usually,handling of tasks in cloud system is an NP-hard problem and moreover,nonpreemptive independent tasks are crucial in cloud computing.This paper goes with the introduction of a new optimal load balancing model by considering three major objectives:minimum makespan,priority,and load balancing,respectively.Moreover,a new single-objective function is also defined that incorporates all the three objectives mentioned above.Furthermore,the deployment of tasks must be optimal and for this a new hybrid optimization algorithm referred as Firefly Movement insistedWOA(FM-WOA)is introduced.This FM-WOA is the conceptual amalgamation of standard Whale Optimization Algorithm(WOA)and Firefly(FF)algorithm.Finally,the performances of the proposed FM-WOA model is compared over the conventional models with the intention of proving its efficiency in terms of makespan,task completion(priority),and degree of imbalance as well.展开更多
为了解决复杂任务群调度过程中资源利用不均、任务完成时间较长等问题,以最小化资源负载均方差和最小化任务群完成时间为目标构建复杂任务群资源调度模型,提出一种融合局部搜索和Pareto支配的多目标优化算法BRLSN(multi-objective optim...为了解决复杂任务群调度过程中资源利用不均、任务完成时间较长等问题,以最小化资源负载均方差和最小化任务群完成时间为目标构建复杂任务群资源调度模型,提出一种融合局部搜索和Pareto支配的多目标优化算法BRLSN(multi-objective optimization based on boundary range local search and NSGA-Ⅱ,BRLSN)。该算法采用有效的编码方式与交叉变异算子进行迭代寻优,并利用基于边界区域局部搜索的精英保留策略扩大算法搜索范围,保存种群优良个体。实验结果表明,BRLSN相较于其他多目标算法在收敛性和多样性上有显著的提升,同时算法收敛速度更快,种群质量更高,明显优化了最终目标函数的结果值。展开更多
文摘When the communication time is relatively shorter than the computation time for every task, the task duplication based scheduling (TDS) algorithm proposed by Darbha and Agrawal generates an optimal schedule. Park and Choe also proposed an extended TDS algorithm whose optimality condition is less restricted than that of TDS algorithm, but the condition is very complex and is difficult to satisfy when the number of tasks is large. An efficient algorithm is proposed whose optimality condition is less restricted and simpler than both of the algorithms, and the schedule length is also shorter than both of the algorithms. The time complexity of the proposed algorithm is O(v2), where v represents the number of tasks.
文摘Well organized datacentres with interconnected servers constitute the cloud computing infrastructure.User requests are submitted through an interface to these servers that provide service to them in an on-demand basis.The scientific applications that get executed at cloud by making use of the heterogeneous resources being allocated to them in a dynamic manner are grouped under NP hard problem category.Task scheduling in cloud poses numerous challenges impacting the cloud performance.If not handled properly,user satisfaction becomes questionable.More recently researchers had come up with meta-heuristic type of solutions for enriching the task scheduling activity in the cloud environment.The prime aim of task scheduling is to utilize the resources available in an optimal manner and reduce the time span of task execution.An improvised seagull optimization algorithm which combines the features of the Cuckoo search(CS)and seagull optimization algorithm(SOA)had been proposed in this work to enhance the performance of the scheduling activity inside the cloud computing environment.The proposed algorithm aims to minimize the cost and time parameters that are spent during task scheduling in the heterogeneous cloud environment.Performance evaluation of the proposed algorithm had been performed using the Cloudsim 3.0 toolkit by comparing it with Multi objective-Ant Colony Optimization(MO-ACO),ACO and Min-Min algorithms.The proposed SOA-CS technique had produced an improvement of 1.06%,4.2%,and 2.4%for makespan and had reduced the overall cost to the extent of 1.74%,3.93%and 2.77%when compared with PSO,ACO,IDEA algorithms respectively when 300 vms are considered.The comparative simulation results obtained had shown that the proposed improvised seagull optimization algorithm fares better than other contemporaries.
基金supported in part by the National Key R&D Program of China under Grant 2017YFB1302400the Jinan“20 New Colleges and Universities”Funded Scientific Research Leader Studio under Grant 2021GXRC079+2 种基金the Major Agricultural Applied Technological Innovation Projects of Shandong Province underGrant SD2019NJ014the Shandong Natural Science Foundation under Grant ZR2019MF064the Beijing Advanced Innovation Center for Intelligent Robots and Systems under Grant 2019IRS19.
文摘The solution strategy of the heuristic algorithm is pre-set and has good performance in the conventional cloud resource scheduling process.However,for complex and dynamic cloud service scheduling tasks,due to the difference in service attributes,the solution efficiency of a single strategy is low for such problems.In this paper,we presents a hyper-heuristic algorithm based on reinforcement learning(HHRL)to optimize the completion time of the task sequence.Firstly,In the reward table setting stage of HHRL,we introduce population diversity and integrate maximum time to comprehensively deter-mine the task scheduling and the selection of low-level heuristic strategies.Secondly,a task computational complexity estimation method integrated with linear regression is proposed to influence task scheduling priorities.Besides,we propose a high-quality candidate solution migration method to ensure the continuity and diversity of the solving process.Compared with HHSA,ACO,GA,F-PSO,etc,HHRL can quickly obtain task complexity,select appropriate heuristic strategies for task scheduling,search for the the best makspan and have stronger disturbance detection ability for population diversity.
基金supported by National Natural Science Foundation of China(Grant No.61806138)the Central Government Guides Local Science and Technology Development Funds(Grant No.YDZJSX2021A038)+2 种基金Key RD Program of Shanxi Province(International Cooperation)under Grant No.201903D421048Outstanding Innovation Project for Graduate Students of Taiyuan University of Science and Technology(Project No.XCX211004)China University Industry-University-Research Collaborative Innovation Fund(Future Network Innovation Research and Application Project)(Grant 2021FNA04014).
文摘Due to the security and scalability features of hybrid cloud architecture,it can bettermeet the diverse requirements of users for cloud services.And a reasonable resource allocation solution is the key to adequately utilize the hybrid cloud.However,most previous studies have not comprehensively optimized the performance of hybrid cloud task scheduling,even ignoring the conflicts between its security privacy features and other requirements.Based on the above problems,a many-objective hybrid cloud task scheduling optimization model(HCTSO)is constructed combining risk rate,resource utilization,total cost,and task completion time.Meanwhile,an opposition-based learning knee point-driven many-objective evolutionary algorithm(OBL-KnEA)is proposed to improve the performance of model solving.The algorithm uses opposition-based learning to generate initial populations for faster convergence.Furthermore,a perturbation-based multipoint crossover operator and a dynamic range mutation operator are designed to extend the search range.By comparing the experiments with other excellent algorithms on HCTSO,OBL-KnEA achieves excellent results in terms of evaluation metrics,initial populations,and model optimization effects.
基金supported by the National Security Fundamental Research Foundation of China (61361)the National Natural Science Foundation of China (61104180)
文摘Task scheduling for electro-magnetic detection satellite is a typical combinatorial optimization problem. The count of constraints that need to be taken into account is of large scale. An algorithm combined integer programming with constraint programming is presented. This algorithm is deployed in this problem through two steps. The first step is to decompose the original problem into master and sub-problem using the logic-based Benders decomposition; then a circus combines master and sub-problem solving process together, and the connection between them is general Benders cut. This hybrid algorithm is tested by a set of derived experiments. The result is compared with corresponding outcomes generated by the strength Pareto evolutionary algorithm and the pure constraint programming solver GECODE, which is an open source software. These tests and comparisons yield promising effect.
文摘Task scheduling is the main problem in cloud computing that reduces system performance;it is an important way to arrange user needs and perform multiple goals.Cloud computing is the most popular technology nowadays and has many research potential in various areas like resource allocation,task scheduling,security,privacy,etc.To improve system performance,an efficient task-scheduling algorithm is required.Existing task-scheduling algorithms focus on task-resource requirements,CPU memory,execution time,and execution cost.In this paper,a task scheduling algorithm based on a Genetic Algorithm(GA)has been presented for assigning and executing different tasks.The proposed algorithm aims to minimize both the completion time and execution cost of tasks and maximize resource utilization.We evaluate our algorithm’s performance by applying it to two examples with a different number of tasks and processors.The first example contains ten tasks and four processors;the computation costs are generated randomly.The last example has eight processors,and the number of tasks ranges from twenty to seventy;the computation cost of each task on different processors is generated randomly.The achieved results show that the proposed approach significantly succeeded in finding the optimal solutions for the three objectives;completion time,execution cost,and resource utilization.
基金Project(2012B091100444)supported by the Production,Education and Research Cooperative Program of Guangdong Province and Ministry of Education,ChinaProject(2013ZM0091)supported by Fundamental Research Funds for the Central Universities of China
文摘To cope with the task scheduling problem under multi-task and transportation consideration in large-scale service oriented manufacturing systems(SOMS), a service allocation optimization mathematical model was established, and then a hybrid discrete particle swarm optimization-genetic algorithm(HDPSOGA) was proposed. In SOMS, each resource involved in the whole life cycle of a product, whether it is provided by a piece of software or a hardware device, is encapsulated into a service. So, the transportation during production of a task should be taken into account because the hard-services selected are possibly provided by various providers in different areas. In the service allocation optimization mathematical model, multi-task and transportation were considered simultaneously. In the proposed HDPSOGA algorithm, integer coding method was applied to establish the mapping between the particle location matrix and the service allocation scheme. The position updating process was performed according to the cognition part, the social part, and the previous velocity and position while introducing the crossover and mutation idea of genetic algorithm to fit the discrete space. Finally, related simulation experiments were carried out to compare with other two previous algorithms. The results indicate the effectiveness and efficiency of the proposed hybrid algorithm.
基金supported in part by the National Natural Science Foundation of China(61802015,61703011)the Major Science and Technology Program for Water Pollution Control and Treatment of China(2018ZX07111005)+1 种基金the National Defense Pre-Research Foundation of China(41401020401,41401050102)the Deanship of Scientific Research(DSR)at King Abdulaziz University,Jeddah(D-422-135-1441)。
文摘An increasing number of enterprises have adopted cloud computing to manage their important business applications in distributed green cloud(DGC)systems for low response time and high cost-effectiveness in recent years.Task scheduling and resource allocation in DGCs have gained more attention in both academia and industry as they are costly to manage because of high energy consumption.Many factors in DGCs,e.g.,prices of power grid,and the amount of green energy express strong spatial variations.The dramatic increase of arriving tasks brings a big challenge to minimize the energy cost of a DGC provider in a market where above factors all possess spatial variations.This work adopts a G/G/1 queuing system to analyze the performance of servers in DGCs.Based on it,a single-objective constrained optimization problem is formulated and solved by a proposed simulated-annealing-based bees algorithm(SBA)to find SBA can minimize the energy cost of a DGC provider by optimally allocating tasks of heterogeneous applications among multiple DGCs,and specifying the running speed of each server and the number of powered-on servers in each GC while strictly meeting response time limits of tasks of all applications.Realistic databased experimental results prove that SBA achieves lower energy cost than several benchmark scheduling methods do.
基金This work was partly supported by the Institute of Information&communications Technology Planning&Evaluation(IITP)grant funded by theKorean government(MSIT)(No.2021-0-02068,Artificial Intelligence Innovation Hub)(No.RS-2022-00155966,Artificial Intelligence Convergence Innovation Human Resources Development(Ewha University)).
文摘AI(Artificial Intelligence)workloads are proliferating in modernreal-time systems.As the tasks of AI workloads fluctuate over time,resourceplanning policies used for traditional fixed real-time tasks should be reexamined.In particular,it is difficult to immediately handle changes inreal-time tasks without violating the deadline constraints.To cope with thissituation,this paper analyzes the task situations of AI workloads and findsthe following two observations.First,resource planning for AI workloadsis a complicated search problem that requires much time for optimization.Second,although the task set of an AI workload may change over time,thepossible combinations of the task sets are known in advance.Based on theseobservations,this paper proposes a new resource planning scheme for AIworkloads that supports the re-planning of resources.Instead of generatingresource plans on the fly,the proposed scheme pre-determines resourceplans for various combinations of tasks.Thus,in any case,the workload isimmediately executed according to the resource plan maintained.Specifically,the proposed scheme maintains an optimized CPU(Central Processing Unit)and memory resource plan using genetic algorithms and applies it as soonas the workload changes.The proposed scheme is implemented in the opensourcesimulator SimRTS for the validation of its effectiveness.Simulationexperiments show that the proposed scheme reduces the energy consumptionof CPU and memory by 45.5%on average without deadline misses.
基金in part by the Hubei Natural Science and Research Project under Grant 2020418in part by the 2021 Light of Taihu Science and Technology Projectin part by the 2022 Wuxi Science and Technology Innovation and Entrepreneurship Program.
文摘More devices in the Intelligent Internet of Things(AIoT)result in an increased number of tasks that require low latency and real-time responsiveness,leading to an increased demand for computational resources.Cloud computing’s low-latency performance issues in AIoT scenarios have led researchers to explore fog computing as a complementary extension.However,the effective allocation of resources for task execution within fog environments,characterized by limitations and heterogeneity in computational resources,remains a formidable challenge.To tackle this challenge,in this study,we integrate fog computing and cloud computing.We begin by establishing a fog-cloud environment framework,followed by the formulation of a mathematical model for task scheduling.Lastly,we introduce an enhanced hybrid Equilibrium Optimizer(EHEO)tailored for AIoT task scheduling.The overarching objective is to decrease both the makespan and energy consumption of the fog-cloud system while accounting for task deadlines.The proposed EHEO method undergoes a thorough evaluation against multiple benchmark algorithms,encompassing metrics likemakespan,total energy consumption,success rate,and average waiting time.Comprehensive experimental results unequivocally demonstrate the superior performance of EHEO across all assessed metrics.Notably,in the most favorable conditions,EHEO significantly diminishes both the makespan and energy consumption by approximately 50%and 35.5%,respectively,compared to the secondbest performing approach,which affirms its efficacy in advancing the efficiency of AIoT task scheduling within fog-cloud networks.
文摘In cloud computing,a lot of challenges like the server failures,loss of confidentiality,improper workloads,etc.are still bounding the efficiency of cloud systems in real-world scenarios.For this reason,many research works are being performed to overcome the shortcoming of existing systems.Among them,load balancing seems to be the most critical issue that worsen the performance of the cloud sector,and hence there necessitates the optimal load balancing with optimal task scheduling.With the intention of accomplishing optimal load balancing by effectual task deployment,this paper plans to develop an advanced load balancing model with the assistance acquired from the metaheuristic algorithms.Usually,handling of tasks in cloud system is an NP-hard problem and moreover,nonpreemptive independent tasks are crucial in cloud computing.This paper goes with the introduction of a new optimal load balancing model by considering three major objectives:minimum makespan,priority,and load balancing,respectively.Moreover,a new single-objective function is also defined that incorporates all the three objectives mentioned above.Furthermore,the deployment of tasks must be optimal and for this a new hybrid optimization algorithm referred as Firefly Movement insistedWOA(FM-WOA)is introduced.This FM-WOA is the conceptual amalgamation of standard Whale Optimization Algorithm(WOA)and Firefly(FF)algorithm.Finally,the performances of the proposed FM-WOA model is compared over the conventional models with the intention of proving its efficiency in terms of makespan,task completion(priority),and degree of imbalance as well.
文摘为了解决复杂任务群调度过程中资源利用不均、任务完成时间较长等问题,以最小化资源负载均方差和最小化任务群完成时间为目标构建复杂任务群资源调度模型,提出一种融合局部搜索和Pareto支配的多目标优化算法BRLSN(multi-objective optimization based on boundary range local search and NSGA-Ⅱ,BRLSN)。该算法采用有效的编码方式与交叉变异算子进行迭代寻优,并利用基于边界区域局部搜索的精英保留策略扩大算法搜索范围,保存种群优良个体。实验结果表明,BRLSN相较于其他多目标算法在收敛性和多样性上有显著的提升,同时算法收敛速度更快,种群质量更高,明显优化了最终目标函数的结果值。