Unbalanced traffic distribution in cellular networks results in congestion and degrades spectrum efficiency.To tackle this problem,we propose an Unmanned Aerial Vehicle(UAV)-assisted wireless network in which the UAV ...Unbalanced traffic distribution in cellular networks results in congestion and degrades spectrum efficiency.To tackle this problem,we propose an Unmanned Aerial Vehicle(UAV)-assisted wireless network in which the UAV acts as an aerial relay to divert some traffic from the overloaded cell to its adjacent underloaded cell.To fully exploit its potential,we jointly optimize the UAV position,user association,spectrum allocation,and power allocation to maximize the sum-log-rate of all users in two adjacent cells.To tackle the complicated joint optimization problem,we first design a genetic-based algorithm to optimize the UAV position.Then,we simplify the problem by theoretical analysis and devise a low-complexity algorithm according to the branch-and-bound method,so as to obtain the optimal user association and spectrum allocation schemes.We further propose an iterative power allocation algorithm based on the sequential convex approximation theory.The simulation results indicate that the proposed UAV-assisted wireless network is superior to the terrestrial network in both utility and throughput,and the proposed algorithms can substantially improve the network performance in comparison with the other schemes.展开更多
This paper focuses on the scheduling problem of workflow tasks that exhibit interdependencies.Unlike indepen-dent batch tasks,workflows typically consist of multiple subtasks with intrinsic correlations and dependenci...This paper focuses on the scheduling problem of workflow tasks that exhibit interdependencies.Unlike indepen-dent batch tasks,workflows typically consist of multiple subtasks with intrinsic correlations and dependencies.It necessitates the distribution of various computational tasks to appropriate computing node resources in accor-dance with task dependencies to ensure the smooth completion of the entire workflow.Workflow scheduling must consider an array of factors,including task dependencies,availability of computational resources,and the schedulability of tasks.Therefore,this paper delves into the distributed graph database workflow task scheduling problem and proposes a workflow scheduling methodology based on deep reinforcement learning(DRL).The method optimizes the maximum completion time(makespan)and response time of workflow tasks,aiming to enhance the responsiveness of workflow tasks while ensuring the minimization of the makespan.The experimental results indicate that the Q-learning Deep Reinforcement Learning(Q-DRL)algorithm markedly diminishes the makespan and refines the average response time within distributed graph database environments.In quantifying makespan,Q-DRL achieves mean reductions of 12.4%and 11.9%over established First-fit and Random scheduling strategies,respectively.Additionally,Q-DRL surpasses the performance of both DRL-Cloud and Improved Deep Q-learning Network(IDQN)algorithms,with improvements standing at 4.4%and 2.6%,respectively.With reference to average response time,the Q-DRL approach exhibits a significantly enhanced performance in the scheduling of workflow tasks,decreasing the average by 2.27%and 4.71%when compared to IDQN and DRL-Cloud,respectively.The Q-DRL algorithm also demonstrates a notable increase in the efficiency of system resource utilization,reducing the average idle rate by 5.02%and 9.30%in comparison to IDQN and DRL-Cloud,respectively.These findings support the assertion that Q-DRL not only upholds a lower average idle rate but also effectively curtails the average response time,thereby substantially improving processing efficiency and optimizing resource utilization within distributed graph database systems.展开更多
Cloud Computing has the ability to provide on-demand access to a shared resource pool.It has completely changed the way businesses are managed,implement applications,and provide services.The rise in popularity has led...Cloud Computing has the ability to provide on-demand access to a shared resource pool.It has completely changed the way businesses are managed,implement applications,and provide services.The rise in popularity has led to a significant increase in the user demand for services.However,in cloud environments efficient load balancing is essential to ensure optimal performance and resource utilization.This systematic review targets a detailed description of load balancing techniques including static and dynamic load balancing algorithms.Specifically,metaheuristic-based dynamic load balancing algorithms are identified as the optimal solution in case of increased traffic.In a cloud-based context,this paper describes load balancing measurements,including the benefits and drawbacks associated with the selected load balancing techniques.It also summarizes the algorithms based on implementation,time complexity,adaptability,associated issue(s),and targeted QoS parameters.Additionally,the analysis evaluates the tools and instruments utilized in each investigated study.Moreover,comparative analysis among static,traditional dynamic and metaheuristic algorithms based on response time by using the CloudSim simulation tool is also performed.Finally,the key open problems and potential directions for the state-of-the-art metaheuristic-based approaches are also addressed.展开更多
With the continuous expansion of the data center network scale, changing network requirements, and increasing pressure on network bandwidth, the traditional network architecture can no longer meet people’s needs. The...With the continuous expansion of the data center network scale, changing network requirements, and increasing pressure on network bandwidth, the traditional network architecture can no longer meet people’s needs. The development of software defined networks has brought new opportunities and challenges to future networks. The data and control separation characteristics of SDN improve the performance of the entire network. Researchers have integrated SDN architecture into data centers to improve network resource utilization and performance. This paper first introduces the basic concepts of SDN and data center networks. Then it discusses SDN-based load balancing mechanisms for data centers from different perspectives. Finally, it summarizes and looks forward to the study on SDN-based load balancing mechanisms and its development trend.展开更多
In this paper, a sender-initiated protocol is applied which uses fuzzy logic control method to improve computer networks performance by balancing loads among computers. This new model devises sender-initiated protocol...In this paper, a sender-initiated protocol is applied which uses fuzzy logic control method to improve computer networks performance by balancing loads among computers. This new model devises sender-initiated protocol for load transfer for load balancing. Groups are formed and every group has a node called a designated representative (DR). During load transferring processes, loads are transferred using the DR in each group to achieve load balancing purposes. The simulation results show that the performance of the protocol proposed is better than the compared conventional method. This protocol is more stable than the method without using the fuzzy logic control.展开更多
At present,the flow table of the SDN switch is stored in the costly Ternary Content Addressable Memory(TCAM)cache.Due to the cost problem,the number of flow tables that the SDN switch can store is extremely limited,wh...At present,the flow table of the SDN switch is stored in the costly Ternary Content Addressable Memory(TCAM)cache.Due to the cost problem,the number of flow tables that the SDN switch can store is extremely limited,which is far less than the number of traffic,so it is prone to overflow problem,and leads to network paralysis.That has become a bottleneck in restricting the processing capacity of the data center,and will become a weak point focused by attackers.In this paper,we propose an algorithm for the Alarm Switch Remove(ASR)that fully loads the flow table space in SDN,and further put forward an integrated load balancing scheme in SDN.Finally,we use Mininet to verify that the scheme can ease the SDN switch flow table overflow problem and increase network throughput.展开更多
Integrating the blockchain technology into mobile-edge computing(MEC)networks with multiple cooperative MEC servers(MECS)providing a promising solution to improving resource utilization,and helping establish a secure ...Integrating the blockchain technology into mobile-edge computing(MEC)networks with multiple cooperative MEC servers(MECS)providing a promising solution to improving resource utilization,and helping establish a secure reward mechanism that can facilitate load balancing among MECS.In addition,intelligent management of service caching and load balancing can improve the network utility in MEC blockchain networks with multiple types of workloads.In this paper,we investigate a learningbased joint service caching and load balancing policy for optimizing the communication and computation resources allocation,so as to improve the resource utilization of MEC blockchain networks.We formulate the problem as a challenging long-term network revenue maximization Markov decision process(MDP)problem.To address the highly dynamic and high dimension of system states,we design a joint service caching and load balancing algorithm based on the double-dueling Deep Q network(DQN)approach.The simulation results validate the feasibility and superior performance of our proposed algorithm over several baseline schemes.展开更多
This paper presents a novel fuzzy firefly-based intelligent algorithm for load balancing in mobile cloud computing while reducing makespan.The proposed technique implicitly acts intelligently by using inherent traits ...This paper presents a novel fuzzy firefly-based intelligent algorithm for load balancing in mobile cloud computing while reducing makespan.The proposed technique implicitly acts intelligently by using inherent traits of fuzzy and firefly.It automatically adjusts its behavior or converges depending on the information gathered during the search process and objective function.It works for 3-tier architecture,including cloudlet and public cloud.As cloudlets have limited resources,fuzzy logic is used for cloudlet selection using capacity and waiting time as input.Fuzzy provides human-like decisions without using any mathematical model.Firefly is a powerful meta-heuristic optimization technique to balance diversification and solution speed.It balances the load on cloud and cloudlet while minimizing makespan and execution time.However,it may trap in local optimum;levy flight can handle it.Hybridization of fuzzy fireflywith levy flight is a novel technique that provides reduced makespan,execution time,and Degree of imbalance while balancing the load.Simulation has been carried out on the Cloud Analyst platform with National Aeronautics and Space Administration(NASA)and Clarknet datasets.Results show that the proposed algorithm outperforms Ant Colony Optimization Queue Decision Maker(ACOQDM),Distributed Scheduling Optimization Algorithm(DSOA),andUtility-based Firefly Algorithm(UFA)when compared in terms of makespan,Degree of imbalance,and Figure of Merit.展开更多
A major challenge for the future wireless network is to design the self-organizing architecture.The reactive self-organizing model of traditional networks needs to be transformed into an active self-organizing network...A major challenge for the future wireless network is to design the self-organizing architecture.The reactive self-organizing model of traditional networks needs to be transformed into an active self-organizing network.Due to the user mobility and the coverage of small cells,the network load often becomes unbalanced,resulting in poor network performance.Mobility management has become an important issue to ensure seamless communication when users move between cells,and proactive mobility management is one of the important functions of the active Self-Organizing Network(SON).This paper proposes a proactive mobility management framework for active SON,which transforms the original reactive load balancing into a forward-aware and proactive load balancing.The proposed framework firstly uses the BART model to predict the users’temporal and spatial mobility based on a weekly cycle and then formulate the MLB optimization problem based on the soft load.Two solutions are proposed to solve the above MLB problem.The simulation results show that the proposed method can better optimize the network performance and realize intelligent mobile management for the future network.展开更多
According to the advances in users’service requirements,physical hardware accessibility,and speed of resource delivery,Cloud Computing(CC)is an essential technology to be used in many fields.Moreover,the Internet of ...According to the advances in users’service requirements,physical hardware accessibility,and speed of resource delivery,Cloud Computing(CC)is an essential technology to be used in many fields.Moreover,the Internet of Things(IoT)is employed for more communication flexibility and richness that are required to obtain fruitful services.A multi-agent system might be a proper solution to control the load balancing of interaction and communication among agents.This paper proposes a multi-agent load balancing framework that consists of two phases to optimize the workload among different servers with large-scale CC power with various utilities and a significant number of IoT devices with low resources.Different agents are integrated based on relevant features of behavioral interaction using classification techniques to balance the workload.Aload balancing algorithm is developed to serve users’requests to improve the solution of workload problems with an efficient distribution.The activity task from IoT devices has been classified by feature selection methods in the preparatory phase to optimize the scalability ofCC.Then,the server’s availability is checked and the classified task is assigned to its suitable server in the main phase to enhance the cloud environment performance.Multi-agent load balancing framework is succeeded to cope with the importance of using large-scale requirements of CC and(low resources and large number)of IoT.展开更多
Healthcare is a fundamental part of every individual’s life.The healthcare industry is developing very rapidly with the help of advanced technologies.Many researchers are trying to build cloud-based healthcare applic...Healthcare is a fundamental part of every individual’s life.The healthcare industry is developing very rapidly with the help of advanced technologies.Many researchers are trying to build cloud-based healthcare applications that can be accessed by healthcare professionals from their premises,as well as by patients from their mobile devices through communication interfaces.These systems promote reliable and remote interactions between patients and healthcare professionals.However,there are several limitations to these innovative cloud computing-based systems,namely network availability,latency,battery life and resource availability.We propose a hybrid mobile cloud computing(HMCC)architecture to address these challenges.Furthermore,we also evaluate the performance of heuristic and dynamic machine learning based task scheduling and load balancing algorithms on our proposed architecture.We compare them,to identify the strengths and weaknesses of each algorithm;and provide their comparative results,to show latency and energy consumption performance.Challenging issues for cloudbased healthcare systems are discussed in detail.展开更多
Recently,the fundamental problem with Hybrid Mobile Ad-hoc Net-works(H-MANETs)is tofind a suitable and secure way of balancing the load through Internet gateways.Moreover,the selection of the gateway and overload of th...Recently,the fundamental problem with Hybrid Mobile Ad-hoc Net-works(H-MANETs)is tofind a suitable and secure way of balancing the load through Internet gateways.Moreover,the selection of the gateway and overload of the network results in packet loss and Delay(DL).For optimal performance,it is important to load balance between different gateways.As a result,a stable load balancing procedure is implemented,which selects gateways based on Fuzzy Logic(FL)and increases the efficiency of the network.In this case,since gate-ways are selected based on the number of nodes,the Energy Consumption(EC)was high.This paper presents a novel Node Quality-based Clustering Algo-rithm(NQCA)based on Fuzzy-Genetic for Cluster Head and Gateway Selection(FGCHGS).This algorithm combines NQCA with the Improved Weighted Clus-tering Algorithm(IWCA).The NQCA algorithm divides the network into clusters based upon node priority,transmission range,and neighbourfidelity.In addition,the simulation results tend to evaluate the performance effectiveness of the FFFCHGS algorithm in terms of EC,packet loss rate(PLR),etc.展开更多
Every day,more and more data is being produced by the Internet of Things(IoT)applications.IoT data differ in amount,diversity,veracity,and velocity.Because of latency,various types of data handling in cloud computing ...Every day,more and more data is being produced by the Internet of Things(IoT)applications.IoT data differ in amount,diversity,veracity,and velocity.Because of latency,various types of data handling in cloud computing are not suitable for many time-sensitive applications.When users move from one site to another,mobility also adds to the latency.By placing computing close to IoT devices with mobility support,fog computing addresses these problems.An efficient Load Balancing Algorithm(LBA)improves user experience and Quality of Service(QoS).Classification of Request(CoR)based Resource Adaptive LBA is suggested in this research.This technique clusters fog nodes using an efficient K-means clustering algorithm and then uses a Decision Tree approach to categorize the request.The decision-making process for time-sensitive and delay-tolerable requests is facilitated by the classification of requests.LBA does the operation based on these classifications.The MobFogSim simulation program is utilized to assess how well the algorithm with mobility features performs.The outcome demonstrates that the LBA algorithm’s performance enhances the total system performance,which was attained by(90.8%).Using LBA,several metrics may be examined,including Response Time(RT),delay(d),Energy Consumption(EC),and latency.Through the on-demand provisioning of necessary resources to IoT users,our suggested LBA assures effective resource usage.展开更多
Advancements in cloud computing and virtualization technologies have revolutionized Enterprise Application Development with innovative ways to design and develop complex systems.Microservices Architecture is one of th...Advancements in cloud computing and virtualization technologies have revolutionized Enterprise Application Development with innovative ways to design and develop complex systems.Microservices Architecture is one of the recent techniques in which Enterprise Systems can be developed as fine-grained smaller components and deployed independently.This methodology brings numerous benefits like scalability,resilience,flexibility in development,faster time to market,etc.and the advantages;Microservices bring some challenges too.Multiple microservices need to be invoked one by one as a chain.In most applications,more than one chain of microservices runs in parallel to complete a particular requirement To complete a user’s request.It results in competition for resources and the need for more inter-service communication among the services,which increases the overall latency of the application.A new approach has been proposed in this paper to handle a complex chain of microservices and reduce the latency of user requests.A machine learning technique is followed to predict the weighting time of different types of requests.The communication time among services distributed among different physical machines are estimated based on that and obtained insights are applied to an algorithm to calculate their priorities dynamically and select suitable service instances to minimize the latency based on the shortest queue waiting time.Experiments were done for both interactive as well as non interactive workloads to test the effectiveness of the solution.The approach has been proved to be very effective in reducing latency in the case of long service chains.展开更多
Cloud Technology is a new platform that offers on-demand computing Peripheral such as storage,processing power,and other computer system resources.It is also referred to as a system that will let the consumers utilize...Cloud Technology is a new platform that offers on-demand computing Peripheral such as storage,processing power,and other computer system resources.It is also referred to as a system that will let the consumers utilize computational resources like databases,servers,storage,and intelligence over the Internet.In a cloud network,load balancing is the process of dividing network traffic among a cluster of available servers to increase efficiency.It is also known as a server pool or server farm.When a single node is overwhelmed,balancing the workload is needed to manage unpredictable workflows.The load balancer sends the load to another free node in this case.We focus on the Balancing of workflows with the proposed approach,and we present a novel method to balance the load that manages the dynamic scheduling process.One of the preexisting load balancing techniques is considered,however it is somewhat modified to fit the scenario at hand.Depending on the experimentation’s findings,it is concluded that this suggested approach improves load balancing consistency,response time,and throughput by 6%.展开更多
In wireless sensor network(WSN),the gateways which are placed far away from the base station(BS)forward the collected data to the BS through the gateways which are nearer to the BS.This leads to more energy consumptio...In wireless sensor network(WSN),the gateways which are placed far away from the base station(BS)forward the collected data to the BS through the gateways which are nearer to the BS.This leads to more energy consumption because the gateways nearer to the BS manages heavy traffic load.So,to over-come this issue,loads around the gateways are to be balanced by presenting energy efficient clustering approach.Besides,to enhance the lifetime of the net-work,optimal routing path is to be established between the source node and BS.For energy efficient load balancing and routing,multi objective based beetle swarm optimization(BSO)algorithm is presented in this paper.Using this algo-rithm,optimal clustering and routing are performed depend on the objective func-tions routingfitness and clusteringfitness.This approach leads to decrease the power consumption.Simulation results show that the performance of the pro-posed BSO based clustering and routing scheme attains better results than that of the existing algorithms in terms of energy consumption,delivery ratio,through-put and network lifetime.Namely,the proposed scheme increases throughput to 72%and network lifetime to 37%as well as it reduces delay to 37%than the existing optimization algorithms based clustering and routing schemes.展开更多
A low-Earth-orbit(LEO)satellite network can provide full-coverage access services worldwide and is an essential candidate for future 6G networking.However,the large variability of the geographic distribution of the Ea...A low-Earth-orbit(LEO)satellite network can provide full-coverage access services worldwide and is an essential candidate for future 6G networking.However,the large variability of the geographic distribution of the Earth’s population leads to an uneven service volume distribution of access service.Moreover,the limitations on the resources of satellites are far from being able to serve the traffic in hotspot areas.To enhance the forwarding capability of satellite networks,we first assess how hotspot areas under different load cases and spatial scales significantly affect the network throughput of an LEO satellite network overall.Then,we propose a multi-region cooperative traffic scheduling algorithm.The algorithm migrates low-grade traffic from hotspot areas to coldspot areas for forwarding,significantly increasing the overall throughput of the satellite network while sacrificing some latency of end-to-end forwarding.This algorithm can utilize all the global satellite resources and improve the utilization of network resources.We model the cooperative multi-region scheduling of large-scale LEO satellites.Based on the model,we build a system testbed using OMNET++to compare the proposed method with existing techniques.The simulations show that our proposed method can reduce the packet loss probability by 30%and improve the resource utilization ratio by 3.69%.展开更多
Task scheduling plays a key role in effectively managing and allocating computing resources to meet various computing tasks in a cloud computing environment.Short execution time and low load imbalance may be the chall...Task scheduling plays a key role in effectively managing and allocating computing resources to meet various computing tasks in a cloud computing environment.Short execution time and low load imbalance may be the challenges for some algorithms in resource scheduling scenarios.In this work,the Hierarchical Particle Swarm Optimization-Evolutionary Artificial Bee Colony Algorithm(HPSO-EABC)has been proposed,which hybrids our presented Evolutionary Artificial Bee Colony(EABC),and Hierarchical Particle Swarm Optimization(HPSO)algorithm.The HPSO-EABC algorithm incorporates both the advantages of the HPSO and the EABC algorithm.Comprehensive testing including evaluations of algorithm convergence speed,resource execution time,load balancing,and operational costs has been done.The results indicate that the EABC algorithm exhibits greater parallelism compared to the Artificial Bee Colony algorithm.Compared with the Particle Swarm Optimization algorithm,the HPSO algorithmnot only improves the global search capability but also effectively mitigates getting stuck in local optima.As a result,the hybrid HPSO-EABC algorithm demonstrates significant improvements in terms of stability and convergence speed.Moreover,it exhibits enhanced resource scheduling performance in both homogeneous and heterogeneous environments,effectively reducing execution time and cost,which also is verified by the ablation experimental.展开更多
To solve the load balancing problem in a triplet-based hierarchical interconnection network(THIN) system, a dynamic load balancing (DLB)algorithm--THINDLBA, which adopts multicast tree (MT)technology to improve ...To solve the load balancing problem in a triplet-based hierarchical interconnection network(THIN) system, a dynamic load balancing (DLB)algorithm--THINDLBA, which adopts multicast tree (MT)technology to improve the efficiency of interchanging load information, is presented. To support the algorithm, a complete set of DLB messages and a schema of maintaining DLB information in each processing node are designed. The load migration request messages from the heavily loaded node (HLN)are spread along an MT whose root is the HLN. And the lightly loaded nodes(LLNs) covered by the MT are the candidate destinations of load migration; the load information interchanged between the LLNs and the HLN can be transmitted along the MT. So the HLN can migrate excess loads out as many as possible during a one time execution of the THINDLBA, and its load state can be improved as quickly as possible. To avoid wrongly transmitted or redundant DLB messages due to MT overlapping, the MT construction is restricted in the design of the THINDLBA. Through experiments, the effectiveness of four DLB algorithms are compared, and the results show that the THINDLBA can effectively decrease the time costs of THIN systems in dealing with large scale computeintensive tasks more than others.展开更多
This paper focuses on solving a problem of improving system robustness and the efficiency of a distributed system at the same time. Fault tolerance with active replication and load balancing techniques are used. The p...This paper focuses on solving a problem of improving system robustness and the efficiency of a distributed system at the same time. Fault tolerance with active replication and load balancing techniques are used. The pros and cons of both techniques are analyzed, and a novel load balancing framework for fault tolerant systems with active replication is presented. Hierarchical architecture is described in detail. The framework can dynamically adjust fault tolerant groups and their memberships with respect to system loads. Three potential task scheduler group selection methods are proposed and simulation tests are made. Further analysis of test data is done and helpful observations for system design are also pointed out, including effects of task arrival intensity and task set size, relationship between total task execution time and single task execution time.展开更多
基金supported in part by the National Key Research and Development Program of China under Grant 2020YFB1807003in part by the National Natural Science Foundation of China under Grants 61901381,62171385,and 61901378+3 种基金in part by the Aeronautical Science Foundation of China under Grant 2020z073053004in part by the Foundation of the State Key Laboratory of Integrated Services Networks of Xidian University under Grant ISN21-06in part by the Key Research Program and Industrial Innovation Chain Project of Shaanxi Province under Grant 2019ZDLGY07-10in part by the Natural Science Fundamental Research Program of Shaanxi Province under Grant 2021JM-069.
文摘Unbalanced traffic distribution in cellular networks results in congestion and degrades spectrum efficiency.To tackle this problem,we propose an Unmanned Aerial Vehicle(UAV)-assisted wireless network in which the UAV acts as an aerial relay to divert some traffic from the overloaded cell to its adjacent underloaded cell.To fully exploit its potential,we jointly optimize the UAV position,user association,spectrum allocation,and power allocation to maximize the sum-log-rate of all users in two adjacent cells.To tackle the complicated joint optimization problem,we first design a genetic-based algorithm to optimize the UAV position.Then,we simplify the problem by theoretical analysis and devise a low-complexity algorithm according to the branch-and-bound method,so as to obtain the optimal user association and spectrum allocation schemes.We further propose an iterative power allocation algorithm based on the sequential convex approximation theory.The simulation results indicate that the proposed UAV-assisted wireless network is superior to the terrestrial network in both utility and throughput,and the proposed algorithms can substantially improve the network performance in comparison with the other schemes.
基金funded by the Science and Technology Foundation of State Grid Corporation of China(Grant No.5108-202218280A-2-397-XG).
文摘This paper focuses on the scheduling problem of workflow tasks that exhibit interdependencies.Unlike indepen-dent batch tasks,workflows typically consist of multiple subtasks with intrinsic correlations and dependencies.It necessitates the distribution of various computational tasks to appropriate computing node resources in accor-dance with task dependencies to ensure the smooth completion of the entire workflow.Workflow scheduling must consider an array of factors,including task dependencies,availability of computational resources,and the schedulability of tasks.Therefore,this paper delves into the distributed graph database workflow task scheduling problem and proposes a workflow scheduling methodology based on deep reinforcement learning(DRL).The method optimizes the maximum completion time(makespan)and response time of workflow tasks,aiming to enhance the responsiveness of workflow tasks while ensuring the minimization of the makespan.The experimental results indicate that the Q-learning Deep Reinforcement Learning(Q-DRL)algorithm markedly diminishes the makespan and refines the average response time within distributed graph database environments.In quantifying makespan,Q-DRL achieves mean reductions of 12.4%and 11.9%over established First-fit and Random scheduling strategies,respectively.Additionally,Q-DRL surpasses the performance of both DRL-Cloud and Improved Deep Q-learning Network(IDQN)algorithms,with improvements standing at 4.4%and 2.6%,respectively.With reference to average response time,the Q-DRL approach exhibits a significantly enhanced performance in the scheduling of workflow tasks,decreasing the average by 2.27%and 4.71%when compared to IDQN and DRL-Cloud,respectively.The Q-DRL algorithm also demonstrates a notable increase in the efficiency of system resource utilization,reducing the average idle rate by 5.02%and 9.30%in comparison to IDQN and DRL-Cloud,respectively.These findings support the assertion that Q-DRL not only upholds a lower average idle rate but also effectively curtails the average response time,thereby substantially improving processing efficiency and optimizing resource utilization within distributed graph database systems.
文摘Cloud Computing has the ability to provide on-demand access to a shared resource pool.It has completely changed the way businesses are managed,implement applications,and provide services.The rise in popularity has led to a significant increase in the user demand for services.However,in cloud environments efficient load balancing is essential to ensure optimal performance and resource utilization.This systematic review targets a detailed description of load balancing techniques including static and dynamic load balancing algorithms.Specifically,metaheuristic-based dynamic load balancing algorithms are identified as the optimal solution in case of increased traffic.In a cloud-based context,this paper describes load balancing measurements,including the benefits and drawbacks associated with the selected load balancing techniques.It also summarizes the algorithms based on implementation,time complexity,adaptability,associated issue(s),and targeted QoS parameters.Additionally,the analysis evaluates the tools and instruments utilized in each investigated study.Moreover,comparative analysis among static,traditional dynamic and metaheuristic algorithms based on response time by using the CloudSim simulation tool is also performed.Finally,the key open problems and potential directions for the state-of-the-art metaheuristic-based approaches are also addressed.
文摘With the continuous expansion of the data center network scale, changing network requirements, and increasing pressure on network bandwidth, the traditional network architecture can no longer meet people’s needs. The development of software defined networks has brought new opportunities and challenges to future networks. The data and control separation characteristics of SDN improve the performance of the entire network. Researchers have integrated SDN architecture into data centers to improve network resource utilization and performance. This paper first introduces the basic concepts of SDN and data center networks. Then it discusses SDN-based load balancing mechanisms for data centers from different perspectives. Finally, it summarizes and looks forward to the study on SDN-based load balancing mechanisms and its development trend.
文摘In this paper, a sender-initiated protocol is applied which uses fuzzy logic control method to improve computer networks performance by balancing loads among computers. This new model devises sender-initiated protocol for load transfer for load balancing. Groups are formed and every group has a node called a designated representative (DR). During load transferring processes, loads are transferred using the DR in each group to achieve load balancing purposes. The simulation results show that the performance of the protocol proposed is better than the compared conventional method. This protocol is more stable than the method without using the fuzzy logic control.
基金supported supported by the National Key Research and Development Program of China(No.2020YFE0200500)CERNET Innovation Project(NGII20190806)。
文摘At present,the flow table of the SDN switch is stored in the costly Ternary Content Addressable Memory(TCAM)cache.Due to the cost problem,the number of flow tables that the SDN switch can store is extremely limited,which is far less than the number of traffic,so it is prone to overflow problem,and leads to network paralysis.That has become a bottleneck in restricting the processing capacity of the data center,and will become a weak point focused by attackers.In this paper,we propose an algorithm for the Alarm Switch Remove(ASR)that fully loads the flow table space in SDN,and further put forward an integrated load balancing scheme in SDN.Finally,we use Mininet to verify that the scheme can ease the SDN switch flow table overflow problem and increase network throughput.
基金supported in part by the National Natural Science Foundation of China 62072096the Fundamental Research Funds for the Central Universities under Grant 2232020A-12+4 种基金the International S&T Cooperation Program of Shanghai Science and Technology Commission under Grant 20220713000the Young Top-notch Talent Program in Shanghaithe"Shuguang Program"of Shanghai Education Development Foundation and Shanghai Municipal Education Commissionthe Fundamental Research Funds for the Central Universities and Graduate Student Innovation Fund of Donghua University CUSF-DH-D-2019093supported in part by the NSF under grants CNS-2107190 and ECCS-1923717。
文摘Integrating the blockchain technology into mobile-edge computing(MEC)networks with multiple cooperative MEC servers(MECS)providing a promising solution to improving resource utilization,and helping establish a secure reward mechanism that can facilitate load balancing among MECS.In addition,intelligent management of service caching and load balancing can improve the network utility in MEC blockchain networks with multiple types of workloads.In this paper,we investigate a learningbased joint service caching and load balancing policy for optimizing the communication and computation resources allocation,so as to improve the resource utilization of MEC blockchain networks.We formulate the problem as a challenging long-term network revenue maximization Markov decision process(MDP)problem.To address the highly dynamic and high dimension of system states,we design a joint service caching and load balancing algorithm based on the double-dueling Deep Q network(DQN)approach.The simulation results validate the feasibility and superior performance of our proposed algorithm over several baseline schemes.
基金funded by University Grant Commission with UGC-Ref.No.:3364/(NET-JUNE 2015).
文摘This paper presents a novel fuzzy firefly-based intelligent algorithm for load balancing in mobile cloud computing while reducing makespan.The proposed technique implicitly acts intelligently by using inherent traits of fuzzy and firefly.It automatically adjusts its behavior or converges depending on the information gathered during the search process and objective function.It works for 3-tier architecture,including cloudlet and public cloud.As cloudlets have limited resources,fuzzy logic is used for cloudlet selection using capacity and waiting time as input.Fuzzy provides human-like decisions without using any mathematical model.Firefly is a powerful meta-heuristic optimization technique to balance diversification and solution speed.It balances the load on cloud and cloudlet while minimizing makespan and execution time.However,it may trap in local optimum;levy flight can handle it.Hybridization of fuzzy fireflywith levy flight is a novel technique that provides reduced makespan,execution time,and Degree of imbalance while balancing the load.Simulation has been carried out on the Cloud Analyst platform with National Aeronautics and Space Administration(NASA)and Clarknet datasets.Results show that the proposed algorithm outperforms Ant Colony Optimization Queue Decision Maker(ACOQDM),Distributed Scheduling Optimization Algorithm(DSOA),andUtility-based Firefly Algorithm(UFA)when compared in terms of makespan,Degree of imbalance,and Figure of Merit.
基金supported in part by the Guangdong Basic and Applied Basic Research Foundation under grant 2020A1515110269.
文摘A major challenge for the future wireless network is to design the self-organizing architecture.The reactive self-organizing model of traditional networks needs to be transformed into an active self-organizing network.Due to the user mobility and the coverage of small cells,the network load often becomes unbalanced,resulting in poor network performance.Mobility management has become an important issue to ensure seamless communication when users move between cells,and proactive mobility management is one of the important functions of the active Self-Organizing Network(SON).This paper proposes a proactive mobility management framework for active SON,which transforms the original reactive load balancing into a forward-aware and proactive load balancing.The proposed framework firstly uses the BART model to predict the users’temporal and spatial mobility based on a weekly cycle and then formulate the MLB optimization problem based on the soft load.Two solutions are proposed to solve the above MLB problem.The simulation results show that the proposed method can better optimize the network performance and realize intelligent mobile management for the future network.
文摘According to the advances in users’service requirements,physical hardware accessibility,and speed of resource delivery,Cloud Computing(CC)is an essential technology to be used in many fields.Moreover,the Internet of Things(IoT)is employed for more communication flexibility and richness that are required to obtain fruitful services.A multi-agent system might be a proper solution to control the load balancing of interaction and communication among agents.This paper proposes a multi-agent load balancing framework that consists of two phases to optimize the workload among different servers with large-scale CC power with various utilities and a significant number of IoT devices with low resources.Different agents are integrated based on relevant features of behavioral interaction using classification techniques to balance the workload.Aload balancing algorithm is developed to serve users’requests to improve the solution of workload problems with an efficient distribution.The activity task from IoT devices has been classified by feature selection methods in the preparatory phase to optimize the scalability ofCC.Then,the server’s availability is checked and the classified task is assigned to its suitable server in the main phase to enhance the cloud environment performance.Multi-agent load balancing framework is succeeded to cope with the importance of using large-scale requirements of CC and(low resources and large number)of IoT.
基金supported by the Bio and Medical Technology Development Program of the National Research Foundation(NRF)funded by the Korean government(MSIT)(No.NRF-2019M3E5D1A02069073)supported by the Soonchunhyang University Research Fund.
文摘Healthcare is a fundamental part of every individual’s life.The healthcare industry is developing very rapidly with the help of advanced technologies.Many researchers are trying to build cloud-based healthcare applications that can be accessed by healthcare professionals from their premises,as well as by patients from their mobile devices through communication interfaces.These systems promote reliable and remote interactions between patients and healthcare professionals.However,there are several limitations to these innovative cloud computing-based systems,namely network availability,latency,battery life and resource availability.We propose a hybrid mobile cloud computing(HMCC)architecture to address these challenges.Furthermore,we also evaluate the performance of heuristic and dynamic machine learning based task scheduling and load balancing algorithms on our proposed architecture.We compare them,to identify the strengths and weaknesses of each algorithm;and provide their comparative results,to show latency and energy consumption performance.Challenging issues for cloudbased healthcare systems are discussed in detail.
文摘Recently,the fundamental problem with Hybrid Mobile Ad-hoc Net-works(H-MANETs)is tofind a suitable and secure way of balancing the load through Internet gateways.Moreover,the selection of the gateway and overload of the network results in packet loss and Delay(DL).For optimal performance,it is important to load balance between different gateways.As a result,a stable load balancing procedure is implemented,which selects gateways based on Fuzzy Logic(FL)and increases the efficiency of the network.In this case,since gate-ways are selected based on the number of nodes,the Energy Consumption(EC)was high.This paper presents a novel Node Quality-based Clustering Algo-rithm(NQCA)based on Fuzzy-Genetic for Cluster Head and Gateway Selection(FGCHGS).This algorithm combines NQCA with the Improved Weighted Clus-tering Algorithm(IWCA).The NQCA algorithm divides the network into clusters based upon node priority,transmission range,and neighbourfidelity.In addition,the simulation results tend to evaluate the performance effectiveness of the FFFCHGS algorithm in terms of EC,packet loss rate(PLR),etc.
文摘Every day,more and more data is being produced by the Internet of Things(IoT)applications.IoT data differ in amount,diversity,veracity,and velocity.Because of latency,various types of data handling in cloud computing are not suitable for many time-sensitive applications.When users move from one site to another,mobility also adds to the latency.By placing computing close to IoT devices with mobility support,fog computing addresses these problems.An efficient Load Balancing Algorithm(LBA)improves user experience and Quality of Service(QoS).Classification of Request(CoR)based Resource Adaptive LBA is suggested in this research.This technique clusters fog nodes using an efficient K-means clustering algorithm and then uses a Decision Tree approach to categorize the request.The decision-making process for time-sensitive and delay-tolerable requests is facilitated by the classification of requests.LBA does the operation based on these classifications.The MobFogSim simulation program is utilized to assess how well the algorithm with mobility features performs.The outcome demonstrates that the LBA algorithm’s performance enhances the total system performance,which was attained by(90.8%).Using LBA,several metrics may be examined,including Response Time(RT),delay(d),Energy Consumption(EC),and latency.Through the on-demand provisioning of necessary resources to IoT users,our suggested LBA assures effective resource usage.
文摘Advancements in cloud computing and virtualization technologies have revolutionized Enterprise Application Development with innovative ways to design and develop complex systems.Microservices Architecture is one of the recent techniques in which Enterprise Systems can be developed as fine-grained smaller components and deployed independently.This methodology brings numerous benefits like scalability,resilience,flexibility in development,faster time to market,etc.and the advantages;Microservices bring some challenges too.Multiple microservices need to be invoked one by one as a chain.In most applications,more than one chain of microservices runs in parallel to complete a particular requirement To complete a user’s request.It results in competition for resources and the need for more inter-service communication among the services,which increases the overall latency of the application.A new approach has been proposed in this paper to handle a complex chain of microservices and reduce the latency of user requests.A machine learning technique is followed to predict the weighting time of different types of requests.The communication time among services distributed among different physical machines are estimated based on that and obtained insights are applied to an algorithm to calculate their priorities dynamically and select suitable service instances to minimize the latency based on the shortest queue waiting time.Experiments were done for both interactive as well as non interactive workloads to test the effectiveness of the solution.The approach has been proved to be very effective in reducing latency in the case of long service chains.
基金supported by the project:“Research and Implementation of Innovative Solutions for Monitoring Consumption in Technical Installations Using Artificial Intelligence”,beneficiary S.C.REMONI TECHNOLOGIES RO S.R.L in partnership with“Gheorghe Asachi”Technical University of Iasi,Financing Contract No.400/390076/26.11.2021,SMIS Code 121866,financed by POC/163/1/3.
文摘Cloud Technology is a new platform that offers on-demand computing Peripheral such as storage,processing power,and other computer system resources.It is also referred to as a system that will let the consumers utilize computational resources like databases,servers,storage,and intelligence over the Internet.In a cloud network,load balancing is the process of dividing network traffic among a cluster of available servers to increase efficiency.It is also known as a server pool or server farm.When a single node is overwhelmed,balancing the workload is needed to manage unpredictable workflows.The load balancer sends the load to another free node in this case.We focus on the Balancing of workflows with the proposed approach,and we present a novel method to balance the load that manages the dynamic scheduling process.One of the preexisting load balancing techniques is considered,however it is somewhat modified to fit the scenario at hand.Depending on the experimentation’s findings,it is concluded that this suggested approach improves load balancing consistency,response time,and throughput by 6%.
文摘In wireless sensor network(WSN),the gateways which are placed far away from the base station(BS)forward the collected data to the BS through the gateways which are nearer to the BS.This leads to more energy consumption because the gateways nearer to the BS manages heavy traffic load.So,to over-come this issue,loads around the gateways are to be balanced by presenting energy efficient clustering approach.Besides,to enhance the lifetime of the net-work,optimal routing path is to be established between the source node and BS.For energy efficient load balancing and routing,multi objective based beetle swarm optimization(BSO)algorithm is presented in this paper.Using this algo-rithm,optimal clustering and routing are performed depend on the objective func-tions routingfitness and clusteringfitness.This approach leads to decrease the power consumption.Simulation results show that the performance of the pro-posed BSO based clustering and routing scheme attains better results than that of the existing algorithms in terms of energy consumption,delivery ratio,through-put and network lifetime.Namely,the proposed scheme increases throughput to 72%and network lifetime to 37%as well as it reduces delay to 37%than the existing optimization algorithms based clustering and routing schemes.
基金This work was supported by the National Key R&D Program of China(2021YFB2900604).
文摘A low-Earth-orbit(LEO)satellite network can provide full-coverage access services worldwide and is an essential candidate for future 6G networking.However,the large variability of the geographic distribution of the Earth’s population leads to an uneven service volume distribution of access service.Moreover,the limitations on the resources of satellites are far from being able to serve the traffic in hotspot areas.To enhance the forwarding capability of satellite networks,we first assess how hotspot areas under different load cases and spatial scales significantly affect the network throughput of an LEO satellite network overall.Then,we propose a multi-region cooperative traffic scheduling algorithm.The algorithm migrates low-grade traffic from hotspot areas to coldspot areas for forwarding,significantly increasing the overall throughput of the satellite network while sacrificing some latency of end-to-end forwarding.This algorithm can utilize all the global satellite resources and improve the utilization of network resources.We model the cooperative multi-region scheduling of large-scale LEO satellites.Based on the model,we build a system testbed using OMNET++to compare the proposed method with existing techniques.The simulations show that our proposed method can reduce the packet loss probability by 30%and improve the resource utilization ratio by 3.69%.
基金jointly supported by the Jiangsu Postgraduate Research and Practice Innovation Project under Grant KYCX22_1030,SJCX22_0283 and SJCX23_0293the NUPTSF under Grant NY220201.
文摘Task scheduling plays a key role in effectively managing and allocating computing resources to meet various computing tasks in a cloud computing environment.Short execution time and low load imbalance may be the challenges for some algorithms in resource scheduling scenarios.In this work,the Hierarchical Particle Swarm Optimization-Evolutionary Artificial Bee Colony Algorithm(HPSO-EABC)has been proposed,which hybrids our presented Evolutionary Artificial Bee Colony(EABC),and Hierarchical Particle Swarm Optimization(HPSO)algorithm.The HPSO-EABC algorithm incorporates both the advantages of the HPSO and the EABC algorithm.Comprehensive testing including evaluations of algorithm convergence speed,resource execution time,load balancing,and operational costs has been done.The results indicate that the EABC algorithm exhibits greater parallelism compared to the Artificial Bee Colony algorithm.Compared with the Particle Swarm Optimization algorithm,the HPSO algorithmnot only improves the global search capability but also effectively mitigates getting stuck in local optima.As a result,the hybrid HPSO-EABC algorithm demonstrates significant improvements in terms of stability and convergence speed.Moreover,it exhibits enhanced resource scheduling performance in both homogeneous and heterogeneous environments,effectively reducing execution time and cost,which also is verified by the ablation experimental.
基金The National Natural Science Foundation of China(No.69973007).
文摘To solve the load balancing problem in a triplet-based hierarchical interconnection network(THIN) system, a dynamic load balancing (DLB)algorithm--THINDLBA, which adopts multicast tree (MT)technology to improve the efficiency of interchanging load information, is presented. To support the algorithm, a complete set of DLB messages and a schema of maintaining DLB information in each processing node are designed. The load migration request messages from the heavily loaded node (HLN)are spread along an MT whose root is the HLN. And the lightly loaded nodes(LLNs) covered by the MT are the candidate destinations of load migration; the load information interchanged between the LLNs and the HLN can be transmitted along the MT. So the HLN can migrate excess loads out as many as possible during a one time execution of the THINDLBA, and its load state can be improved as quickly as possible. To avoid wrongly transmitted or redundant DLB messages due to MT overlapping, the MT construction is restricted in the design of the THINDLBA. Through experiments, the effectiveness of four DLB algorithms are compared, and the results show that the THINDLBA can effectively decrease the time costs of THIN systems in dealing with large scale computeintensive tasks more than others.
文摘This paper focuses on solving a problem of improving system robustness and the efficiency of a distributed system at the same time. Fault tolerance with active replication and load balancing techniques are used. The pros and cons of both techniques are analyzed, and a novel load balancing framework for fault tolerant systems with active replication is presented. Hierarchical architecture is described in detail. The framework can dynamically adjust fault tolerant groups and their memberships with respect to system loads. Three potential task scheduler group selection methods are proposed and simulation tests are made. Further analysis of test data is done and helpful observations for system design are also pointed out, including effects of task arrival intensity and task set size, relationship between total task execution time and single task execution time.