This paper focuses on the scheduling problem of workflow tasks that exhibit interdependencies.Unlike indepen-dent batch tasks,workflows typically consist of multiple subtasks with intrinsic correlations and dependenci...This paper focuses on the scheduling problem of workflow tasks that exhibit interdependencies.Unlike indepen-dent batch tasks,workflows typically consist of multiple subtasks with intrinsic correlations and dependencies.It necessitates the distribution of various computational tasks to appropriate computing node resources in accor-dance with task dependencies to ensure the smooth completion of the entire workflow.Workflow scheduling must consider an array of factors,including task dependencies,availability of computational resources,and the schedulability of tasks.Therefore,this paper delves into the distributed graph database workflow task scheduling problem and proposes a workflow scheduling methodology based on deep reinforcement learning(DRL).The method optimizes the maximum completion time(makespan)and response time of workflow tasks,aiming to enhance the responsiveness of workflow tasks while ensuring the minimization of the makespan.The experimental results indicate that the Q-learning Deep Reinforcement Learning(Q-DRL)algorithm markedly diminishes the makespan and refines the average response time within distributed graph database environments.In quantifying makespan,Q-DRL achieves mean reductions of 12.4%and 11.9%over established First-fit and Random scheduling strategies,respectively.Additionally,Q-DRL surpasses the performance of both DRL-Cloud and Improved Deep Q-learning Network(IDQN)algorithms,with improvements standing at 4.4%and 2.6%,respectively.With reference to average response time,the Q-DRL approach exhibits a significantly enhanced performance in the scheduling of workflow tasks,decreasing the average by 2.27%and 4.71%when compared to IDQN and DRL-Cloud,respectively.The Q-DRL algorithm also demonstrates a notable increase in the efficiency of system resource utilization,reducing the average idle rate by 5.02%and 9.30%in comparison to IDQN and DRL-Cloud,respectively.These findings support the assertion that Q-DRL not only upholds a lower average idle rate but also effectively curtails the average response time,thereby substantially improving processing efficiency and optimizing resource utilization within distributed graph database systems.展开更多
Unbalanced traffic distribution in cellular networks results in congestion and degrades spectrum efficiency.To tackle this problem,we propose an Unmanned Aerial Vehicle(UAV)-assisted wireless network in which the UAV ...Unbalanced traffic distribution in cellular networks results in congestion and degrades spectrum efficiency.To tackle this problem,we propose an Unmanned Aerial Vehicle(UAV)-assisted wireless network in which the UAV acts as an aerial relay to divert some traffic from the overloaded cell to its adjacent underloaded cell.To fully exploit its potential,we jointly optimize the UAV position,user association,spectrum allocation,and power allocation to maximize the sum-log-rate of all users in two adjacent cells.To tackle the complicated joint optimization problem,we first design a genetic-based algorithm to optimize the UAV position.Then,we simplify the problem by theoretical analysis and devise a low-complexity algorithm according to the branch-and-bound method,so as to obtain the optimal user association and spectrum allocation schemes.We further propose an iterative power allocation algorithm based on the sequential convex approximation theory.The simulation results indicate that the proposed UAV-assisted wireless network is superior to the terrestrial network in both utility and throughput,and the proposed algorithms can substantially improve the network performance in comparison with the other schemes.展开更多
Cloud Computing has the ability to provide on-demand access to a shared resource pool.It has completely changed the way businesses are managed,implement applications,and provide services.The rise in popularity has led...Cloud Computing has the ability to provide on-demand access to a shared resource pool.It has completely changed the way businesses are managed,implement applications,and provide services.The rise in popularity has led to a significant increase in the user demand for services.However,in cloud environments efficient load balancing is essential to ensure optimal performance and resource utilization.This systematic review targets a detailed description of load balancing techniques including static and dynamic load balancing algorithms.Specifically,metaheuristic-based dynamic load balancing algorithms are identified as the optimal solution in case of increased traffic.In a cloud-based context,this paper describes load balancing measurements,including the benefits and drawbacks associated with the selected load balancing techniques.It also summarizes the algorithms based on implementation,time complexity,adaptability,associated issue(s),and targeted QoS parameters.Additionally,the analysis evaluates the tools and instruments utilized in each investigated study.Moreover,comparative analysis among static,traditional dynamic and metaheuristic algorithms based on response time by using the CloudSim simulation tool is also performed.Finally,the key open problems and potential directions for the state-of-the-art metaheuristic-based approaches are also addressed.展开更多
With the continuous expansion of the data center network scale, changing network requirements, and increasing pressure on network bandwidth, the traditional network architecture can no longer meet people’s needs. The...With the continuous expansion of the data center network scale, changing network requirements, and increasing pressure on network bandwidth, the traditional network architecture can no longer meet people’s needs. The development of software defined networks has brought new opportunities and challenges to future networks. The data and control separation characteristics of SDN improve the performance of the entire network. Researchers have integrated SDN architecture into data centers to improve network resource utilization and performance. This paper first introduces the basic concepts of SDN and data center networks. Then it discusses SDN-based load balancing mechanisms for data centers from different perspectives. Finally, it summarizes and looks forward to the study on SDN-based load balancing mechanisms and its development trend.展开更多
In this paper, a sender-initiated protocol is applied which uses fuzzy logic control method to improve computer networks performance by balancing loads among computers. This new model devises sender-initiated protocol...In this paper, a sender-initiated protocol is applied which uses fuzzy logic control method to improve computer networks performance by balancing loads among computers. This new model devises sender-initiated protocol for load transfer for load balancing. Groups are formed and every group has a node called a designated representative (DR). During load transferring processes, loads are transferred using the DR in each group to achieve load balancing purposes. The simulation results show that the performance of the protocol proposed is better than the compared conventional method. This protocol is more stable than the method without using the fuzzy logic control.展开更多
Sensors are considered as important elements of electronic devices.In many applications and service,Wireless Sensor Networks(WSNs)are involved in significant data sharing that are delivered to the sink node in energy ...Sensors are considered as important elements of electronic devices.In many applications and service,Wireless Sensor Networks(WSNs)are involved in significant data sharing that are delivered to the sink node in energy efficient man-ner using multi-hop communications.But,the major challenge in WSN is the nodes are having limited battery resources,it is important to monitor the consumption rate of energy is very much needed.However,reducing energy con-sumption can increase the network lifetime in effective manner.For that,clustering methods are widely used for optimizing the rate of energy consumption among the sensor nodes.In that concern,this paper involves in deriving a novel model called Improved Load-Balanced Clustering for Energy-Aware Routing(ILBC-EAR),which mainly concentrates on optimal energy utilization with load-balanced process among cluster heads and member nodes.For providing equal rate of energy consumption among nodes,the dimensions of framed clusters are measured.Moreover,the model develops a Finest Routing Scheme based on Load-Balanced Clustering to transmit the sensed information to the sink or base station.The evaluation results depict that the derived energy aware model attains higher rate of life time than other works and also achieves balanced energy rate among head node.Additionally,the model also provides higher throughput and minimal delay in delivering data packets.展开更多
Real-time applications based on Wireless Sensor Network(WSN)tech-nologies are quickly increasing due to intelligent surroundings.Among the most significant resources in the WSN are battery power and security.Clustering...Real-time applications based on Wireless Sensor Network(WSN)tech-nologies are quickly increasing due to intelligent surroundings.Among the most significant resources in the WSN are battery power and security.Clustering stra-tegies improve the power factor and secure the WSN environment.It takes more electricity to forward data in a WSN.Though numerous clustering methods have been developed to provide energy consumption,there is indeed a risk of unequal load balancing,resulting in a decrease in the network’s lifetime due to network inequalities and less security.These possibilities arise due to the cluster head’s limited life span.These cluster heads(CH)are in charge of all activities and con-trol intra-cluster and inter-cluster interactions.The proposed method uses Lifetime centric load balancing mechanisms(LCLBM)and Cluster-based energy optimiza-tion using a mobile sink algorithm(CEOMS).LCLBM emphasizes the selection of CH,system architectures,and optimal distribution of CH.In addition,the LCLBM was added with an assistant cluster head(ACH)for load balancing.Power consumption,communications latency,the frequency of failing nodes,high security,and one-way delay are essential variables to consider while evaluating LCLBM.CEOMS will choose a cluster leader based on the influence of the fol-lowing parameters on the energy balance of WSNs.According to simulatedfind-ings,the suggested LCLBM-CEOMS method increases cluster head selection self-adaptability,improves the network’s lifetime,decreases data latency,and bal-ances network capacity.展开更多
Internet of Things(IoT)empowers imaginative applications and permits new services when mobile nodes are included.For IoT-enabled low-power and lossy networks(LLN),the Routing Protocol for Low-power and Lossy Networks(...Internet of Things(IoT)empowers imaginative applications and permits new services when mobile nodes are included.For IoT-enabled low-power and lossy networks(LLN),the Routing Protocol for Low-power and Lossy Networks(RPL)has become an established standard routing protocol.Mobility under standard RPL remains a difficult issue as it causes continuous path disturbance,energy loss,and increases the end-to-end delay in the network.In this unique circumstance,a Balanced-load and Energy-efficient RPL(BE-RPL)is proposed.It is a routing technique that is both energy-efficient and mobility-aware.It responds quicker to link breakage through received signal strength-based mobility monitoring and selecting a new preferred parent reactively.The proposed system also implements load balancing among stationary nodes for leaf node allocation.Static nodes with more leaf nodes are restricted from participating in the election for a new preferred parent.The performance of BE-RPL is assessed using the COOJA simulator.It improves the energy use,network control overhead,frame acknowledgment ratio,and packet delivery ratio of the network.展开更多
One of the very important functions of three-phase inverter is to maintain the symmetric three-phase output voltage when the three-phase loads are unbalanced. Although the traditional symmetrical component decomposing...One of the very important functions of three-phase inverter is to maintain the symmetric three-phase output voltage when the three-phase loads are unbalanced. Although the traditional symmetrical component decomposing and superimpose theory can keep the voltage balance through compensating the positive-, negative- and zero-sequence components of the output voltage of inverter, however, this method is time-consuming and not suitable for control. Aiming at high power medium frequency inverter source, a P+Resonant (Proportion and Resonant) controller which ensured a balanced three phase output voltage under unbalanced load is proposed in this paper. The regulator was proved to be applicable to both three-phase three-wire system and three-phase four-wire system and developed two methods of realization. The simulation results verified that this method can suppressed effectively the output voltage distorted caused by the unbalanced load and attained a high quality voltage waveforms.展开更多
At present,the flow table of the SDN switch is stored in the costly Ternary Content Addressable Memory(TCAM)cache.Due to the cost problem,the number of flow tables that the SDN switch can store is extremely limited,wh...At present,the flow table of the SDN switch is stored in the costly Ternary Content Addressable Memory(TCAM)cache.Due to the cost problem,the number of flow tables that the SDN switch can store is extremely limited,which is far less than the number of traffic,so it is prone to overflow problem,and leads to network paralysis.That has become a bottleneck in restricting the processing capacity of the data center,and will become a weak point focused by attackers.In this paper,we propose an algorithm for the Alarm Switch Remove(ASR)that fully loads the flow table space in SDN,and further put forward an integrated load balancing scheme in SDN.Finally,we use Mininet to verify that the scheme can ease the SDN switch flow table overflow problem and increase network throughput.展开更多
A major challenge for the future wireless network is to design the self-organizing architecture.The reactive self-organizing model of traditional networks needs to be transformed into an active self-organizing network...A major challenge for the future wireless network is to design the self-organizing architecture.The reactive self-organizing model of traditional networks needs to be transformed into an active self-organizing network.Due to the user mobility and the coverage of small cells,the network load often becomes unbalanced,resulting in poor network performance.Mobility management has become an important issue to ensure seamless communication when users move between cells,and proactive mobility management is one of the important functions of the active Self-Organizing Network(SON).This paper proposes a proactive mobility management framework for active SON,which transforms the original reactive load balancing into a forward-aware and proactive load balancing.The proposed framework firstly uses the BART model to predict the users’temporal and spatial mobility based on a weekly cycle and then formulate the MLB optimization problem based on the soft load.Two solutions are proposed to solve the above MLB problem.The simulation results show that the proposed method can better optimize the network performance and realize intelligent mobile management for the future network.展开更多
According to the advances in users’service requirements,physical hardware accessibility,and speed of resource delivery,Cloud Computing(CC)is an essential technology to be used in many fields.Moreover,the Internet of ...According to the advances in users’service requirements,physical hardware accessibility,and speed of resource delivery,Cloud Computing(CC)is an essential technology to be used in many fields.Moreover,the Internet of Things(IoT)is employed for more communication flexibility and richness that are required to obtain fruitful services.A multi-agent system might be a proper solution to control the load balancing of interaction and communication among agents.This paper proposes a multi-agent load balancing framework that consists of two phases to optimize the workload among different servers with large-scale CC power with various utilities and a significant number of IoT devices with low resources.Different agents are integrated based on relevant features of behavioral interaction using classification techniques to balance the workload.Aload balancing algorithm is developed to serve users’requests to improve the solution of workload problems with an efficient distribution.The activity task from IoT devices has been classified by feature selection methods in the preparatory phase to optimize the scalability ofCC.Then,the server’s availability is checked and the classified task is assigned to its suitable server in the main phase to enhance the cloud environment performance.Multi-agent load balancing framework is succeeded to cope with the importance of using large-scale requirements of CC and(low resources and large number)of IoT.展开更多
Every day,more and more data is being produced by the Internet of Things(IoT)applications.IoT data differ in amount,diversity,veracity,and velocity.Because of latency,various types of data handling in cloud computing ...Every day,more and more data is being produced by the Internet of Things(IoT)applications.IoT data differ in amount,diversity,veracity,and velocity.Because of latency,various types of data handling in cloud computing are not suitable for many time-sensitive applications.When users move from one site to another,mobility also adds to the latency.By placing computing close to IoT devices with mobility support,fog computing addresses these problems.An efficient Load Balancing Algorithm(LBA)improves user experience and Quality of Service(QoS).Classification of Request(CoR)based Resource Adaptive LBA is suggested in this research.This technique clusters fog nodes using an efficient K-means clustering algorithm and then uses a Decision Tree approach to categorize the request.The decision-making process for time-sensitive and delay-tolerable requests is facilitated by the classification of requests.LBA does the operation based on these classifications.The MobFogSim simulation program is utilized to assess how well the algorithm with mobility features performs.The outcome demonstrates that the LBA algorithm’s performance enhances the total system performance,which was attained by(90.8%).Using LBA,several metrics may be examined,including Response Time(RT),delay(d),Energy Consumption(EC),and latency.Through the on-demand provisioning of necessary resources to IoT users,our suggested LBA assures effective resource usage.展开更多
As a new networking paradigm,Software-Defined Networking(SDN)enables us to cope with the limitations of traditional networks.SDN uses a controller that has a global view of the network and switch devices which act as ...As a new networking paradigm,Software-Defined Networking(SDN)enables us to cope with the limitations of traditional networks.SDN uses a controller that has a global view of the network and switch devices which act as packet forwarding hardware,known as“OpenFlow switches”.Since load balancing service is essential to distribute workload across servers in data centers,we propose an effective load balancing scheme in SDN,using a genetic programming approach,called Genetic Programming based Load Balancing(GPLB).We formulate the problem to find a path:1)with the best bottleneck switch which has the lowest capacity within bottleneck switches of each path,2)with the shortest path,and 3)requiring the less possible operations.For the purpose of choosing the real-time least loaded path,GPLB immediately calculates the integrated load of paths based on the information that receives from the SDN controller.Hence,in this design,the controller sends the load information of each path to the load balancing algorithm periodically and then the load balancing algorithm returns a least loaded path to the controller.In this paper,we use the Mininet emulator and the OpenDaylight controller to evaluate the effectiveness of the GPLB.The simulative study of the GPLB shows that there is a big improvement in performance metrics and the latency and the jitter are minimized.The GPLB also has the maximum throughput in comparison with related works and has performed better in the heavy traffic situation.The results show that our model stands smartly while not increasing further overhead.展开更多
Cloud Technology is a new platform that offers on-demand computing Peripheral such as storage,processing power,and other computer system resources.It is also referred to as a system that will let the consumers utilize...Cloud Technology is a new platform that offers on-demand computing Peripheral such as storage,processing power,and other computer system resources.It is also referred to as a system that will let the consumers utilize computational resources like databases,servers,storage,and intelligence over the Internet.In a cloud network,load balancing is the process of dividing network traffic among a cluster of available servers to increase efficiency.It is also known as a server pool or server farm.When a single node is overwhelmed,balancing the workload is needed to manage unpredictable workflows.The load balancer sends the load to another free node in this case.We focus on the Balancing of workflows with the proposed approach,and we present a novel method to balance the load that manages the dynamic scheduling process.One of the preexisting load balancing techniques is considered,however it is somewhat modified to fit the scenario at hand.Depending on the experimentation’s findings,it is concluded that this suggested approach improves load balancing consistency,response time,and throughput by 6%.展开更多
Advancements in cloud computing and virtualization technologies have revolutionized Enterprise Application Development with innovative ways to design and develop complex systems.Microservices Architecture is one of th...Advancements in cloud computing and virtualization technologies have revolutionized Enterprise Application Development with innovative ways to design and develop complex systems.Microservices Architecture is one of the recent techniques in which Enterprise Systems can be developed as fine-grained smaller components and deployed independently.This methodology brings numerous benefits like scalability,resilience,flexibility in development,faster time to market,etc.and the advantages;Microservices bring some challenges too.Multiple microservices need to be invoked one by one as a chain.In most applications,more than one chain of microservices runs in parallel to complete a particular requirement To complete a user’s request.It results in competition for resources and the need for more inter-service communication among the services,which increases the overall latency of the application.A new approach has been proposed in this paper to handle a complex chain of microservices and reduce the latency of user requests.A machine learning technique is followed to predict the weighting time of different types of requests.The communication time among services distributed among different physical machines are estimated based on that and obtained insights are applied to an algorithm to calculate their priorities dynamically and select suitable service instances to minimize the latency based on the shortest queue waiting time.Experiments were done for both interactive as well as non interactive workloads to test the effectiveness of the solution.The approach has been proved to be very effective in reducing latency in the case of long service chains.展开更多
This paper presents a novel fuzzy firefly-based intelligent algorithm for load balancing in mobile cloud computing while reducing makespan.The proposed technique implicitly acts intelligently by using inherent traits ...This paper presents a novel fuzzy firefly-based intelligent algorithm for load balancing in mobile cloud computing while reducing makespan.The proposed technique implicitly acts intelligently by using inherent traits of fuzzy and firefly.It automatically adjusts its behavior or converges depending on the information gathered during the search process and objective function.It works for 3-tier architecture,including cloudlet and public cloud.As cloudlets have limited resources,fuzzy logic is used for cloudlet selection using capacity and waiting time as input.Fuzzy provides human-like decisions without using any mathematical model.Firefly is a powerful meta-heuristic optimization technique to balance diversification and solution speed.It balances the load on cloud and cloudlet while minimizing makespan and execution time.However,it may trap in local optimum;levy flight can handle it.Hybridization of fuzzy fireflywith levy flight is a novel technique that provides reduced makespan,execution time,and Degree of imbalance while balancing the load.Simulation has been carried out on the Cloud Analyst platform with National Aeronautics and Space Administration(NASA)and Clarknet datasets.Results show that the proposed algorithm outperforms Ant Colony Optimization Queue Decision Maker(ACOQDM),Distributed Scheduling Optimization Algorithm(DSOA),andUtility-based Firefly Algorithm(UFA)when compared in terms of makespan,Degree of imbalance,and Figure of Merit.展开更多
Integrating the blockchain technology into mobile-edge computing(MEC)networks with multiple cooperative MEC servers(MECS)providing a promising solution to improving resource utilization,and helping establish a secure ...Integrating the blockchain technology into mobile-edge computing(MEC)networks with multiple cooperative MEC servers(MECS)providing a promising solution to improving resource utilization,and helping establish a secure reward mechanism that can facilitate load balancing among MECS.In addition,intelligent management of service caching and load balancing can improve the network utility in MEC blockchain networks with multiple types of workloads.In this paper,we investigate a learningbased joint service caching and load balancing policy for optimizing the communication and computation resources allocation,so as to improve the resource utilization of MEC blockchain networks.We formulate the problem as a challenging long-term network revenue maximization Markov decision process(MDP)problem.To address the highly dynamic and high dimension of system states,we design a joint service caching and load balancing algorithm based on the double-dueling Deep Q network(DQN)approach.The simulation results validate the feasibility and superior performance of our proposed algorithm over several baseline schemes.展开更多
Healthcare is a fundamental part of every individual’s life.The healthcare industry is developing very rapidly with the help of advanced technologies.Many researchers are trying to build cloud-based healthcare applic...Healthcare is a fundamental part of every individual’s life.The healthcare industry is developing very rapidly with the help of advanced technologies.Many researchers are trying to build cloud-based healthcare applications that can be accessed by healthcare professionals from their premises,as well as by patients from their mobile devices through communication interfaces.These systems promote reliable and remote interactions between patients and healthcare professionals.However,there are several limitations to these innovative cloud computing-based systems,namely network availability,latency,battery life and resource availability.We propose a hybrid mobile cloud computing(HMCC)architecture to address these challenges.Furthermore,we also evaluate the performance of heuristic and dynamic machine learning based task scheduling and load balancing algorithms on our proposed architecture.We compare them,to identify the strengths and weaknesses of each algorithm;and provide their comparative results,to show latency and energy consumption performance.Challenging issues for cloudbased healthcare systems are discussed in detail.展开更多
In wireless sensor network(WSN),the gateways which are placed far away from the base station(BS)forward the collected data to the BS through the gateways which are nearer to the BS.This leads to more energy consumptio...In wireless sensor network(WSN),the gateways which are placed far away from the base station(BS)forward the collected data to the BS through the gateways which are nearer to the BS.This leads to more energy consumption because the gateways nearer to the BS manages heavy traffic load.So,to over-come this issue,loads around the gateways are to be balanced by presenting energy efficient clustering approach.Besides,to enhance the lifetime of the net-work,optimal routing path is to be established between the source node and BS.For energy efficient load balancing and routing,multi objective based beetle swarm optimization(BSO)algorithm is presented in this paper.Using this algo-rithm,optimal clustering and routing are performed depend on the objective func-tions routingfitness and clusteringfitness.This approach leads to decrease the power consumption.Simulation results show that the performance of the pro-posed BSO based clustering and routing scheme attains better results than that of the existing algorithms in terms of energy consumption,delivery ratio,through-put and network lifetime.Namely,the proposed scheme increases throughput to 72%and network lifetime to 37%as well as it reduces delay to 37%than the existing optimization algorithms based clustering and routing schemes.展开更多
基金funded by the Science and Technology Foundation of State Grid Corporation of China(Grant No.5108-202218280A-2-397-XG).
文摘This paper focuses on the scheduling problem of workflow tasks that exhibit interdependencies.Unlike indepen-dent batch tasks,workflows typically consist of multiple subtasks with intrinsic correlations and dependencies.It necessitates the distribution of various computational tasks to appropriate computing node resources in accor-dance with task dependencies to ensure the smooth completion of the entire workflow.Workflow scheduling must consider an array of factors,including task dependencies,availability of computational resources,and the schedulability of tasks.Therefore,this paper delves into the distributed graph database workflow task scheduling problem and proposes a workflow scheduling methodology based on deep reinforcement learning(DRL).The method optimizes the maximum completion time(makespan)and response time of workflow tasks,aiming to enhance the responsiveness of workflow tasks while ensuring the minimization of the makespan.The experimental results indicate that the Q-learning Deep Reinforcement Learning(Q-DRL)algorithm markedly diminishes the makespan and refines the average response time within distributed graph database environments.In quantifying makespan,Q-DRL achieves mean reductions of 12.4%and 11.9%over established First-fit and Random scheduling strategies,respectively.Additionally,Q-DRL surpasses the performance of both DRL-Cloud and Improved Deep Q-learning Network(IDQN)algorithms,with improvements standing at 4.4%and 2.6%,respectively.With reference to average response time,the Q-DRL approach exhibits a significantly enhanced performance in the scheduling of workflow tasks,decreasing the average by 2.27%and 4.71%when compared to IDQN and DRL-Cloud,respectively.The Q-DRL algorithm also demonstrates a notable increase in the efficiency of system resource utilization,reducing the average idle rate by 5.02%and 9.30%in comparison to IDQN and DRL-Cloud,respectively.These findings support the assertion that Q-DRL not only upholds a lower average idle rate but also effectively curtails the average response time,thereby substantially improving processing efficiency and optimizing resource utilization within distributed graph database systems.
基金supported in part by the National Key Research and Development Program of China under Grant 2020YFB1807003in part by the National Natural Science Foundation of China under Grants 61901381,62171385,and 61901378+3 种基金in part by the Aeronautical Science Foundation of China under Grant 2020z073053004in part by the Foundation of the State Key Laboratory of Integrated Services Networks of Xidian University under Grant ISN21-06in part by the Key Research Program and Industrial Innovation Chain Project of Shaanxi Province under Grant 2019ZDLGY07-10in part by the Natural Science Fundamental Research Program of Shaanxi Province under Grant 2021JM-069.
文摘Unbalanced traffic distribution in cellular networks results in congestion and degrades spectrum efficiency.To tackle this problem,we propose an Unmanned Aerial Vehicle(UAV)-assisted wireless network in which the UAV acts as an aerial relay to divert some traffic from the overloaded cell to its adjacent underloaded cell.To fully exploit its potential,we jointly optimize the UAV position,user association,spectrum allocation,and power allocation to maximize the sum-log-rate of all users in two adjacent cells.To tackle the complicated joint optimization problem,we first design a genetic-based algorithm to optimize the UAV position.Then,we simplify the problem by theoretical analysis and devise a low-complexity algorithm according to the branch-and-bound method,so as to obtain the optimal user association and spectrum allocation schemes.We further propose an iterative power allocation algorithm based on the sequential convex approximation theory.The simulation results indicate that the proposed UAV-assisted wireless network is superior to the terrestrial network in both utility and throughput,and the proposed algorithms can substantially improve the network performance in comparison with the other schemes.
文摘Cloud Computing has the ability to provide on-demand access to a shared resource pool.It has completely changed the way businesses are managed,implement applications,and provide services.The rise in popularity has led to a significant increase in the user demand for services.However,in cloud environments efficient load balancing is essential to ensure optimal performance and resource utilization.This systematic review targets a detailed description of load balancing techniques including static and dynamic load balancing algorithms.Specifically,metaheuristic-based dynamic load balancing algorithms are identified as the optimal solution in case of increased traffic.In a cloud-based context,this paper describes load balancing measurements,including the benefits and drawbacks associated with the selected load balancing techniques.It also summarizes the algorithms based on implementation,time complexity,adaptability,associated issue(s),and targeted QoS parameters.Additionally,the analysis evaluates the tools and instruments utilized in each investigated study.Moreover,comparative analysis among static,traditional dynamic and metaheuristic algorithms based on response time by using the CloudSim simulation tool is also performed.Finally,the key open problems and potential directions for the state-of-the-art metaheuristic-based approaches are also addressed.
文摘With the continuous expansion of the data center network scale, changing network requirements, and increasing pressure on network bandwidth, the traditional network architecture can no longer meet people’s needs. The development of software defined networks has brought new opportunities and challenges to future networks. The data and control separation characteristics of SDN improve the performance of the entire network. Researchers have integrated SDN architecture into data centers to improve network resource utilization and performance. This paper first introduces the basic concepts of SDN and data center networks. Then it discusses SDN-based load balancing mechanisms for data centers from different perspectives. Finally, it summarizes and looks forward to the study on SDN-based load balancing mechanisms and its development trend.
文摘In this paper, a sender-initiated protocol is applied which uses fuzzy logic control method to improve computer networks performance by balancing loads among computers. This new model devises sender-initiated protocol for load transfer for load balancing. Groups are formed and every group has a node called a designated representative (DR). During load transferring processes, loads are transferred using the DR in each group to achieve load balancing purposes. The simulation results show that the performance of the protocol proposed is better than the compared conventional method. This protocol is more stable than the method without using the fuzzy logic control.
文摘Sensors are considered as important elements of electronic devices.In many applications and service,Wireless Sensor Networks(WSNs)are involved in significant data sharing that are delivered to the sink node in energy efficient man-ner using multi-hop communications.But,the major challenge in WSN is the nodes are having limited battery resources,it is important to monitor the consumption rate of energy is very much needed.However,reducing energy con-sumption can increase the network lifetime in effective manner.For that,clustering methods are widely used for optimizing the rate of energy consumption among the sensor nodes.In that concern,this paper involves in deriving a novel model called Improved Load-Balanced Clustering for Energy-Aware Routing(ILBC-EAR),which mainly concentrates on optimal energy utilization with load-balanced process among cluster heads and member nodes.For providing equal rate of energy consumption among nodes,the dimensions of framed clusters are measured.Moreover,the model develops a Finest Routing Scheme based on Load-Balanced Clustering to transmit the sensed information to the sink or base station.The evaluation results depict that the derived energy aware model attains higher rate of life time than other works and also achieves balanced energy rate among head node.Additionally,the model also provides higher throughput and minimal delay in delivering data packets.
文摘Real-time applications based on Wireless Sensor Network(WSN)tech-nologies are quickly increasing due to intelligent surroundings.Among the most significant resources in the WSN are battery power and security.Clustering stra-tegies improve the power factor and secure the WSN environment.It takes more electricity to forward data in a WSN.Though numerous clustering methods have been developed to provide energy consumption,there is indeed a risk of unequal load balancing,resulting in a decrease in the network’s lifetime due to network inequalities and less security.These possibilities arise due to the cluster head’s limited life span.These cluster heads(CH)are in charge of all activities and con-trol intra-cluster and inter-cluster interactions.The proposed method uses Lifetime centric load balancing mechanisms(LCLBM)and Cluster-based energy optimiza-tion using a mobile sink algorithm(CEOMS).LCLBM emphasizes the selection of CH,system architectures,and optimal distribution of CH.In addition,the LCLBM was added with an assistant cluster head(ACH)for load balancing.Power consumption,communications latency,the frequency of failing nodes,high security,and one-way delay are essential variables to consider while evaluating LCLBM.CEOMS will choose a cluster leader based on the influence of the fol-lowing parameters on the energy balance of WSNs.According to simulatedfind-ings,the suggested LCLBM-CEOMS method increases cluster head selection self-adaptability,improves the network’s lifetime,decreases data latency,and bal-ances network capacity.
文摘Internet of Things(IoT)empowers imaginative applications and permits new services when mobile nodes are included.For IoT-enabled low-power and lossy networks(LLN),the Routing Protocol for Low-power and Lossy Networks(RPL)has become an established standard routing protocol.Mobility under standard RPL remains a difficult issue as it causes continuous path disturbance,energy loss,and increases the end-to-end delay in the network.In this unique circumstance,a Balanced-load and Energy-efficient RPL(BE-RPL)is proposed.It is a routing technique that is both energy-efficient and mobility-aware.It responds quicker to link breakage through received signal strength-based mobility monitoring and selecting a new preferred parent reactively.The proposed system also implements load balancing among stationary nodes for leaf node allocation.Static nodes with more leaf nodes are restricted from participating in the election for a new preferred parent.The performance of BE-RPL is assessed using the COOJA simulator.It improves the energy use,network control overhead,frame acknowledgment ratio,and packet delivery ratio of the network.
文摘One of the very important functions of three-phase inverter is to maintain the symmetric three-phase output voltage when the three-phase loads are unbalanced. Although the traditional symmetrical component decomposing and superimpose theory can keep the voltage balance through compensating the positive-, negative- and zero-sequence components of the output voltage of inverter, however, this method is time-consuming and not suitable for control. Aiming at high power medium frequency inverter source, a P+Resonant (Proportion and Resonant) controller which ensured a balanced three phase output voltage under unbalanced load is proposed in this paper. The regulator was proved to be applicable to both three-phase three-wire system and three-phase four-wire system and developed two methods of realization. The simulation results verified that this method can suppressed effectively the output voltage distorted caused by the unbalanced load and attained a high quality voltage waveforms.
基金supported supported by the National Key Research and Development Program of China(No.2020YFE0200500)CERNET Innovation Project(NGII20190806)。
文摘At present,the flow table of the SDN switch is stored in the costly Ternary Content Addressable Memory(TCAM)cache.Due to the cost problem,the number of flow tables that the SDN switch can store is extremely limited,which is far less than the number of traffic,so it is prone to overflow problem,and leads to network paralysis.That has become a bottleneck in restricting the processing capacity of the data center,and will become a weak point focused by attackers.In this paper,we propose an algorithm for the Alarm Switch Remove(ASR)that fully loads the flow table space in SDN,and further put forward an integrated load balancing scheme in SDN.Finally,we use Mininet to verify that the scheme can ease the SDN switch flow table overflow problem and increase network throughput.
基金supported in part by the Guangdong Basic and Applied Basic Research Foundation under grant 2020A1515110269.
文摘A major challenge for the future wireless network is to design the self-organizing architecture.The reactive self-organizing model of traditional networks needs to be transformed into an active self-organizing network.Due to the user mobility and the coverage of small cells,the network load often becomes unbalanced,resulting in poor network performance.Mobility management has become an important issue to ensure seamless communication when users move between cells,and proactive mobility management is one of the important functions of the active Self-Organizing Network(SON).This paper proposes a proactive mobility management framework for active SON,which transforms the original reactive load balancing into a forward-aware and proactive load balancing.The proposed framework firstly uses the BART model to predict the users’temporal and spatial mobility based on a weekly cycle and then formulate the MLB optimization problem based on the soft load.Two solutions are proposed to solve the above MLB problem.The simulation results show that the proposed method can better optimize the network performance and realize intelligent mobile management for the future network.
文摘According to the advances in users’service requirements,physical hardware accessibility,and speed of resource delivery,Cloud Computing(CC)is an essential technology to be used in many fields.Moreover,the Internet of Things(IoT)is employed for more communication flexibility and richness that are required to obtain fruitful services.A multi-agent system might be a proper solution to control the load balancing of interaction and communication among agents.This paper proposes a multi-agent load balancing framework that consists of two phases to optimize the workload among different servers with large-scale CC power with various utilities and a significant number of IoT devices with low resources.Different agents are integrated based on relevant features of behavioral interaction using classification techniques to balance the workload.Aload balancing algorithm is developed to serve users’requests to improve the solution of workload problems with an efficient distribution.The activity task from IoT devices has been classified by feature selection methods in the preparatory phase to optimize the scalability ofCC.Then,the server’s availability is checked and the classified task is assigned to its suitable server in the main phase to enhance the cloud environment performance.Multi-agent load balancing framework is succeeded to cope with the importance of using large-scale requirements of CC and(low resources and large number)of IoT.
文摘Every day,more and more data is being produced by the Internet of Things(IoT)applications.IoT data differ in amount,diversity,veracity,and velocity.Because of latency,various types of data handling in cloud computing are not suitable for many time-sensitive applications.When users move from one site to another,mobility also adds to the latency.By placing computing close to IoT devices with mobility support,fog computing addresses these problems.An efficient Load Balancing Algorithm(LBA)improves user experience and Quality of Service(QoS).Classification of Request(CoR)based Resource Adaptive LBA is suggested in this research.This technique clusters fog nodes using an efficient K-means clustering algorithm and then uses a Decision Tree approach to categorize the request.The decision-making process for time-sensitive and delay-tolerable requests is facilitated by the classification of requests.LBA does the operation based on these classifications.The MobFogSim simulation program is utilized to assess how well the algorithm with mobility features performs.The outcome demonstrates that the LBA algorithm’s performance enhances the total system performance,which was attained by(90.8%).Using LBA,several metrics may be examined,including Response Time(RT),delay(d),Energy Consumption(EC),and latency.Through the on-demand provisioning of necessary resources to IoT users,our suggested LBA assures effective resource usage.
文摘As a new networking paradigm,Software-Defined Networking(SDN)enables us to cope with the limitations of traditional networks.SDN uses a controller that has a global view of the network and switch devices which act as packet forwarding hardware,known as“OpenFlow switches”.Since load balancing service is essential to distribute workload across servers in data centers,we propose an effective load balancing scheme in SDN,using a genetic programming approach,called Genetic Programming based Load Balancing(GPLB).We formulate the problem to find a path:1)with the best bottleneck switch which has the lowest capacity within bottleneck switches of each path,2)with the shortest path,and 3)requiring the less possible operations.For the purpose of choosing the real-time least loaded path,GPLB immediately calculates the integrated load of paths based on the information that receives from the SDN controller.Hence,in this design,the controller sends the load information of each path to the load balancing algorithm periodically and then the load balancing algorithm returns a least loaded path to the controller.In this paper,we use the Mininet emulator and the OpenDaylight controller to evaluate the effectiveness of the GPLB.The simulative study of the GPLB shows that there is a big improvement in performance metrics and the latency and the jitter are minimized.The GPLB also has the maximum throughput in comparison with related works and has performed better in the heavy traffic situation.The results show that our model stands smartly while not increasing further overhead.
基金supported by the project:“Research and Implementation of Innovative Solutions for Monitoring Consumption in Technical Installations Using Artificial Intelligence”,beneficiary S.C.REMONI TECHNOLOGIES RO S.R.L in partnership with“Gheorghe Asachi”Technical University of Iasi,Financing Contract No.400/390076/26.11.2021,SMIS Code 121866,financed by POC/163/1/3.
文摘Cloud Technology is a new platform that offers on-demand computing Peripheral such as storage,processing power,and other computer system resources.It is also referred to as a system that will let the consumers utilize computational resources like databases,servers,storage,and intelligence over the Internet.In a cloud network,load balancing is the process of dividing network traffic among a cluster of available servers to increase efficiency.It is also known as a server pool or server farm.When a single node is overwhelmed,balancing the workload is needed to manage unpredictable workflows.The load balancer sends the load to another free node in this case.We focus on the Balancing of workflows with the proposed approach,and we present a novel method to balance the load that manages the dynamic scheduling process.One of the preexisting load balancing techniques is considered,however it is somewhat modified to fit the scenario at hand.Depending on the experimentation’s findings,it is concluded that this suggested approach improves load balancing consistency,response time,and throughput by 6%.
文摘Advancements in cloud computing and virtualization technologies have revolutionized Enterprise Application Development with innovative ways to design and develop complex systems.Microservices Architecture is one of the recent techniques in which Enterprise Systems can be developed as fine-grained smaller components and deployed independently.This methodology brings numerous benefits like scalability,resilience,flexibility in development,faster time to market,etc.and the advantages;Microservices bring some challenges too.Multiple microservices need to be invoked one by one as a chain.In most applications,more than one chain of microservices runs in parallel to complete a particular requirement To complete a user’s request.It results in competition for resources and the need for more inter-service communication among the services,which increases the overall latency of the application.A new approach has been proposed in this paper to handle a complex chain of microservices and reduce the latency of user requests.A machine learning technique is followed to predict the weighting time of different types of requests.The communication time among services distributed among different physical machines are estimated based on that and obtained insights are applied to an algorithm to calculate their priorities dynamically and select suitable service instances to minimize the latency based on the shortest queue waiting time.Experiments were done for both interactive as well as non interactive workloads to test the effectiveness of the solution.The approach has been proved to be very effective in reducing latency in the case of long service chains.
基金funded by University Grant Commission with UGC-Ref.No.:3364/(NET-JUNE 2015).
文摘This paper presents a novel fuzzy firefly-based intelligent algorithm for load balancing in mobile cloud computing while reducing makespan.The proposed technique implicitly acts intelligently by using inherent traits of fuzzy and firefly.It automatically adjusts its behavior or converges depending on the information gathered during the search process and objective function.It works for 3-tier architecture,including cloudlet and public cloud.As cloudlets have limited resources,fuzzy logic is used for cloudlet selection using capacity and waiting time as input.Fuzzy provides human-like decisions without using any mathematical model.Firefly is a powerful meta-heuristic optimization technique to balance diversification and solution speed.It balances the load on cloud and cloudlet while minimizing makespan and execution time.However,it may trap in local optimum;levy flight can handle it.Hybridization of fuzzy fireflywith levy flight is a novel technique that provides reduced makespan,execution time,and Degree of imbalance while balancing the load.Simulation has been carried out on the Cloud Analyst platform with National Aeronautics and Space Administration(NASA)and Clarknet datasets.Results show that the proposed algorithm outperforms Ant Colony Optimization Queue Decision Maker(ACOQDM),Distributed Scheduling Optimization Algorithm(DSOA),andUtility-based Firefly Algorithm(UFA)when compared in terms of makespan,Degree of imbalance,and Figure of Merit.
基金supported in part by the National Natural Science Foundation of China 62072096the Fundamental Research Funds for the Central Universities under Grant 2232020A-12+4 种基金the International S&T Cooperation Program of Shanghai Science and Technology Commission under Grant 20220713000the Young Top-notch Talent Program in Shanghaithe"Shuguang Program"of Shanghai Education Development Foundation and Shanghai Municipal Education Commissionthe Fundamental Research Funds for the Central Universities and Graduate Student Innovation Fund of Donghua University CUSF-DH-D-2019093supported in part by the NSF under grants CNS-2107190 and ECCS-1923717。
文摘Integrating the blockchain technology into mobile-edge computing(MEC)networks with multiple cooperative MEC servers(MECS)providing a promising solution to improving resource utilization,and helping establish a secure reward mechanism that can facilitate load balancing among MECS.In addition,intelligent management of service caching and load balancing can improve the network utility in MEC blockchain networks with multiple types of workloads.In this paper,we investigate a learningbased joint service caching and load balancing policy for optimizing the communication and computation resources allocation,so as to improve the resource utilization of MEC blockchain networks.We formulate the problem as a challenging long-term network revenue maximization Markov decision process(MDP)problem.To address the highly dynamic and high dimension of system states,we design a joint service caching and load balancing algorithm based on the double-dueling Deep Q network(DQN)approach.The simulation results validate the feasibility and superior performance of our proposed algorithm over several baseline schemes.
基金supported by the Bio and Medical Technology Development Program of the National Research Foundation(NRF)funded by the Korean government(MSIT)(No.NRF-2019M3E5D1A02069073)supported by the Soonchunhyang University Research Fund.
文摘Healthcare is a fundamental part of every individual’s life.The healthcare industry is developing very rapidly with the help of advanced technologies.Many researchers are trying to build cloud-based healthcare applications that can be accessed by healthcare professionals from their premises,as well as by patients from their mobile devices through communication interfaces.These systems promote reliable and remote interactions between patients and healthcare professionals.However,there are several limitations to these innovative cloud computing-based systems,namely network availability,latency,battery life and resource availability.We propose a hybrid mobile cloud computing(HMCC)architecture to address these challenges.Furthermore,we also evaluate the performance of heuristic and dynamic machine learning based task scheduling and load balancing algorithms on our proposed architecture.We compare them,to identify the strengths and weaknesses of each algorithm;and provide their comparative results,to show latency and energy consumption performance.Challenging issues for cloudbased healthcare systems are discussed in detail.
文摘In wireless sensor network(WSN),the gateways which are placed far away from the base station(BS)forward the collected data to the BS through the gateways which are nearer to the BS.This leads to more energy consumption because the gateways nearer to the BS manages heavy traffic load.So,to over-come this issue,loads around the gateways are to be balanced by presenting energy efficient clustering approach.Besides,to enhance the lifetime of the net-work,optimal routing path is to be established between the source node and BS.For energy efficient load balancing and routing,multi objective based beetle swarm optimization(BSO)algorithm is presented in this paper.Using this algo-rithm,optimal clustering and routing are performed depend on the objective func-tions routingfitness and clusteringfitness.This approach leads to decrease the power consumption.Simulation results show that the performance of the pro-posed BSO based clustering and routing scheme attains better results than that of the existing algorithms in terms of energy consumption,delivery ratio,through-put and network lifetime.Namely,the proposed scheme increases throughput to 72%and network lifetime to 37%as well as it reduces delay to 37%than the existing optimization algorithms based clustering and routing schemes.