This paper focuses on the scheduling problem of workflow tasks that exhibit interdependencies.Unlike indepen-dent batch tasks,workflows typically consist of multiple subtasks with intrinsic correlations and dependenci...This paper focuses on the scheduling problem of workflow tasks that exhibit interdependencies.Unlike indepen-dent batch tasks,workflows typically consist of multiple subtasks with intrinsic correlations and dependencies.It necessitates the distribution of various computational tasks to appropriate computing node resources in accor-dance with task dependencies to ensure the smooth completion of the entire workflow.Workflow scheduling must consider an array of factors,including task dependencies,availability of computational resources,and the schedulability of tasks.Therefore,this paper delves into the distributed graph database workflow task scheduling problem and proposes a workflow scheduling methodology based on deep reinforcement learning(DRL).The method optimizes the maximum completion time(makespan)and response time of workflow tasks,aiming to enhance the responsiveness of workflow tasks while ensuring the minimization of the makespan.The experimental results indicate that the Q-learning Deep Reinforcement Learning(Q-DRL)algorithm markedly diminishes the makespan and refines the average response time within distributed graph database environments.In quantifying makespan,Q-DRL achieves mean reductions of 12.4%and 11.9%over established First-fit and Random scheduling strategies,respectively.Additionally,Q-DRL surpasses the performance of both DRL-Cloud and Improved Deep Q-learning Network(IDQN)algorithms,with improvements standing at 4.4%and 2.6%,respectively.With reference to average response time,the Q-DRL approach exhibits a significantly enhanced performance in the scheduling of workflow tasks,decreasing the average by 2.27%and 4.71%when compared to IDQN and DRL-Cloud,respectively.The Q-DRL algorithm also demonstrates a notable increase in the efficiency of system resource utilization,reducing the average idle rate by 5.02%and 9.30%in comparison to IDQN and DRL-Cloud,respectively.These findings support the assertion that Q-DRL not only upholds a lower average idle rate but also effectively curtails the average response time,thereby substantially improving processing efficiency and optimizing resource utilization within distributed graph database systems.展开更多
Unbalanced traffic distribution in cellular networks results in congestion and degrades spectrum efficiency.To tackle this problem,we propose an Unmanned Aerial Vehicle(UAV)-assisted wireless network in which the UAV ...Unbalanced traffic distribution in cellular networks results in congestion and degrades spectrum efficiency.To tackle this problem,we propose an Unmanned Aerial Vehicle(UAV)-assisted wireless network in which the UAV acts as an aerial relay to divert some traffic from the overloaded cell to its adjacent underloaded cell.To fully exploit its potential,we jointly optimize the UAV position,user association,spectrum allocation,and power allocation to maximize the sum-log-rate of all users in two adjacent cells.To tackle the complicated joint optimization problem,we first design a genetic-based algorithm to optimize the UAV position.Then,we simplify the problem by theoretical analysis and devise a low-complexity algorithm according to the branch-and-bound method,so as to obtain the optimal user association and spectrum allocation schemes.We further propose an iterative power allocation algorithm based on the sequential convex approximation theory.The simulation results indicate that the proposed UAV-assisted wireless network is superior to the terrestrial network in both utility and throughput,and the proposed algorithms can substantially improve the network performance in comparison with the other schemes.展开更多
Cloud Computing has the ability to provide on-demand access to a shared resource pool.It has completely changed the way businesses are managed,implement applications,and provide services.The rise in popularity has led...Cloud Computing has the ability to provide on-demand access to a shared resource pool.It has completely changed the way businesses are managed,implement applications,and provide services.The rise in popularity has led to a significant increase in the user demand for services.However,in cloud environments efficient load balancing is essential to ensure optimal performance and resource utilization.This systematic review targets a detailed description of load balancing techniques including static and dynamic load balancing algorithms.Specifically,metaheuristic-based dynamic load balancing algorithms are identified as the optimal solution in case of increased traffic.In a cloud-based context,this paper describes load balancing measurements,including the benefits and drawbacks associated with the selected load balancing techniques.It also summarizes the algorithms based on implementation,time complexity,adaptability,associated issue(s),and targeted QoS parameters.Additionally,the analysis evaluates the tools and instruments utilized in each investigated study.Moreover,comparative analysis among static,traditional dynamic and metaheuristic algorithms based on response time by using the CloudSim simulation tool is also performed.Finally,the key open problems and potential directions for the state-of-the-art metaheuristic-based approaches are also addressed.展开更多
This work investigates a multi-product parallel disassembly line balancing problem considering multi-skilled workers.A mathematical model for the parallel disassembly line is established to achieve maximized disassemb...This work investigates a multi-product parallel disassembly line balancing problem considering multi-skilled workers.A mathematical model for the parallel disassembly line is established to achieve maximized disassembly profit and minimized workstation cycle time.Based on a product’s AND/OR graph,matrices for task-skill,worker-skill,precedence relationships,and disassembly correlations are developed.A multi-objective discrete chemical reaction optimization algorithm is designed.To enhance solution diversity,improvements are made to four reactions:decomposition,synthesis,intermolecular ineffective collision,and wall invalid collision reaction,completing the evolution of molecular individuals.The established model and improved algorithm are applied to ball pen,flashlight,washing machine,and radio combinations,respectively.Introducing a Collaborative Resource Allocation(CRA)strategy based on a Decomposition-Based Multi-Objective Evolutionary Algorithm,the experimental results are compared with four classical algorithms:MOEA/D,MOEAD-CRA,Non-dominated Sorting Genetic Algorithm Ⅱ(NSGA-Ⅱ),and Non-dominated Sorting Genetic Algorithm Ⅲ(NSGA-Ⅲ).This validates the feasibility and superiority of the proposed algorithm in parallel disassembly production lines.展开更多
In this paper, a sender-initiated protocol is applied which uses fuzzy logic control method to improve computer networks performance by balancing loads among computers. This new model devises sender-initiated protocol...In this paper, a sender-initiated protocol is applied which uses fuzzy logic control method to improve computer networks performance by balancing loads among computers. This new model devises sender-initiated protocol for load transfer for load balancing. Groups are formed and every group has a node called a designated representative (DR). During load transferring processes, loads are transferred using the DR in each group to achieve load balancing purposes. The simulation results show that the performance of the protocol proposed is better than the compared conventional method. This protocol is more stable than the method without using the fuzzy logic control.展开更多
This paper introduces an innovative approach to the synchronized demand-capacity balance with special focus on sector capacity uncertainty within a centrally controlled collaborative air traffic flow management(ATFM)f...This paper introduces an innovative approach to the synchronized demand-capacity balance with special focus on sector capacity uncertainty within a centrally controlled collaborative air traffic flow management(ATFM)framework.Further with previous study,the uncertainty in capacity is considered as a non-negligible issue regarding multiple reasons,like the impact of weather,the strike of air traffic controllers(ATCOs),the military use of airspace and the spatiotemporal distribution of nonscheduled flights,etc.These recessive factors affect the outcome of traffic flow optimization.In this research,the focus is placed on the impact of sector capacity uncertainty on demand and capacity balancing(DCB)optimization and ATFM,and multiple options,such as delay assignment and rerouting,are intended for regulating the traffic flow.A scenario optimization method for sector capacity in the presence of uncertainties is used to find the approximately optimal solution.The results show that the proposed approach can achieve better demand and capacity balancing and determine perfect integer solutions to ATFM problems,solving large-scale instances(24 h on seven capacity scenarios,with 6255 flights and 8949 trajectories)in 5-15 min.To the best of our knowledge,our experiment is the first to tackle large-scale instances of stochastic ATFM problems within the collaborative ATFM framework.展开更多
With the continuous expansion of the data center network scale, changing network requirements, and increasing pressure on network bandwidth, the traditional network architecture can no longer meet people’s needs. The...With the continuous expansion of the data center network scale, changing network requirements, and increasing pressure on network bandwidth, the traditional network architecture can no longer meet people’s needs. The development of software defined networks has brought new opportunities and challenges to future networks. The data and control separation characteristics of SDN improve the performance of the entire network. Researchers have integrated SDN architecture into data centers to improve network resource utilization and performance. This paper first introduces the basic concepts of SDN and data center networks. Then it discusses SDN-based load balancing mechanisms for data centers from different perspectives. Finally, it summarizes and looks forward to the study on SDN-based load balancing mechanisms and its development trend.展开更多
Cascaded H-bridge inverter(CHBI) with supercapacitors(SCs) and dc-dc stage shows significant promise for medium to high voltage energy storage applications. This paper investigates the voltage balance of capacitors wi...Cascaded H-bridge inverter(CHBI) with supercapacitors(SCs) and dc-dc stage shows significant promise for medium to high voltage energy storage applications. This paper investigates the voltage balance of capacitors within the CHBI, including both the dc-link capacitors and SCs. Balance control over the dc-link capacitor voltages is realized by the dcdc stage in each submodule(SM), while a hybrid modulation strategy(HMS) is implemented in the H-bridge to balance the SC voltages among the SMs. Meanwhile, the dc-link voltage fluctuations are analyzed under the HMS. A virtual voltage variable is introduced to coordinate the balancing of dc-link capacitor voltages and SC voltages. Compared to the balancing method that solely considers the SC voltages, the presented method reduces the dc-link voltage fluctuations without affecting the voltage balance of SCs. Finally, both simulation and experimental results verify the effectiveness of the presented method.展开更多
Electrocatalytic urea synthesis via coupling of nitrate with CO_(2)is considered as a promising alternative to the industrial urea synthetic process.However,the requirement of sub-reaction(NO_(3)RR and CO_(2)RR)activi...Electrocatalytic urea synthesis via coupling of nitrate with CO_(2)is considered as a promising alternative to the industrial urea synthetic process.However,the requirement of sub-reaction(NO_(3)RR and CO_(2)RR)activities for efficient urea synthesis is not clear and the related reaction mechanisms remain obscure.Here,the construction,breaking,and rebuilding of the sub-reaction activity balance would be accompanied by the corresponding regulation in urea synthesis,and the balance of sub-reaction activities was proven to play a vital role in efficient urea synthesis.With rational design,a urea yield rate of 610.6 mg h−1 gcat.−1 was realized on the N-doped carbon electrocatalyst,superior to that of noble-metal electrocatalysts.Based on the operando SRFTIR measurements,we proposed that urea synthesis arises from the coupling of^(*)NO and^(*)CO to generate the key intermediate of^(*)OCNO.This work provides new insights and guidelines into urea synthesis from the aspect of activity balance.展开更多
Basic helix–loop–helix(bHLH)proteins play pivotal roles in plant growth,development,and stress responses.However,the molecular and functional properties of bHLHs have not been fully characterized.In this study,a nov...Basic helix–loop–helix(bHLH)proteins play pivotal roles in plant growth,development,and stress responses.However,the molecular and functional properties of bHLHs have not been fully characterized.In this study,a novel XI subgroup of the bHLH protein gene RcbHLH59 was isolated and identified in rose(Rosa sp.).This gene was induced by salinity stress in both rose leaves and roots,and functioned as a transactivator.Accordingly,silencing RcbHLH59 affected the antioxidant system,^(Na+/K+)balance,and photosynthetic system,thereby reducing salt tolerance,while the transient overexpression of RcbHLH59 improved salinity stress tolerance.Additionally,RcbLHLH59 was found to regulate the expression of sets of pathogenesis-related(PR)genes in RcbHLH59-silenced(TRV-RcbHLH59)and RcbHLH59-overexpressing(RcbHLH59-OE)rose plants.The RcPR4/1 and RcPR5/1 transcript levels showed opposite changes in the TRVRcbHLH59 and RcbHLH59-OE lines,suggesting that these two genes are regulated by RcbHLH59.Further analysis revealed that RcbHLH59 binds to the promoters of RcPR4/1 and RcPR5/1,and that the silencing of RcPR4/1 or RcPR5/1 led to decreased tolerance to salinity stress.Moreover,callose degradation-and deposition-related genes were impaired in RcPR4/1-or RcPR5/1-silenced plants,which displayed a salt tolerance phenotype by balancing the ^(Na+/K+)ratio through callose deposition.Collectively,our data highlight a new RcbLHLH59-RcPRs module that positively regulates salinity stress tolerance by balancing Na^(+)/K^(+)and through callose deposition in rose plants.展开更多
At present,the flow table of the SDN switch is stored in the costly Ternary Content Addressable Memory(TCAM)cache.Due to the cost problem,the number of flow tables that the SDN switch can store is extremely limited,wh...At present,the flow table of the SDN switch is stored in the costly Ternary Content Addressable Memory(TCAM)cache.Due to the cost problem,the number of flow tables that the SDN switch can store is extremely limited,which is far less than the number of traffic,so it is prone to overflow problem,and leads to network paralysis.That has become a bottleneck in restricting the processing capacity of the data center,and will become a weak point focused by attackers.In this paper,we propose an algorithm for the Alarm Switch Remove(ASR)that fully loads the flow table space in SDN,and further put forward an integrated load balancing scheme in SDN.Finally,we use Mininet to verify that the scheme can ease the SDN switch flow table overflow problem and increase network throughput.展开更多
Every day,more and more data is being produced by the Internet of Things(IoT)applications.IoT data differ in amount,diversity,veracity,and velocity.Because of latency,various types of data handling in cloud computing ...Every day,more and more data is being produced by the Internet of Things(IoT)applications.IoT data differ in amount,diversity,veracity,and velocity.Because of latency,various types of data handling in cloud computing are not suitable for many time-sensitive applications.When users move from one site to another,mobility also adds to the latency.By placing computing close to IoT devices with mobility support,fog computing addresses these problems.An efficient Load Balancing Algorithm(LBA)improves user experience and Quality of Service(QoS).Classification of Request(CoR)based Resource Adaptive LBA is suggested in this research.This technique clusters fog nodes using an efficient K-means clustering algorithm and then uses a Decision Tree approach to categorize the request.The decision-making process for time-sensitive and delay-tolerable requests is facilitated by the classification of requests.LBA does the operation based on these classifications.The MobFogSim simulation program is utilized to assess how well the algorithm with mobility features performs.The outcome demonstrates that the LBA algorithm’s performance enhances the total system performance,which was attained by(90.8%).Using LBA,several metrics may be examined,including Response Time(RT),delay(d),Energy Consumption(EC),and latency.Through the on-demand provisioning of necessary resources to IoT users,our suggested LBA assures effective resource usage.展开更多
According to the advances in users’service requirements,physical hardware accessibility,and speed of resource delivery,Cloud Computing(CC)is an essential technology to be used in many fields.Moreover,the Internet of ...According to the advances in users’service requirements,physical hardware accessibility,and speed of resource delivery,Cloud Computing(CC)is an essential technology to be used in many fields.Moreover,the Internet of Things(IoT)is employed for more communication flexibility and richness that are required to obtain fruitful services.A multi-agent system might be a proper solution to control the load balancing of interaction and communication among agents.This paper proposes a multi-agent load balancing framework that consists of two phases to optimize the workload among different servers with large-scale CC power with various utilities and a significant number of IoT devices with low resources.Different agents are integrated based on relevant features of behavioral interaction using classification techniques to balance the workload.Aload balancing algorithm is developed to serve users’requests to improve the solution of workload problems with an efficient distribution.The activity task from IoT devices has been classified by feature selection methods in the preparatory phase to optimize the scalability ofCC.Then,the server’s availability is checked and the classified task is assigned to its suitable server in the main phase to enhance the cloud environment performance.Multi-agent load balancing framework is succeeded to cope with the importance of using large-scale requirements of CC and(low resources and large number)of IoT.展开更多
Cloud Technology is a new platform that offers on-demand computing Peripheral such as storage,processing power,and other computer system resources.It is also referred to as a system that will let the consumers utilize...Cloud Technology is a new platform that offers on-demand computing Peripheral such as storage,processing power,and other computer system resources.It is also referred to as a system that will let the consumers utilize computational resources like databases,servers,storage,and intelligence over the Internet.In a cloud network,load balancing is the process of dividing network traffic among a cluster of available servers to increase efficiency.It is also known as a server pool or server farm.When a single node is overwhelmed,balancing the workload is needed to manage unpredictable workflows.The load balancer sends the load to another free node in this case.We focus on the Balancing of workflows with the proposed approach,and we present a novel method to balance the load that manages the dynamic scheduling process.One of the preexisting load balancing techniques is considered,however it is somewhat modified to fit the scenario at hand.Depending on the experimentation’s findings,it is concluded that this suggested approach improves load balancing consistency,response time,and throughput by 6%.展开更多
Advancements in cloud computing and virtualization technologies have revolutionized Enterprise Application Development with innovative ways to design and develop complex systems.Microservices Architecture is one of th...Advancements in cloud computing and virtualization technologies have revolutionized Enterprise Application Development with innovative ways to design and develop complex systems.Microservices Architecture is one of the recent techniques in which Enterprise Systems can be developed as fine-grained smaller components and deployed independently.This methodology brings numerous benefits like scalability,resilience,flexibility in development,faster time to market,etc.and the advantages;Microservices bring some challenges too.Multiple microservices need to be invoked one by one as a chain.In most applications,more than one chain of microservices runs in parallel to complete a particular requirement To complete a user’s request.It results in competition for resources and the need for more inter-service communication among the services,which increases the overall latency of the application.A new approach has been proposed in this paper to handle a complex chain of microservices and reduce the latency of user requests.A machine learning technique is followed to predict the weighting time of different types of requests.The communication time among services distributed among different physical machines are estimated based on that and obtained insights are applied to an algorithm to calculate their priorities dynamically and select suitable service instances to minimize the latency based on the shortest queue waiting time.Experiments were done for both interactive as well as non interactive workloads to test the effectiveness of the solution.The approach has been proved to be very effective in reducing latency in the case of long service chains.展开更多
A major challenge for the future wireless network is to design the self-organizing architecture.The reactive self-organizing model of traditional networks needs to be transformed into an active self-organizing network...A major challenge for the future wireless network is to design the self-organizing architecture.The reactive self-organizing model of traditional networks needs to be transformed into an active self-organizing network.Due to the user mobility and the coverage of small cells,the network load often becomes unbalanced,resulting in poor network performance.Mobility management has become an important issue to ensure seamless communication when users move between cells,and proactive mobility management is one of the important functions of the active Self-Organizing Network(SON).This paper proposes a proactive mobility management framework for active SON,which transforms the original reactive load balancing into a forward-aware and proactive load balancing.The proposed framework firstly uses the BART model to predict the users’temporal and spatial mobility based on a weekly cycle and then formulate the MLB optimization problem based on the soft load.Two solutions are proposed to solve the above MLB problem.The simulation results show that the proposed method can better optimize the network performance and realize intelligent mobile management for the future network.展开更多
The amount of oxygen blown into the converter is one of the key parameters for the control of the converter blowing process,which directly affects the tap-to-tap time of converter. In this study, a hybrid model based ...The amount of oxygen blown into the converter is one of the key parameters for the control of the converter blowing process,which directly affects the tap-to-tap time of converter. In this study, a hybrid model based on oxygen balance mechanism (OBM) and deep neural network (DNN) was established for predicting oxygen blowing time in converter. A three-step method was utilized in the hybrid model. First, the oxygen consumption volume was predicted by the OBM model and DNN model, respectively. Second, a more accurate oxygen consumption volume was obtained by integrating the OBM model and DNN model. Finally, the converter oxygen blowing time was calculated according to the oxygen consumption volume and the oxygen supply intensity of each heat. The proposed hybrid model was verified using the actual data collected from an integrated steel plant in China, and compared with multiple linear regression model, OBM model, and neural network model including extreme learning machine, back propagation neural network, and DNN. The test results indicate that the hybrid model with a network structure of 3 hidden layer layers, 32-16-8 neurons per hidden layer, and 0.1 learning rate has the best prediction accuracy and stronger generalization ability compared with other models. The predicted hit ratio of oxygen consumption volume within the error±300 m^(3)is 96.67%;determination coefficient (R^(2)) and root mean square error (RMSE) are0.6984 and 150.03 m^(3), respectively. The oxygen blow time prediction hit ratio within the error±0.6 min is 89.50%;R2and RMSE are0.9486 and 0.3592 min, respectively. As a result, the proposed model can effectively predict the oxygen consumption volume and oxygen blowing time in the converter.展开更多
Healthcare is a fundamental part of every individual’s life.The healthcare industry is developing very rapidly with the help of advanced technologies.Many researchers are trying to build cloud-based healthcare applic...Healthcare is a fundamental part of every individual’s life.The healthcare industry is developing very rapidly with the help of advanced technologies.Many researchers are trying to build cloud-based healthcare applications that can be accessed by healthcare professionals from their premises,as well as by patients from their mobile devices through communication interfaces.These systems promote reliable and remote interactions between patients and healthcare professionals.However,there are several limitations to these innovative cloud computing-based systems,namely network availability,latency,battery life and resource availability.We propose a hybrid mobile cloud computing(HMCC)architecture to address these challenges.Furthermore,we also evaluate the performance of heuristic and dynamic machine learning based task scheduling and load balancing algorithms on our proposed architecture.We compare them,to identify the strengths and weaknesses of each algorithm;and provide their comparative results,to show latency and energy consumption performance.Challenging issues for cloudbased healthcare systems are discussed in detail.展开更多
This paper presents a novel fuzzy firefly-based intelligent algorithm for load balancing in mobile cloud computing while reducing makespan.The proposed technique implicitly acts intelligently by using inherent traits ...This paper presents a novel fuzzy firefly-based intelligent algorithm for load balancing in mobile cloud computing while reducing makespan.The proposed technique implicitly acts intelligently by using inherent traits of fuzzy and firefly.It automatically adjusts its behavior or converges depending on the information gathered during the search process and objective function.It works for 3-tier architecture,including cloudlet and public cloud.As cloudlets have limited resources,fuzzy logic is used for cloudlet selection using capacity and waiting time as input.Fuzzy provides human-like decisions without using any mathematical model.Firefly is a powerful meta-heuristic optimization technique to balance diversification and solution speed.It balances the load on cloud and cloudlet while minimizing makespan and execution time.However,it may trap in local optimum;levy flight can handle it.Hybridization of fuzzy fireflywith levy flight is a novel technique that provides reduced makespan,execution time,and Degree of imbalance while balancing the load.Simulation has been carried out on the Cloud Analyst platform with National Aeronautics and Space Administration(NASA)and Clarknet datasets.Results show that the proposed algorithm outperforms Ant Colony Optimization Queue Decision Maker(ACOQDM),Distributed Scheduling Optimization Algorithm(DSOA),andUtility-based Firefly Algorithm(UFA)when compared in terms of makespan,Degree of imbalance,and Figure of Merit.展开更多
Integrating the blockchain technology into mobile-edge computing(MEC)networks with multiple cooperative MEC servers(MECS)providing a promising solution to improving resource utilization,and helping establish a secure ...Integrating the blockchain technology into mobile-edge computing(MEC)networks with multiple cooperative MEC servers(MECS)providing a promising solution to improving resource utilization,and helping establish a secure reward mechanism that can facilitate load balancing among MECS.In addition,intelligent management of service caching and load balancing can improve the network utility in MEC blockchain networks with multiple types of workloads.In this paper,we investigate a learningbased joint service caching and load balancing policy for optimizing the communication and computation resources allocation,so as to improve the resource utilization of MEC blockchain networks.We formulate the problem as a challenging long-term network revenue maximization Markov decision process(MDP)problem.To address the highly dynamic and high dimension of system states,we design a joint service caching and load balancing algorithm based on the double-dueling Deep Q network(DQN)approach.The simulation results validate the feasibility and superior performance of our proposed algorithm over several baseline schemes.展开更多
基金funded by the Science and Technology Foundation of State Grid Corporation of China(Grant No.5108-202218280A-2-397-XG).
文摘This paper focuses on the scheduling problem of workflow tasks that exhibit interdependencies.Unlike indepen-dent batch tasks,workflows typically consist of multiple subtasks with intrinsic correlations and dependencies.It necessitates the distribution of various computational tasks to appropriate computing node resources in accor-dance with task dependencies to ensure the smooth completion of the entire workflow.Workflow scheduling must consider an array of factors,including task dependencies,availability of computational resources,and the schedulability of tasks.Therefore,this paper delves into the distributed graph database workflow task scheduling problem and proposes a workflow scheduling methodology based on deep reinforcement learning(DRL).The method optimizes the maximum completion time(makespan)and response time of workflow tasks,aiming to enhance the responsiveness of workflow tasks while ensuring the minimization of the makespan.The experimental results indicate that the Q-learning Deep Reinforcement Learning(Q-DRL)algorithm markedly diminishes the makespan and refines the average response time within distributed graph database environments.In quantifying makespan,Q-DRL achieves mean reductions of 12.4%and 11.9%over established First-fit and Random scheduling strategies,respectively.Additionally,Q-DRL surpasses the performance of both DRL-Cloud and Improved Deep Q-learning Network(IDQN)algorithms,with improvements standing at 4.4%and 2.6%,respectively.With reference to average response time,the Q-DRL approach exhibits a significantly enhanced performance in the scheduling of workflow tasks,decreasing the average by 2.27%and 4.71%when compared to IDQN and DRL-Cloud,respectively.The Q-DRL algorithm also demonstrates a notable increase in the efficiency of system resource utilization,reducing the average idle rate by 5.02%and 9.30%in comparison to IDQN and DRL-Cloud,respectively.These findings support the assertion that Q-DRL not only upholds a lower average idle rate but also effectively curtails the average response time,thereby substantially improving processing efficiency and optimizing resource utilization within distributed graph database systems.
基金supported in part by the National Key Research and Development Program of China under Grant 2020YFB1807003in part by the National Natural Science Foundation of China under Grants 61901381,62171385,and 61901378+3 种基金in part by the Aeronautical Science Foundation of China under Grant 2020z073053004in part by the Foundation of the State Key Laboratory of Integrated Services Networks of Xidian University under Grant ISN21-06in part by the Key Research Program and Industrial Innovation Chain Project of Shaanxi Province under Grant 2019ZDLGY07-10in part by the Natural Science Fundamental Research Program of Shaanxi Province under Grant 2021JM-069.
文摘Unbalanced traffic distribution in cellular networks results in congestion and degrades spectrum efficiency.To tackle this problem,we propose an Unmanned Aerial Vehicle(UAV)-assisted wireless network in which the UAV acts as an aerial relay to divert some traffic from the overloaded cell to its adjacent underloaded cell.To fully exploit its potential,we jointly optimize the UAV position,user association,spectrum allocation,and power allocation to maximize the sum-log-rate of all users in two adjacent cells.To tackle the complicated joint optimization problem,we first design a genetic-based algorithm to optimize the UAV position.Then,we simplify the problem by theoretical analysis and devise a low-complexity algorithm according to the branch-and-bound method,so as to obtain the optimal user association and spectrum allocation schemes.We further propose an iterative power allocation algorithm based on the sequential convex approximation theory.The simulation results indicate that the proposed UAV-assisted wireless network is superior to the terrestrial network in both utility and throughput,and the proposed algorithms can substantially improve the network performance in comparison with the other schemes.
文摘Cloud Computing has the ability to provide on-demand access to a shared resource pool.It has completely changed the way businesses are managed,implement applications,and provide services.The rise in popularity has led to a significant increase in the user demand for services.However,in cloud environments efficient load balancing is essential to ensure optimal performance and resource utilization.This systematic review targets a detailed description of load balancing techniques including static and dynamic load balancing algorithms.Specifically,metaheuristic-based dynamic load balancing algorithms are identified as the optimal solution in case of increased traffic.In a cloud-based context,this paper describes load balancing measurements,including the benefits and drawbacks associated with the selected load balancing techniques.It also summarizes the algorithms based on implementation,time complexity,adaptability,associated issue(s),and targeted QoS parameters.Additionally,the analysis evaluates the tools and instruments utilized in each investigated study.Moreover,comparative analysis among static,traditional dynamic and metaheuristic algorithms based on response time by using the CloudSim simulation tool is also performed.Finally,the key open problems and potential directions for the state-of-the-art metaheuristic-based approaches are also addressed.
文摘This work investigates a multi-product parallel disassembly line balancing problem considering multi-skilled workers.A mathematical model for the parallel disassembly line is established to achieve maximized disassembly profit and minimized workstation cycle time.Based on a product’s AND/OR graph,matrices for task-skill,worker-skill,precedence relationships,and disassembly correlations are developed.A multi-objective discrete chemical reaction optimization algorithm is designed.To enhance solution diversity,improvements are made to four reactions:decomposition,synthesis,intermolecular ineffective collision,and wall invalid collision reaction,completing the evolution of molecular individuals.The established model and improved algorithm are applied to ball pen,flashlight,washing machine,and radio combinations,respectively.Introducing a Collaborative Resource Allocation(CRA)strategy based on a Decomposition-Based Multi-Objective Evolutionary Algorithm,the experimental results are compared with four classical algorithms:MOEA/D,MOEAD-CRA,Non-dominated Sorting Genetic Algorithm Ⅱ(NSGA-Ⅱ),and Non-dominated Sorting Genetic Algorithm Ⅲ(NSGA-Ⅲ).This validates the feasibility and superiority of the proposed algorithm in parallel disassembly production lines.
文摘In this paper, a sender-initiated protocol is applied which uses fuzzy logic control method to improve computer networks performance by balancing loads among computers. This new model devises sender-initiated protocol for load transfer for load balancing. Groups are formed and every group has a node called a designated representative (DR). During load transferring processes, loads are transferred using the DR in each group to achieve load balancing purposes. The simulation results show that the performance of the protocol proposed is better than the compared conventional method. This protocol is more stable than the method without using the fuzzy logic control.
文摘This paper introduces an innovative approach to the synchronized demand-capacity balance with special focus on sector capacity uncertainty within a centrally controlled collaborative air traffic flow management(ATFM)framework.Further with previous study,the uncertainty in capacity is considered as a non-negligible issue regarding multiple reasons,like the impact of weather,the strike of air traffic controllers(ATCOs),the military use of airspace and the spatiotemporal distribution of nonscheduled flights,etc.These recessive factors affect the outcome of traffic flow optimization.In this research,the focus is placed on the impact of sector capacity uncertainty on demand and capacity balancing(DCB)optimization and ATFM,and multiple options,such as delay assignment and rerouting,are intended for regulating the traffic flow.A scenario optimization method for sector capacity in the presence of uncertainties is used to find the approximately optimal solution.The results show that the proposed approach can achieve better demand and capacity balancing and determine perfect integer solutions to ATFM problems,solving large-scale instances(24 h on seven capacity scenarios,with 6255 flights and 8949 trajectories)in 5-15 min.To the best of our knowledge,our experiment is the first to tackle large-scale instances of stochastic ATFM problems within the collaborative ATFM framework.
文摘With the continuous expansion of the data center network scale, changing network requirements, and increasing pressure on network bandwidth, the traditional network architecture can no longer meet people’s needs. The development of software defined networks has brought new opportunities and challenges to future networks. The data and control separation characteristics of SDN improve the performance of the entire network. Researchers have integrated SDN architecture into data centers to improve network resource utilization and performance. This paper first introduces the basic concepts of SDN and data center networks. Then it discusses SDN-based load balancing mechanisms for data centers from different perspectives. Finally, it summarizes and looks forward to the study on SDN-based load balancing mechanisms and its development trend.
基金supported in part by the CAS Project for Young Scientists in Basic Research under Grant No. YSBR-045the Youth Innovation Promotion Association CAS under Grant 2022137the Institute of Electrical Engineering CAS under Grant E155320101。
文摘Cascaded H-bridge inverter(CHBI) with supercapacitors(SCs) and dc-dc stage shows significant promise for medium to high voltage energy storage applications. This paper investigates the voltage balance of capacitors within the CHBI, including both the dc-link capacitors and SCs. Balance control over the dc-link capacitor voltages is realized by the dcdc stage in each submodule(SM), while a hybrid modulation strategy(HMS) is implemented in the H-bridge to balance the SC voltages among the SMs. Meanwhile, the dc-link voltage fluctuations are analyzed under the HMS. A virtual voltage variable is introduced to coordinate the balancing of dc-link capacitor voltages and SC voltages. Compared to the balancing method that solely considers the SC voltages, the presented method reduces the dc-link voltage fluctuations without affecting the voltage balance of SCs. Finally, both simulation and experimental results verify the effectiveness of the presented method.
基金National Key R&D Program of China,Grant/Award Number:2020YFA0710000National Natural Science Foundation of China,Grant/Award Numbers:21573066,21902047,21825201,22075075,22173048,and U1932212China Postdoctoral Science Foundation,Grant/Award Numbers:2020M682540,BX20200116。
文摘Electrocatalytic urea synthesis via coupling of nitrate with CO_(2)is considered as a promising alternative to the industrial urea synthetic process.However,the requirement of sub-reaction(NO_(3)RR and CO_(2)RR)activities for efficient urea synthesis is not clear and the related reaction mechanisms remain obscure.Here,the construction,breaking,and rebuilding of the sub-reaction activity balance would be accompanied by the corresponding regulation in urea synthesis,and the balance of sub-reaction activities was proven to play a vital role in efficient urea synthesis.With rational design,a urea yield rate of 610.6 mg h−1 gcat.−1 was realized on the N-doped carbon electrocatalyst,superior to that of noble-metal electrocatalysts.Based on the operando SRFTIR measurements,we proposed that urea synthesis arises from the coupling of^(*)NO and^(*)CO to generate the key intermediate of^(*)OCNO.This work provides new insights and guidelines into urea synthesis from the aspect of activity balance.
基金supported by the National Key Research and Development Program(2018YFD1000400)National Natural Science Foundation of China(Grant No.32002084).
文摘Basic helix–loop–helix(bHLH)proteins play pivotal roles in plant growth,development,and stress responses.However,the molecular and functional properties of bHLHs have not been fully characterized.In this study,a novel XI subgroup of the bHLH protein gene RcbHLH59 was isolated and identified in rose(Rosa sp.).This gene was induced by salinity stress in both rose leaves and roots,and functioned as a transactivator.Accordingly,silencing RcbHLH59 affected the antioxidant system,^(Na+/K+)balance,and photosynthetic system,thereby reducing salt tolerance,while the transient overexpression of RcbHLH59 improved salinity stress tolerance.Additionally,RcbLHLH59 was found to regulate the expression of sets of pathogenesis-related(PR)genes in RcbHLH59-silenced(TRV-RcbHLH59)and RcbHLH59-overexpressing(RcbHLH59-OE)rose plants.The RcPR4/1 and RcPR5/1 transcript levels showed opposite changes in the TRVRcbHLH59 and RcbHLH59-OE lines,suggesting that these two genes are regulated by RcbHLH59.Further analysis revealed that RcbHLH59 binds to the promoters of RcPR4/1 and RcPR5/1,and that the silencing of RcPR4/1 or RcPR5/1 led to decreased tolerance to salinity stress.Moreover,callose degradation-and deposition-related genes were impaired in RcPR4/1-or RcPR5/1-silenced plants,which displayed a salt tolerance phenotype by balancing the ^(Na+/K+)ratio through callose deposition.Collectively,our data highlight a new RcbLHLH59-RcPRs module that positively regulates salinity stress tolerance by balancing Na^(+)/K^(+)and through callose deposition in rose plants.
基金supported supported by the National Key Research and Development Program of China(No.2020YFE0200500)CERNET Innovation Project(NGII20190806)。
文摘At present,the flow table of the SDN switch is stored in the costly Ternary Content Addressable Memory(TCAM)cache.Due to the cost problem,the number of flow tables that the SDN switch can store is extremely limited,which is far less than the number of traffic,so it is prone to overflow problem,and leads to network paralysis.That has become a bottleneck in restricting the processing capacity of the data center,and will become a weak point focused by attackers.In this paper,we propose an algorithm for the Alarm Switch Remove(ASR)that fully loads the flow table space in SDN,and further put forward an integrated load balancing scheme in SDN.Finally,we use Mininet to verify that the scheme can ease the SDN switch flow table overflow problem and increase network throughput.
文摘Every day,more and more data is being produced by the Internet of Things(IoT)applications.IoT data differ in amount,diversity,veracity,and velocity.Because of latency,various types of data handling in cloud computing are not suitable for many time-sensitive applications.When users move from one site to another,mobility also adds to the latency.By placing computing close to IoT devices with mobility support,fog computing addresses these problems.An efficient Load Balancing Algorithm(LBA)improves user experience and Quality of Service(QoS).Classification of Request(CoR)based Resource Adaptive LBA is suggested in this research.This technique clusters fog nodes using an efficient K-means clustering algorithm and then uses a Decision Tree approach to categorize the request.The decision-making process for time-sensitive and delay-tolerable requests is facilitated by the classification of requests.LBA does the operation based on these classifications.The MobFogSim simulation program is utilized to assess how well the algorithm with mobility features performs.The outcome demonstrates that the LBA algorithm’s performance enhances the total system performance,which was attained by(90.8%).Using LBA,several metrics may be examined,including Response Time(RT),delay(d),Energy Consumption(EC),and latency.Through the on-demand provisioning of necessary resources to IoT users,our suggested LBA assures effective resource usage.
文摘According to the advances in users’service requirements,physical hardware accessibility,and speed of resource delivery,Cloud Computing(CC)is an essential technology to be used in many fields.Moreover,the Internet of Things(IoT)is employed for more communication flexibility and richness that are required to obtain fruitful services.A multi-agent system might be a proper solution to control the load balancing of interaction and communication among agents.This paper proposes a multi-agent load balancing framework that consists of two phases to optimize the workload among different servers with large-scale CC power with various utilities and a significant number of IoT devices with low resources.Different agents are integrated based on relevant features of behavioral interaction using classification techniques to balance the workload.Aload balancing algorithm is developed to serve users’requests to improve the solution of workload problems with an efficient distribution.The activity task from IoT devices has been classified by feature selection methods in the preparatory phase to optimize the scalability ofCC.Then,the server’s availability is checked and the classified task is assigned to its suitable server in the main phase to enhance the cloud environment performance.Multi-agent load balancing framework is succeeded to cope with the importance of using large-scale requirements of CC and(low resources and large number)of IoT.
基金supported by the project:“Research and Implementation of Innovative Solutions for Monitoring Consumption in Technical Installations Using Artificial Intelligence”,beneficiary S.C.REMONI TECHNOLOGIES RO S.R.L in partnership with“Gheorghe Asachi”Technical University of Iasi,Financing Contract No.400/390076/26.11.2021,SMIS Code 121866,financed by POC/163/1/3.
文摘Cloud Technology is a new platform that offers on-demand computing Peripheral such as storage,processing power,and other computer system resources.It is also referred to as a system that will let the consumers utilize computational resources like databases,servers,storage,and intelligence over the Internet.In a cloud network,load balancing is the process of dividing network traffic among a cluster of available servers to increase efficiency.It is also known as a server pool or server farm.When a single node is overwhelmed,balancing the workload is needed to manage unpredictable workflows.The load balancer sends the load to another free node in this case.We focus on the Balancing of workflows with the proposed approach,and we present a novel method to balance the load that manages the dynamic scheduling process.One of the preexisting load balancing techniques is considered,however it is somewhat modified to fit the scenario at hand.Depending on the experimentation’s findings,it is concluded that this suggested approach improves load balancing consistency,response time,and throughput by 6%.
文摘Advancements in cloud computing and virtualization technologies have revolutionized Enterprise Application Development with innovative ways to design and develop complex systems.Microservices Architecture is one of the recent techniques in which Enterprise Systems can be developed as fine-grained smaller components and deployed independently.This methodology brings numerous benefits like scalability,resilience,flexibility in development,faster time to market,etc.and the advantages;Microservices bring some challenges too.Multiple microservices need to be invoked one by one as a chain.In most applications,more than one chain of microservices runs in parallel to complete a particular requirement To complete a user’s request.It results in competition for resources and the need for more inter-service communication among the services,which increases the overall latency of the application.A new approach has been proposed in this paper to handle a complex chain of microservices and reduce the latency of user requests.A machine learning technique is followed to predict the weighting time of different types of requests.The communication time among services distributed among different physical machines are estimated based on that and obtained insights are applied to an algorithm to calculate their priorities dynamically and select suitable service instances to minimize the latency based on the shortest queue waiting time.Experiments were done for both interactive as well as non interactive workloads to test the effectiveness of the solution.The approach has been proved to be very effective in reducing latency in the case of long service chains.
基金supported in part by the Guangdong Basic and Applied Basic Research Foundation under grant 2020A1515110269.
文摘A major challenge for the future wireless network is to design the self-organizing architecture.The reactive self-organizing model of traditional networks needs to be transformed into an active self-organizing network.Due to the user mobility and the coverage of small cells,the network load often becomes unbalanced,resulting in poor network performance.Mobility management has become an important issue to ensure seamless communication when users move between cells,and proactive mobility management is one of the important functions of the active Self-Organizing Network(SON).This paper proposes a proactive mobility management framework for active SON,which transforms the original reactive load balancing into a forward-aware and proactive load balancing.The proposed framework firstly uses the BART model to predict the users’temporal and spatial mobility based on a weekly cycle and then formulate the MLB optimization problem based on the soft load.Two solutions are proposed to solve the above MLB problem.The simulation results show that the proposed method can better optimize the network performance and realize intelligent mobile management for the future network.
基金financially supported by the National Natural Science Foundation of China (Nos.51974023 and52374321)the funding of State Key Laboratory of Advanced Metallurgy,University of Science and Technology Beijing,China (No.41620007)。
文摘The amount of oxygen blown into the converter is one of the key parameters for the control of the converter blowing process,which directly affects the tap-to-tap time of converter. In this study, a hybrid model based on oxygen balance mechanism (OBM) and deep neural network (DNN) was established for predicting oxygen blowing time in converter. A three-step method was utilized in the hybrid model. First, the oxygen consumption volume was predicted by the OBM model and DNN model, respectively. Second, a more accurate oxygen consumption volume was obtained by integrating the OBM model and DNN model. Finally, the converter oxygen blowing time was calculated according to the oxygen consumption volume and the oxygen supply intensity of each heat. The proposed hybrid model was verified using the actual data collected from an integrated steel plant in China, and compared with multiple linear regression model, OBM model, and neural network model including extreme learning machine, back propagation neural network, and DNN. The test results indicate that the hybrid model with a network structure of 3 hidden layer layers, 32-16-8 neurons per hidden layer, and 0.1 learning rate has the best prediction accuracy and stronger generalization ability compared with other models. The predicted hit ratio of oxygen consumption volume within the error±300 m^(3)is 96.67%;determination coefficient (R^(2)) and root mean square error (RMSE) are0.6984 and 150.03 m^(3), respectively. The oxygen blow time prediction hit ratio within the error±0.6 min is 89.50%;R2and RMSE are0.9486 and 0.3592 min, respectively. As a result, the proposed model can effectively predict the oxygen consumption volume and oxygen blowing time in the converter.
基金supported by the Bio and Medical Technology Development Program of the National Research Foundation(NRF)funded by the Korean government(MSIT)(No.NRF-2019M3E5D1A02069073)supported by the Soonchunhyang University Research Fund.
文摘Healthcare is a fundamental part of every individual’s life.The healthcare industry is developing very rapidly with the help of advanced technologies.Many researchers are trying to build cloud-based healthcare applications that can be accessed by healthcare professionals from their premises,as well as by patients from their mobile devices through communication interfaces.These systems promote reliable and remote interactions between patients and healthcare professionals.However,there are several limitations to these innovative cloud computing-based systems,namely network availability,latency,battery life and resource availability.We propose a hybrid mobile cloud computing(HMCC)architecture to address these challenges.Furthermore,we also evaluate the performance of heuristic and dynamic machine learning based task scheduling and load balancing algorithms on our proposed architecture.We compare them,to identify the strengths and weaknesses of each algorithm;and provide their comparative results,to show latency and energy consumption performance.Challenging issues for cloudbased healthcare systems are discussed in detail.
基金funded by University Grant Commission with UGC-Ref.No.:3364/(NET-JUNE 2015).
文摘This paper presents a novel fuzzy firefly-based intelligent algorithm for load balancing in mobile cloud computing while reducing makespan.The proposed technique implicitly acts intelligently by using inherent traits of fuzzy and firefly.It automatically adjusts its behavior or converges depending on the information gathered during the search process and objective function.It works for 3-tier architecture,including cloudlet and public cloud.As cloudlets have limited resources,fuzzy logic is used for cloudlet selection using capacity and waiting time as input.Fuzzy provides human-like decisions without using any mathematical model.Firefly is a powerful meta-heuristic optimization technique to balance diversification and solution speed.It balances the load on cloud and cloudlet while minimizing makespan and execution time.However,it may trap in local optimum;levy flight can handle it.Hybridization of fuzzy fireflywith levy flight is a novel technique that provides reduced makespan,execution time,and Degree of imbalance while balancing the load.Simulation has been carried out on the Cloud Analyst platform with National Aeronautics and Space Administration(NASA)and Clarknet datasets.Results show that the proposed algorithm outperforms Ant Colony Optimization Queue Decision Maker(ACOQDM),Distributed Scheduling Optimization Algorithm(DSOA),andUtility-based Firefly Algorithm(UFA)when compared in terms of makespan,Degree of imbalance,and Figure of Merit.
基金supported in part by the National Natural Science Foundation of China 62072096the Fundamental Research Funds for the Central Universities under Grant 2232020A-12+4 种基金the International S&T Cooperation Program of Shanghai Science and Technology Commission under Grant 20220713000the Young Top-notch Talent Program in Shanghaithe"Shuguang Program"of Shanghai Education Development Foundation and Shanghai Municipal Education Commissionthe Fundamental Research Funds for the Central Universities and Graduate Student Innovation Fund of Donghua University CUSF-DH-D-2019093supported in part by the NSF under grants CNS-2107190 and ECCS-1923717。
文摘Integrating the blockchain technology into mobile-edge computing(MEC)networks with multiple cooperative MEC servers(MECS)providing a promising solution to improving resource utilization,and helping establish a secure reward mechanism that can facilitate load balancing among MECS.In addition,intelligent management of service caching and load balancing can improve the network utility in MEC blockchain networks with multiple types of workloads.In this paper,we investigate a learningbased joint service caching and load balancing policy for optimizing the communication and computation resources allocation,so as to improve the resource utilization of MEC blockchain networks.We formulate the problem as a challenging long-term network revenue maximization Markov decision process(MDP)problem.To address the highly dynamic and high dimension of system states,we design a joint service caching and load balancing algorithm based on the double-dueling Deep Q network(DQN)approach.The simulation results validate the feasibility and superior performance of our proposed algorithm over several baseline schemes.