In this paper,a new communication model is built named grouping D2D(GD2D).Different from the traditional D2D coordination,we proposed GD2D communication in licensed and unlicensed spectrum simultaneously.We formulate ...In this paper,a new communication model is built named grouping D2D(GD2D).Different from the traditional D2D coordination,we proposed GD2D communication in licensed and unlicensed spectrum simultaneously.We formulate a resource allocation problem,which aims at maximizing the energy efficiency(EE)of the system while guaranteeing the quality-of-service(Qos)of users.To efficiently solve this problem,the non-convex optimization problem is first transformed into a convex optimization problem.By transforming the fractional-form problem into an equivalent subtractive-form problem,an iterative power allocation algorithm is proposed to maximize the system EE.Moreover,the optimal closedform power allocation expressions are derived by the Lagrangian approach.Simulation results show that our algorithm achieves higher EE performance than the traditional D2D communication scheme.展开更多
With the rapid increasing of maritime activities, maritime wireless networks(MWNs) with high reliability, high energy efficiency, and low delay are required. However, the centralized networking with fixed resource sch...With the rapid increasing of maritime activities, maritime wireless networks(MWNs) with high reliability, high energy efficiency, and low delay are required. However, the centralized networking with fixed resource scheduling is not suitable for MWNs due to the special environment. In this paper,we introduce the collaborative relay communication in distributed MWNs to improve the link reliability, and propose an orthogonal time-frequency resource block reservation based multiple access(RRMA) scheme for both one-hop direct link and two-hop collaborative relay link to reduce the interference. To further improve the network performance, we formulate an energy efficiency(EE) maximization resource allocation problem and solve it by an iterative algorithm based on the Dinkelbach method. Finally, numerical results are provided to investigate the proposed RRMA scheme and resource allocation algorithm, showing that the low outage probability and transmission delay can be attained by the proposed RRMA scheme. Moreover,the proposed resource allocation algorithm is capable of achieving high EE in distributed MWNs.展开更多
The small-cell technology is promising for spectral-efficiency enhancement. However, it usually requires a huge amount of energy consumption. In this paper, queue state information and channel state information are jo...The small-cell technology is promising for spectral-efficiency enhancement. However, it usually requires a huge amount of energy consumption. In this paper, queue state information and channel state information are jointly utilized to minimize the time average of overall energy consumption for a multi-carrier small-cell network, where the inter-cell interference is an intractable problem. Based on the Lyapunov optimization theory, the problem could be solved by dynamically optimizing the problem of user assignment, carrier allocation and power allocation in each time slot. As the optimization problem is NP-hard, we propose a heuristic iteration algorithm to solve it. Numerical results verify that the heuristic algorithm offers an approximate performance as the brute-force algorithm. Moreover, it could bring down the overall energy consumption to different degrees according to the variation of traffic load. Meanwhile, it could achieve the same sum rate as the algorithm which focuses on maximizing system sum rate.展开更多
In order to maximize system energy efficiency(EE) under user quality of service(Qo S) restraints in Long Term Evolution-Advanced(LTE-A) networks,a constrained joint resource optimization allocation scheme is presented...In order to maximize system energy efficiency(EE) under user quality of service(Qo S) restraints in Long Term Evolution-Advanced(LTE-A) networks,a constrained joint resource optimization allocation scheme is presented,which is NP-hard. Hence,we divide it into three sub-problems to reduce computation complexity,i.e.,the resource block(RB) allocation,the power distribution,and the modulation and coding scheme(MCS) assignment for user codewords. Then an enhanced heuristic approach GAPSO is proposed and is adopted in the RB and power allocation respectively to reduce computational complexity further on. Moreover,a novel MCS allocation scheme is put forward,which could make a good balance between the system reliability and availability under different channel conditions. Simulation results show that the proposed GAPSO could achieve better performance in convergence speed and global optimum searching,and that the joint resource allocation scheme could improve energy efficiency effectively under user Qo S requirements.展开更多
The vehicular edge computing(VEC)is a new paradigm that allows vehicles to offload computational tasks to base stations(BSs)with edge servers for computing.In general,the VEC paradigm uses the 5G for wireless communic...The vehicular edge computing(VEC)is a new paradigm that allows vehicles to offload computational tasks to base stations(BSs)with edge servers for computing.In general,the VEC paradigm uses the 5G for wireless communications,where the massive multi-input multi-output(MIMO)technique will be used.However,considering in the VEC environment with many vehicles,the energy consumption of BS may be very large.In this paper,we study the energy optimization problem for the massive MIMO-based VEC network.Aiming at reducing the relevant BS energy consumption,we first propose a joint optimization problem of computation resource allocation,beam allocation and vehicle grouping scheme.Since the original problem is hard to be solved directly,we try to split the original problem into two subproblems and then design a heuristic algorithm to solve them.Simulation results show that our proposed algorithm efficiently reduces the BS energy consumption compared to other schemes.展开更多
Federated Edge Learning(FEL),an emerging distributed Machine Learning(ML)paradigm,enables model training in a distributed environment while ensuring user privacy by using physical separation for each user’s data.Howe...Federated Edge Learning(FEL),an emerging distributed Machine Learning(ML)paradigm,enables model training in a distributed environment while ensuring user privacy by using physical separation for each user’s data.However,with the development of complex application scenarios such as the Internet of Things(IoT)and Smart Earth,the conventional resource allocation schemes can no longer effectively support these growing computational and communication demands.Therefore,joint resource optimization may be the key solution to the scaling problem.This paper simultaneously addresses the multifaceted challenges of computation and communication,with the growing multiple resource demands.We systematically review the joint allocation strategies for different resources(computation,data,communication,and network topology)in FEL,and summarize the advantages in improving system efficiency,reducing latency,enhancing resource utilization,and enhancing robustness.In addition,we present the potential ability of joint optimization to enhance privacy preservation by reducing communication requirements,indirectly.This work not only provides theoretical support for resource management in federated learning(FL)systems,but also provides ideas for potential optimal deployment in multiple real-world scenarios.By thoroughly discussing the current challenges and future research directions,it also provides some important insights into multi-resource optimization in complex application environments.展开更多
The performance of massive MIMO systems relies heavily on the availability of Channel State Information at the Transmitter(CSIT).A large amount of work has been devoted to reducing the CSIT acquisition overhead at the...The performance of massive MIMO systems relies heavily on the availability of Channel State Information at the Transmitter(CSIT).A large amount of work has been devoted to reducing the CSIT acquisition overhead at the pilot training and/or CsI feedback stage.In fact,the downlink communication generally includes three stages,i.e.,pilot training,CsI feedback,and data transmission.These three stages are mutually related and jointly determine the overall system performance.Unfortunately,there exist few studies on the reduction of csIT acquisition overhead from the global point of view.In this paper,we integrate the Minimum Mean Square Error(MMSE)channel estimation,Random Vector Quantization(RVQ)based limited feedback and Maximal Ratio Combining(MRC)precoding into a unified framework for investigating the resource allocation problem.In particular,we first approximate the covariance matrix of the quantization error with a simple expression and derive an analytical expression of the received Signal-to-Noise Ratio(SNR)based on the deterministic equivalence theory.Then the three performance metrics(the spectral efficiency,energy efficiency,and total energy consumption)oriented problems are formulated analytically.With practical system requirements,these three metrics can be collaboratively optimized.Finally,we propose an optimization solver to derive the optimal partition of channel coherence time.Experiment results verify the benefits of the proposed resource allocation schemes under three different scenarios and illustrate the tradeoff of resource allocation between three stages.展开更多
The cloud platform has limited defense resources to fully protect the edge servers used to process crowd sensing data in Internet of Things.To guarantee the network's overall security,we present a network defense ...The cloud platform has limited defense resources to fully protect the edge servers used to process crowd sensing data in Internet of Things.To guarantee the network's overall security,we present a network defense resource allocation with multi-armed bandits to maximize the network's overall benefit.Firstly,we propose the method for dynamic setting of node defense resource thresholds to obtain the defender(attacker)benefit function of edge servers(nodes)and distribution.Secondly,we design a defense resource sharing mechanism for neighboring nodes to obtain the defense capability of nodes.Subsequently,we use the decomposability and Lipschitz conti-nuity of the defender's total expected utility to reduce the difference between the utility's discrete and continuous arms and analyze the difference theoretically.Finally,experimental results show that the method maximizes the defender's total expected utility and reduces the difference between the discrete and continuous arms of the utility.展开更多
With the rapid development of Network Function Virtualization(NFV),the problem of low resource utilizationin traditional data centers is gradually being addressed.However,existing research does not optimize both local...With the rapid development of Network Function Virtualization(NFV),the problem of low resource utilizationin traditional data centers is gradually being addressed.However,existing research does not optimize both localand global allocation of resources in data centers.Hence,we propose an adaptive hybrid optimization strategy thatcombines dynamic programming and neural networks to improve resource utilization and service quality in datacenters.Our approach encompasses a service function chain simulation generator,a parallel architecture servicesystem,a dynamic programming strategy formaximizing the utilization of local server resources,a neural networkfor predicting the global utilization rate of resources and a global resource optimization strategy for bottleneck andredundant resources.With the implementation of our local and global resource allocation strategies,the systemperformance is significantly optimized through simulation.展开更多
Collaborative edge computing is a promising direction to handle the computation intensive tasks in B5G wireless networks.However,edge computing servers(ECSs)from different operators may not trust each other,and thus t...Collaborative edge computing is a promising direction to handle the computation intensive tasks in B5G wireless networks.However,edge computing servers(ECSs)from different operators may not trust each other,and thus the incentives for collaboration cannot be guaranteed.In this paper,we propose a consortium blockchain enabled collaborative edge computing framework,where users can offload computing tasks to ECSs from different operators.To minimize the total delay of users,we formulate a joint task offloading and resource optimization problem,under the constraint of the computing capability of each ECS.We apply the Tammer decomposition method and heuristic optimization algorithms to obtain the optimal solution.Finally,we propose a reputation based node selection approach to facilitate the consensus process,and also consider a completion time based primary node selection to avoid monopolization of certain edge node and enhance the security of the blockchain.Simulation results validate the effectiveness of the proposed algorithm,and the total delay can be reduced by up to 40%compared with the non-cooperative case.展开更多
Users and edge servers are not fullymutually trusted inmobile edge computing(MEC),and hence blockchain can be introduced to provide trustableMEC.In blockchain-basedMEC,each edge server functions as a node in bothMEC a...Users and edge servers are not fullymutually trusted inmobile edge computing(MEC),and hence blockchain can be introduced to provide trustableMEC.In blockchain-basedMEC,each edge server functions as a node in bothMEC and blockchain,processing users’tasks and then uploading the task related information to the blockchain.That is,each edge server runs both users’offloaded tasks and blockchain tasks simultaneously.Note that there is a trade-off between the resource allocation for MEC and blockchain tasks.Therefore,the allocation of the resources of edge servers to the blockchain and theMEC is crucial for the processing delay of blockchain-based MEC.Most of the existing research tackles the problem of resource allocation in either blockchain or MEC,which leads to unfavorable performance of the blockchain-based MEC system.In this paper,we study how to allocate the computing resources of edge servers to the MEC and blockchain tasks with the aimtominimize the total systemprocessing delay.For the problem,we propose a computing resource Allocation algorithmfor Blockchain-based MEC(ABM)which utilizes the Slater’s condition,Karush-Kuhn-Tucker(KKT)conditions,partial derivatives of the Lagrangian function and subgradient projection method to obtain the solution.Simulation results show that ABM converges and effectively reduces the processing delay of blockchain-based MEC.展开更多
Inter-datacenter elastic optical networks(EON)need to provide the service for the requests of cloud computing that require not only connectivity and computing resources but also network survivability.In this paper,to ...Inter-datacenter elastic optical networks(EON)need to provide the service for the requests of cloud computing that require not only connectivity and computing resources but also network survivability.In this paper,to realize joint allocation of computing and connectivity resources in survivable inter-datacenter EONs,a survivable routing,modulation level,spectrum,and computing resource allocation algorithm(SRMLSCRA)algorithm and three datacenter selection strategies,i.e.Computing Resource First(CRF),Shortest Path First(SPF)and Random Destination(RD),are proposed for different scenarios.Unicast and manycast are applied to the communication of computing requests,and the routing strategies are calculated respectively.Simulation results show that SRMLCRA-CRF can serve the largest amount of protected computing tasks,and the requested calculation blocking probability is reduced by 29.2%,28.3%and 30.5%compared with SRMLSCRA-SPF,SRMLSCRA-RD and the benchmark EPS-RMSA algorithms respectively.Therefore,it is more applicable to the networks with huge calculations.Besides,SRMLSCRA-SPF consumes the least spectrum,thereby exhibiting its suitability for scenarios where the amount of calculation is small and communication resources are scarce.The results demonstrate that the proposed methods realize the joint allocation of computing and connectivity resources,and could provide efficient protection for services under single-link failure and occupy less spectrum.展开更多
Unmanned Aerial Vehicles(UAvs)as aerial base stations to provide communication services for ground users is a flexible and cost-effective paradigm in B5G.Besides,dynamic resource allocation and multi-connectivity can ...Unmanned Aerial Vehicles(UAvs)as aerial base stations to provide communication services for ground users is a flexible and cost-effective paradigm in B5G.Besides,dynamic resource allocation and multi-connectivity can be adopted to further harness the potentials of UAVs in improving communication capacity,in such situations such that the interference among users becomes a pivotal disincentive requiring effective solutions.To this end,we investigate the Joint UAV-User Association,Channel Allocation,and transmission Power Control(J-UACAPC)problem in a multi-connectivity-enabled UAV network with constrained backhaul links,where each UAV can determine the reusable channels and transmission power to serve the selected ground users.The goal was to mitigate co-channel interference while maximizing long-term system utility.The problem was modeled as a cooperative stochastic game with hybrid discrete-continuous action space.A Multi-Agent Hybrid Deep Reinforcement Learning(MAHDRL)algorithm was proposed to address this problem.Extensive simulation results demonstrated the effectiveness of the proposed algorithm and showed that it has a higher system utility than the baseline methods.展开更多
Quantum key distribution(QKD)is a technology that can resist the threat of quantum computers to existing conventional cryptographic protocols.However,due to the stringent requirements of the quantum key generation env...Quantum key distribution(QKD)is a technology that can resist the threat of quantum computers to existing conventional cryptographic protocols.However,due to the stringent requirements of the quantum key generation environment,the generated quantum keys are considered valuable,and the slow key generation rate conflicts with the high-speed data transmission in traditional optical networks.In this paper,for the QKD network with a trusted relay,which is mainly based on point-to-point quantum keys and has complex changes in network resources,we aim to allocate resources reasonably for data packet distribution.Firstly,we formulate a linear programming constraint model for the key resource allocation(KRA)problem based on the time-slot scheduling.Secondly,we propose a new scheduling scheme based on the graded key security requirements(GKSR)and a new micro-log key storage algorithm for effective storage and management of key resources.Finally,we propose a key resource consumption(KRC)routing optimization algorithm to properly allocate time slots,routes,and key resources.Simulation results show that the proposed scheme significantly improves the key distribution success rate and key resource utilization rate,among others.展开更多
To meet the communication services with diverse requirements,dynamic resource allocation has shown increasing importance.In this paper,we consider the multi-slot and multi-user resource allocation(MSMU-RA)in a downlin...To meet the communication services with diverse requirements,dynamic resource allocation has shown increasing importance.In this paper,we consider the multi-slot and multi-user resource allocation(MSMU-RA)in a downlink cellular scenario with the aim of maximizing system spectral efficiency while guaranteeing user fairness.We first model the MSMURA problem as a dual-sequence decision-making process,and then solve it by a novel Transformerbased deep reinforcement learning(TDRL)approach.Specifically,the proposed TDRL approach can be achieved based on two aspects:1)To adapt to the dynamic wireless environment,the proximal policy optimization(PPO)algorithm is used to optimize the multi-slot RA strategy.2)To avoid co-channel interference,the Transformer-based PPO algorithm is presented to obtain the optimal multi-user RA scheme by exploring the mapping between user sequence and resource sequence.Experimental results show that:i)the proposed approach outperforms both the traditional and DRL methods in spectral efficiency and user fairness,ii)the proposed algorithm is superior to DRL approaches in terms of convergence speed and generalization performance.展开更多
In this paper,we propose the Two-way Deep Reinforcement Learning(DRL)-Based resource allocation algorithm,which solves the problem of resource allocation in the cognitive downlink network based on the underlay mode.Se...In this paper,we propose the Two-way Deep Reinforcement Learning(DRL)-Based resource allocation algorithm,which solves the problem of resource allocation in the cognitive downlink network based on the underlay mode.Secondary users(SUs)in the cognitive network are multiplexed by a new Power Domain Sparse Code Multiple Access(PD-SCMA)scheme,and the physical resources of the cognitive base station are virtualized into two types of slices:enhanced mobile broadband(eMBB)slice and ultrareliable low latency communication(URLLC)slice.We design the Double Deep Q Network(DDQN)network output the optimal codebook assignment scheme and simultaneously use the Deep Deterministic Policy Gradient(DDPG)network output the optimal power allocation scheme.The objective is to jointly optimize the spectral efficiency of the system and the Quality of Service(QoS)of SUs.Simulation results show that the proposed algorithm outperforms the CNDDQN algorithm and modified JEERA algorithm in terms of spectral efficiency and QoS satisfaction.Additionally,compared with the Power Domain Non-orthogonal Multiple Access(PD-NOMA)slices and the Sparse Code Multiple Access(SCMA)slices,the PD-SCMA slices can dramatically enhance spectral efficiency and increase the number of accessible users.展开更多
Mobile edge computing(MEC)-enabled satellite-terrestrial networks(STNs)can provide Internet of Things(IoT)devices with global computing services.Sometimes,the network state information is uncertain or unknown.To deal ...Mobile edge computing(MEC)-enabled satellite-terrestrial networks(STNs)can provide Internet of Things(IoT)devices with global computing services.Sometimes,the network state information is uncertain or unknown.To deal with this situation,we investigate online learning-based offloading decision and resource allocation in MEC-enabled STNs in this paper.The problem of minimizing the average sum task completion delay of all IoT devices over all time periods is formulated.We decompose this optimization problem into a task offloading decision problem and a computing resource allocation problem.A joint optimization scheme of offloading decision and resource allocation is then proposed,which consists of a task offloading decision algorithm based on the devices cooperation aided upper confidence bound(UCB)algorithm and a computing resource allocation algorithm based on the Lagrange multiplier method.Simulation results validate that the proposed scheme performs better than other baseline schemes.展开更多
With the rapid development of urban rail transit,the existing track detection has some problems such as low efficiency and insufficient detection coverage,so an intelligent and automatic track detectionmethod based on...With the rapid development of urban rail transit,the existing track detection has some problems such as low efficiency and insufficient detection coverage,so an intelligent and automatic track detectionmethod based onUAV is urgently needed to avoid major safety accidents.At the same time,the geographical distribution of IoT devices results in the inefficient use of the significant computing potential held by a large number of devices.As a result,the Dispersed Computing(DCOMP)architecture enables collaborative computing between devices in the Internet of Everything(IoE),promotes low-latency and efficient cross-wide applications,and meets users’growing needs for computing performance and service quality.This paper focuses on examining the resource allocation challenge within a dispersed computing environment that utilizes UAV inspection tracks.Furthermore,the system takes into account both resource constraints and computational constraints and transforms the optimization problem into an energy minimization problem with computational constraints.The Markov Decision Process(MDP)model is employed to capture the connection between the dispersed computing resource allocation strategy and the system environment.Subsequently,a method based on Double Deep Q-Network(DDQN)is introduced to derive the optimal policy.Simultaneously,an experience replay mechanism is implemented to tackle the issue of increasing dimensionality.The experimental simulations validate the efficacy of the method across various scenarios.展开更多
In this paper,we investigate IRS-aided user cooperation(UC)scheme in millimeter wave(mmWave)wirelesspowered sensor networks(WPSN),where two single-antenna users are wireless powered in the wireless energy transfer(WET...In this paper,we investigate IRS-aided user cooperation(UC)scheme in millimeter wave(mmWave)wirelesspowered sensor networks(WPSN),where two single-antenna users are wireless powered in the wireless energy transfer(WET)phase first and then cooperatively transmit information to a hybrid access point(AP)in the wireless information transmission(WIT)phase,following which the IRS is deployed to enhance the system performance of theWET andWIT.We maximized the weighted sum-rate problem by jointly optimizing the transmit time slots,power allocations,and the phase shifts of the IRS.Due to the non-convexity of the original problem,a semidefinite programming relaxation-based approach is proposed to convert the formulated problem to a convex optimization framework,which can obtain the optimal global solution.Simulation results demonstrate that the weighted sum throughput of the proposed UC scheme outperforms the non-UC scheme whether equipped with IRS or not.展开更多
The current resource allocation in 5G vehicular networks for mobile cloud communication faces several challenges,such as low user utilization,unbalanced resource allocation,and extended adaptive allocation time.We pro...The current resource allocation in 5G vehicular networks for mobile cloud communication faces several challenges,such as low user utilization,unbalanced resource allocation,and extended adaptive allocation time.We propose an adaptive allocation algorithm for mobile cloud communication resources in 5G vehicular networks to address these issues.This study analyzes the components of the 5G vehicular network architecture to determine the performance of different components.It is ascertained that the communication modes in 5G vehicular networks for mobile cloud communication include in-band and out-of-band modes.Furthermore,this study analyzes the single-hop and multi-hop modes in mobile cloud communication and calculates the resource transmission rate and bandwidth in different communication modes.The study also determines the scenario of one-way and two-way vehicle lane cloud communication network connectivity,calculates the probability of vehicle network connectivity under different mobile cloud communication radii,and determines the amount of cloud communication resources required by vehicles in different lane scenarios.Based on the communication status of users in 5G vehicular networks,this study calculates the bandwidth and transmission rate of the allocated channels using Shannon’s formula.It determines the adaptive allocation of cloud communication resources,introduces an objective function to obtain the optimal solution after allocation,and completes the adaptive allocation process.The experimental results demonstrate that,with the application of the proposed method,the maximum utilization of user communication resources reaches approximately 99%.The balance coefficient curve approaches 1,and the allocation time remains under 2 s.This indicates that the proposed method has higher adaptive allocation efficiency.展开更多
基金supported in part by the National Natural Science Foundation of China under Grant no.61473066 and Grant no.61601109in part by the Fundamental Research Funds for the Central Universities under Grant No.N152305001.
文摘In this paper,a new communication model is built named grouping D2D(GD2D).Different from the traditional D2D coordination,we proposed GD2D communication in licensed and unlicensed spectrum simultaneously.We formulate a resource allocation problem,which aims at maximizing the energy efficiency(EE)of the system while guaranteeing the quality-of-service(Qos)of users.To efficiently solve this problem,the non-convex optimization problem is first transformed into a convex optimization problem.By transforming the fractional-form problem into an equivalent subtractive-form problem,an iterative power allocation algorithm is proposed to maximize the system EE.Moreover,the optimal closedform power allocation expressions are derived by the Lagrangian approach.Simulation results show that our algorithm achieves higher EE performance than the traditional D2D communication scheme.
基金supported in part by the National Natural Science Foundation of China under Grant 62001056, 61925101, U21A20444in part by the Fundamental Research Funds for the Central Universities under Grant 500421336 and Grant 505021163。
文摘With the rapid increasing of maritime activities, maritime wireless networks(MWNs) with high reliability, high energy efficiency, and low delay are required. However, the centralized networking with fixed resource scheduling is not suitable for MWNs due to the special environment. In this paper,we introduce the collaborative relay communication in distributed MWNs to improve the link reliability, and propose an orthogonal time-frequency resource block reservation based multiple access(RRMA) scheme for both one-hop direct link and two-hop collaborative relay link to reduce the interference. To further improve the network performance, we formulate an energy efficiency(EE) maximization resource allocation problem and solve it by an iterative algorithm based on the Dinkelbach method. Finally, numerical results are provided to investigate the proposed RRMA scheme and resource allocation algorithm, showing that the low outage probability and transmission delay can be attained by the proposed RRMA scheme. Moreover,the proposed resource allocation algorithm is capable of achieving high EE in distributed MWNs.
基金partially supported by National Basic Research Program of China (2013CB329002)National Natural Science Foundation of China (61631013)+6 种基金The National High Technology Research and Development Program of China(2014AA01A703)Science Fund for Creative Research Groups of NSFC (61321061)National Major Project (2017ZX03001011)International Science and Technology Cooperation Program (2014DFT10320)National Science Foundation of China (61701457 \& 61771286)Tsinghua-Qualcomm Joint Research ProgramHuawei Innovation Research Program
文摘The small-cell technology is promising for spectral-efficiency enhancement. However, it usually requires a huge amount of energy consumption. In this paper, queue state information and channel state information are jointly utilized to minimize the time average of overall energy consumption for a multi-carrier small-cell network, where the inter-cell interference is an intractable problem. Based on the Lyapunov optimization theory, the problem could be solved by dynamically optimizing the problem of user assignment, carrier allocation and power allocation in each time slot. As the optimization problem is NP-hard, we propose a heuristic iteration algorithm to solve it. Numerical results verify that the heuristic algorithm offers an approximate performance as the brute-force algorithm. Moreover, it could bring down the overall energy consumption to different degrees according to the variation of traffic load. Meanwhile, it could achieve the same sum rate as the algorithm which focuses on maximizing system sum rate.
基金supported in part by National Natural Science Foundation of China (No.61372070)Natural Science Basic Research Plan in Shaanxi Province of China (2015JM6324)+2 种基金Ningbo Natural Science Foundation (2015A610117)Hong Kong,Macao and Taiwan Science & Technology Cooperation Program of China (2015DFT10160)the 111 Project (B08038)
文摘In order to maximize system energy efficiency(EE) under user quality of service(Qo S) restraints in Long Term Evolution-Advanced(LTE-A) networks,a constrained joint resource optimization allocation scheme is presented,which is NP-hard. Hence,we divide it into three sub-problems to reduce computation complexity,i.e.,the resource block(RB) allocation,the power distribution,and the modulation and coding scheme(MCS) assignment for user codewords. Then an enhanced heuristic approach GAPSO is proposed and is adopted in the RB and power allocation respectively to reduce computational complexity further on. Moreover,a novel MCS allocation scheme is put forward,which could make a good balance between the system reliability and availability under different channel conditions. Simulation results show that the proposed GAPSO could achieve better performance in convergence speed and global optimum searching,and that the joint resource allocation scheme could improve energy efficiency effectively under user Qo S requirements.
基金supported by the major science and technology projects in Anhui Province(202003a05020009)the innovation foundation of the city of Bengbu(JZ2022YDZJ0019)the Key Technology Research and Development Project of Hefei(2021GJ029).
文摘The vehicular edge computing(VEC)is a new paradigm that allows vehicles to offload computational tasks to base stations(BSs)with edge servers for computing.In general,the VEC paradigm uses the 5G for wireless communications,where the massive multi-input multi-output(MIMO)technique will be used.However,considering in the VEC environment with many vehicles,the energy consumption of BS may be very large.In this paper,we study the energy optimization problem for the massive MIMO-based VEC network.Aiming at reducing the relevant BS energy consumption,we first propose a joint optimization problem of computation resource allocation,beam allocation and vehicle grouping scheme.Since the original problem is hard to be solved directly,we try to split the original problem into two subproblems and then design a heuristic algorithm to solve them.Simulation results show that our proposed algorithm efficiently reduces the BS energy consumption compared to other schemes.
基金supported in part by the National Natural Science Foundation of China under Grant No.61701197in part by the National Key Research and Development Program of China under Grant No.2021YFA1000500(4)in part by the 111 Project under Grant No.B23008.
文摘Federated Edge Learning(FEL),an emerging distributed Machine Learning(ML)paradigm,enables model training in a distributed environment while ensuring user privacy by using physical separation for each user’s data.However,with the development of complex application scenarios such as the Internet of Things(IoT)and Smart Earth,the conventional resource allocation schemes can no longer effectively support these growing computational and communication demands.Therefore,joint resource optimization may be the key solution to the scaling problem.This paper simultaneously addresses the multifaceted challenges of computation and communication,with the growing multiple resource demands.We systematically review the joint allocation strategies for different resources(computation,data,communication,and network topology)in FEL,and summarize the advantages in improving system efficiency,reducing latency,enhancing resource utilization,and enhancing robustness.In addition,we present the potential ability of joint optimization to enhance privacy preservation by reducing communication requirements,indirectly.This work not only provides theoretical support for resource management in federated learning(FL)systems,but also provides ideas for potential optimal deployment in multiple real-world scenarios.By thoroughly discussing the current challenges and future research directions,it also provides some important insights into multi-resource optimization in complex application environments.
基金supported by the foundation of National Key Laboratory of Electromagnetic Environment(Grant No.JCKY2020210C 614240304)Natural Science Foundation of ZheJiang province(LQY20F010001)+1 种基金the National Natural Science Foundation of China under grant numbers 82004499State Key Laboratory of Millimeter Waves under grant numbers K202012.
文摘The performance of massive MIMO systems relies heavily on the availability of Channel State Information at the Transmitter(CSIT).A large amount of work has been devoted to reducing the CSIT acquisition overhead at the pilot training and/or CsI feedback stage.In fact,the downlink communication generally includes three stages,i.e.,pilot training,CsI feedback,and data transmission.These three stages are mutually related and jointly determine the overall system performance.Unfortunately,there exist few studies on the reduction of csIT acquisition overhead from the global point of view.In this paper,we integrate the Minimum Mean Square Error(MMSE)channel estimation,Random Vector Quantization(RVQ)based limited feedback and Maximal Ratio Combining(MRC)precoding into a unified framework for investigating the resource allocation problem.In particular,we first approximate the covariance matrix of the quantization error with a simple expression and derive an analytical expression of the received Signal-to-Noise Ratio(SNR)based on the deterministic equivalence theory.Then the three performance metrics(the spectral efficiency,energy efficiency,and total energy consumption)oriented problems are formulated analytically.With practical system requirements,these three metrics can be collaboratively optimized.Finally,we propose an optimization solver to derive the optimal partition of channel coherence time.Experiment results verify the benefits of the proposed resource allocation schemes under three different scenarios and illustrate the tradeoff of resource allocation between three stages.
基金supported by the National Natural Science Foundation of China(NSFC)[grant numbers 62172377,61872205]the Shandong Provincial Natural Science Foundation[grant number ZR2019MF018]the Startup Research Foundation for Distinguished Scholars No.202112016.
文摘The cloud platform has limited defense resources to fully protect the edge servers used to process crowd sensing data in Internet of Things.To guarantee the network's overall security,we present a network defense resource allocation with multi-armed bandits to maximize the network's overall benefit.Firstly,we propose the method for dynamic setting of node defense resource thresholds to obtain the defender(attacker)benefit function of edge servers(nodes)and distribution.Secondly,we design a defense resource sharing mechanism for neighboring nodes to obtain the defense capability of nodes.Subsequently,we use the decomposability and Lipschitz conti-nuity of the defender's total expected utility to reduce the difference between the utility's discrete and continuous arms and analyze the difference theoretically.Finally,experimental results show that the method maximizes the defender's total expected utility and reduces the difference between the discrete and continuous arms of the utility.
基金the Fundamental Research Program of Guangdong,China,under Grants 2020B1515310023 and 2023A1515011281in part by the National Natural Science Foundation of China under Grant 61571005.
文摘With the rapid development of Network Function Virtualization(NFV),the problem of low resource utilizationin traditional data centers is gradually being addressed.However,existing research does not optimize both localand global allocation of resources in data centers.Hence,we propose an adaptive hybrid optimization strategy thatcombines dynamic programming and neural networks to improve resource utilization and service quality in datacenters.Our approach encompasses a service function chain simulation generator,a parallel architecture servicesystem,a dynamic programming strategy formaximizing the utilization of local server resources,a neural networkfor predicting the global utilization rate of resources and a global resource optimization strategy for bottleneck andredundant resources.With the implementation of our local and global resource allocation strategies,the systemperformance is significantly optimized through simulation.
基金supported in part by the National Key R&D Program of China under Grant 2020YFB1005900the National Natural Science Foundation of China under Grant 62001220+3 种基金the Jiangsu Provincial Key Research and Development Program under Grants BE2022068the Natural Science Foundation of Jiangsu Province under Grants BK20200440the Future Network Scientific Research Fund Project FNSRFP-2021-YB-03the Young Elite Scientist Sponsorship Program,China Association for Science and Technology.
文摘Collaborative edge computing is a promising direction to handle the computation intensive tasks in B5G wireless networks.However,edge computing servers(ECSs)from different operators may not trust each other,and thus the incentives for collaboration cannot be guaranteed.In this paper,we propose a consortium blockchain enabled collaborative edge computing framework,where users can offload computing tasks to ECSs from different operators.To minimize the total delay of users,we formulate a joint task offloading and resource optimization problem,under the constraint of the computing capability of each ECS.We apply the Tammer decomposition method and heuristic optimization algorithms to obtain the optimal solution.Finally,we propose a reputation based node selection approach to facilitate the consensus process,and also consider a completion time based primary node selection to avoid monopolization of certain edge node and enhance the security of the blockchain.Simulation results validate the effectiveness of the proposed algorithm,and the total delay can be reduced by up to 40%compared with the non-cooperative case.
基金supported by the Key Research and Development Project in Anhui Province of China(Grant No.202304a05020059)the Fundamental Research Funds for the Central Universities of China(Grant No.PA2023GDSK0055)the Project of Anhui Province Economic and Information Bureau(Grant No.JB20099).
文摘Users and edge servers are not fullymutually trusted inmobile edge computing(MEC),and hence blockchain can be introduced to provide trustableMEC.In blockchain-basedMEC,each edge server functions as a node in bothMEC and blockchain,processing users’tasks and then uploading the task related information to the blockchain.That is,each edge server runs both users’offloaded tasks and blockchain tasks simultaneously.Note that there is a trade-off between the resource allocation for MEC and blockchain tasks.Therefore,the allocation of the resources of edge servers to the blockchain and theMEC is crucial for the processing delay of blockchain-based MEC.Most of the existing research tackles the problem of resource allocation in either blockchain or MEC,which leads to unfavorable performance of the blockchain-based MEC system.In this paper,we study how to allocate the computing resources of edge servers to the MEC and blockchain tasks with the aimtominimize the total systemprocessing delay.For the problem,we propose a computing resource Allocation algorithmfor Blockchain-based MEC(ABM)which utilizes the Slater’s condition,Karush-Kuhn-Tucker(KKT)conditions,partial derivatives of the Lagrangian function and subgradient projection method to obtain the solution.Simulation results show that ABM converges and effectively reduces the processing delay of blockchain-based MEC.
基金supported by the National Natural Science Foundation of China(No.62001045)Beijing Municipal Natural Science Foundation(No.4214059)+1 种基金Fund of State Key Laboratory of IPOC(BUPT)(No.IPOC2021ZT17)Fundamental Research Funds for the Central Universities(No.2022RC09).
文摘Inter-datacenter elastic optical networks(EON)need to provide the service for the requests of cloud computing that require not only connectivity and computing resources but also network survivability.In this paper,to realize joint allocation of computing and connectivity resources in survivable inter-datacenter EONs,a survivable routing,modulation level,spectrum,and computing resource allocation algorithm(SRMLSCRA)algorithm and three datacenter selection strategies,i.e.Computing Resource First(CRF),Shortest Path First(SPF)and Random Destination(RD),are proposed for different scenarios.Unicast and manycast are applied to the communication of computing requests,and the routing strategies are calculated respectively.Simulation results show that SRMLCRA-CRF can serve the largest amount of protected computing tasks,and the requested calculation blocking probability is reduced by 29.2%,28.3%and 30.5%compared with SRMLSCRA-SPF,SRMLSCRA-RD and the benchmark EPS-RMSA algorithms respectively.Therefore,it is more applicable to the networks with huge calculations.Besides,SRMLSCRA-SPF consumes the least spectrum,thereby exhibiting its suitability for scenarios where the amount of calculation is small and communication resources are scarce.The results demonstrate that the proposed methods realize the joint allocation of computing and connectivity resources,and could provide efficient protection for services under single-link failure and occupy less spectrum.
基金supported in part by the National Natural Science Foundation of China(grant nos.61971365,61871339,62171392)Digital Fujian Province Key Laboratory of IoT Communication,Architecture and Safety Technology(grant no.2010499)+1 种基金the State Key Program of the National Natural Science Foundation of China(grant no.61731012)the Natural Science Foundation of Fujian Province of China No.2021J01004.
文摘Unmanned Aerial Vehicles(UAvs)as aerial base stations to provide communication services for ground users is a flexible and cost-effective paradigm in B5G.Besides,dynamic resource allocation and multi-connectivity can be adopted to further harness the potentials of UAVs in improving communication capacity,in such situations such that the interference among users becomes a pivotal disincentive requiring effective solutions.To this end,we investigate the Joint UAV-User Association,Channel Allocation,and transmission Power Control(J-UACAPC)problem in a multi-connectivity-enabled UAV network with constrained backhaul links,where each UAV can determine the reusable channels and transmission power to serve the selected ground users.The goal was to mitigate co-channel interference while maximizing long-term system utility.The problem was modeled as a cooperative stochastic game with hybrid discrete-continuous action space.A Multi-Agent Hybrid Deep Reinforcement Learning(MAHDRL)algorithm was proposed to address this problem.Extensive simulation results demonstrated the effectiveness of the proposed algorithm and showed that it has a higher system utility than the baseline methods.
基金Project supported by the Natural Science Foundation of Jilin Province of China(Grant No.20210101417JC).
文摘Quantum key distribution(QKD)is a technology that can resist the threat of quantum computers to existing conventional cryptographic protocols.However,due to the stringent requirements of the quantum key generation environment,the generated quantum keys are considered valuable,and the slow key generation rate conflicts with the high-speed data transmission in traditional optical networks.In this paper,for the QKD network with a trusted relay,which is mainly based on point-to-point quantum keys and has complex changes in network resources,we aim to allocate resources reasonably for data packet distribution.Firstly,we formulate a linear programming constraint model for the key resource allocation(KRA)problem based on the time-slot scheduling.Secondly,we propose a new scheduling scheme based on the graded key security requirements(GKSR)and a new micro-log key storage algorithm for effective storage and management of key resources.Finally,we propose a key resource consumption(KRC)routing optimization algorithm to properly allocate time slots,routes,and key resources.Simulation results show that the proposed scheme significantly improves the key distribution success rate and key resource utilization rate,among others.
基金supported by the National Natural Science Foundation of China(No.62071354)the Key Research and Development Program of Shaanxi(No.2022ZDLGY05-08)supported by the ISN State Key Laboratory。
文摘To meet the communication services with diverse requirements,dynamic resource allocation has shown increasing importance.In this paper,we consider the multi-slot and multi-user resource allocation(MSMU-RA)in a downlink cellular scenario with the aim of maximizing system spectral efficiency while guaranteeing user fairness.We first model the MSMURA problem as a dual-sequence decision-making process,and then solve it by a novel Transformerbased deep reinforcement learning(TDRL)approach.Specifically,the proposed TDRL approach can be achieved based on two aspects:1)To adapt to the dynamic wireless environment,the proximal policy optimization(PPO)algorithm is used to optimize the multi-slot RA strategy.2)To avoid co-channel interference,the Transformer-based PPO algorithm is presented to obtain the optimal multi-user RA scheme by exploring the mapping between user sequence and resource sequence.Experimental results show that:i)the proposed approach outperforms both the traditional and DRL methods in spectral efficiency and user fairness,ii)the proposed algorithm is superior to DRL approaches in terms of convergence speed and generalization performance.
基金supported by the National Natural Science Foundation of China(Grant No.61971057).
文摘In this paper,we propose the Two-way Deep Reinforcement Learning(DRL)-Based resource allocation algorithm,which solves the problem of resource allocation in the cognitive downlink network based on the underlay mode.Secondary users(SUs)in the cognitive network are multiplexed by a new Power Domain Sparse Code Multiple Access(PD-SCMA)scheme,and the physical resources of the cognitive base station are virtualized into two types of slices:enhanced mobile broadband(eMBB)slice and ultrareliable low latency communication(URLLC)slice.We design the Double Deep Q Network(DDQN)network output the optimal codebook assignment scheme and simultaneously use the Deep Deterministic Policy Gradient(DDPG)network output the optimal power allocation scheme.The objective is to jointly optimize the spectral efficiency of the system and the Quality of Service(QoS)of SUs.Simulation results show that the proposed algorithm outperforms the CNDDQN algorithm and modified JEERA algorithm in terms of spectral efficiency and QoS satisfaction.Additionally,compared with the Power Domain Non-orthogonal Multiple Access(PD-NOMA)slices and the Sparse Code Multiple Access(SCMA)slices,the PD-SCMA slices can dramatically enhance spectral efficiency and increase the number of accessible users.
基金supported by National Key Research and Development Program of China(2018YFC1504502).
文摘Mobile edge computing(MEC)-enabled satellite-terrestrial networks(STNs)can provide Internet of Things(IoT)devices with global computing services.Sometimes,the network state information is uncertain or unknown.To deal with this situation,we investigate online learning-based offloading decision and resource allocation in MEC-enabled STNs in this paper.The problem of minimizing the average sum task completion delay of all IoT devices over all time periods is formulated.We decompose this optimization problem into a task offloading decision problem and a computing resource allocation problem.A joint optimization scheme of offloading decision and resource allocation is then proposed,which consists of a task offloading decision algorithm based on the devices cooperation aided upper confidence bound(UCB)algorithm and a computing resource allocation algorithm based on the Lagrange multiplier method.Simulation results validate that the proposed scheme performs better than other baseline schemes.
文摘With the rapid development of urban rail transit,the existing track detection has some problems such as low efficiency and insufficient detection coverage,so an intelligent and automatic track detectionmethod based onUAV is urgently needed to avoid major safety accidents.At the same time,the geographical distribution of IoT devices results in the inefficient use of the significant computing potential held by a large number of devices.As a result,the Dispersed Computing(DCOMP)architecture enables collaborative computing between devices in the Internet of Everything(IoE),promotes low-latency and efficient cross-wide applications,and meets users’growing needs for computing performance and service quality.This paper focuses on examining the resource allocation challenge within a dispersed computing environment that utilizes UAV inspection tracks.Furthermore,the system takes into account both resource constraints and computational constraints and transforms the optimization problem into an energy minimization problem with computational constraints.The Markov Decision Process(MDP)model is employed to capture the connection between the dispersed computing resource allocation strategy and the system environment.Subsequently,a method based on Double Deep Q-Network(DDQN)is introduced to derive the optimal policy.Simultaneously,an experience replay mechanism is implemented to tackle the issue of increasing dimensionality.The experimental simulations validate the efficacy of the method across various scenarios.
基金This work was supported in part by the open research fund of National Mobile Communications Research Laboratory,Southeast University(No.2023D11)in part by Sponsored by program for Science&Technology Innovation Talents in Universities of Henan Province(23HASTIT019)+2 种基金in part by Natural Science Foundation of Henan Province(20232300421097)in part by the project funded by China Postdoctoral Science Foundation(2020M682345)in part by the Henan Postdoctoral Foundation(202001015).
文摘In this paper,we investigate IRS-aided user cooperation(UC)scheme in millimeter wave(mmWave)wirelesspowered sensor networks(WPSN),where two single-antenna users are wireless powered in the wireless energy transfer(WET)phase first and then cooperatively transmit information to a hybrid access point(AP)in the wireless information transmission(WIT)phase,following which the IRS is deployed to enhance the system performance of theWET andWIT.We maximized the weighted sum-rate problem by jointly optimizing the transmit time slots,power allocations,and the phase shifts of the IRS.Due to the non-convexity of the original problem,a semidefinite programming relaxation-based approach is proposed to convert the formulated problem to a convex optimization framework,which can obtain the optimal global solution.Simulation results demonstrate that the weighted sum throughput of the proposed UC scheme outperforms the non-UC scheme whether equipped with IRS or not.
基金This research was supported by Science and Technology Research Project of Education Department of Jiangxi Province,China(Nos.GJJ2206701,GJJ2206717).
文摘The current resource allocation in 5G vehicular networks for mobile cloud communication faces several challenges,such as low user utilization,unbalanced resource allocation,and extended adaptive allocation time.We propose an adaptive allocation algorithm for mobile cloud communication resources in 5G vehicular networks to address these issues.This study analyzes the components of the 5G vehicular network architecture to determine the performance of different components.It is ascertained that the communication modes in 5G vehicular networks for mobile cloud communication include in-band and out-of-band modes.Furthermore,this study analyzes the single-hop and multi-hop modes in mobile cloud communication and calculates the resource transmission rate and bandwidth in different communication modes.The study also determines the scenario of one-way and two-way vehicle lane cloud communication network connectivity,calculates the probability of vehicle network connectivity under different mobile cloud communication radii,and determines the amount of cloud communication resources required by vehicles in different lane scenarios.Based on the communication status of users in 5G vehicular networks,this study calculates the bandwidth and transmission rate of the allocated channels using Shannon’s formula.It determines the adaptive allocation of cloud communication resources,introduces an objective function to obtain the optimal solution after allocation,and completes the adaptive allocation process.The experimental results demonstrate that,with the application of the proposed method,the maximum utilization of user communication resources reaches approximately 99%.The balance coefficient curve approaches 1,and the allocation time remains under 2 s.This indicates that the proposed method has higher adaptive allocation efficiency.