Brain tissue is one of the softest parts of the human body,composed of white matter and grey matter.The mechanical behavior of the brain tissue plays an essential role in regulating brain morphology and brain function...Brain tissue is one of the softest parts of the human body,composed of white matter and grey matter.The mechanical behavior of the brain tissue plays an essential role in regulating brain morphology and brain function.Besides,traumatic brain injury(TBI)and various brain diseases are also greatly influenced by the brain's mechanical properties.Whether white matter or grey matter,brain tissue contains multiscale structures composed of neurons,glial cells,fibers,blood vessels,etc.,each with different mechanical properties.As such,brain tissue exhibits complex mechanical behavior,usually with strong nonlinearity,heterogeneity,and directional dependence.Building a constitutive law for multiscale brain tissue using traditional function-based approaches can be very challenging.Instead,this paper proposes a data-driven approach to establish the desired mechanical model of brain tissue.We focus on blood vessels with internal pressure embedded in a white or grey matter matrix material to demonstrate our approach.The matrix is described by an isotropic or anisotropic nonlinear elastic model.A representative unit cell(RUC)with blood vessels is built,which is used to generate the stress-strain data under different internal blood pressure and various proportional displacement loading paths.The generated stress-strain data is then used to train a mechanical law using artificial neural networks to predict the macroscopic mechanical response of brain tissue under different internal pressures.Finally,the trained material model is implemented into finite element software to predict the mechanical behavior of a whole brain under intracranial pressure and distributed body forces.Compared with a direct numerical simulation that employs a reference material model,our proposed approach greatly reduces the computational cost and improves modeling efficiency.The predictions made by our trained model demonstrate sufficient accuracy.Specifically,we find that the level of internal blood pressure can greatly influence stress distribution and determine the possible related damage behaviors.展开更多
Through theoretical analysis,we construct a physical model that includes the influence of counter-external driven current opposite to the plasma current direction in the neoclassical tearing mode(NTM).The equation is ...Through theoretical analysis,we construct a physical model that includes the influence of counter-external driven current opposite to the plasma current direction in the neoclassical tearing mode(NTM).The equation is used with this model to obtain the modified Rutherford equation with co-current and counter-current contributions.Consistent with the reported experimental results,numerical simulations have shown that the localized counter external current can only partially suppress NTM when it is far from the resonant magnetic surface.Under some circumstances,the Ohkawa mechanism dominated current drive(OKCD)by electron cyclotron waves can concurrently create both co-current and counter-current.In this instance,the minimal electron cyclotron wave power that suppresses a particular NTM was calculated by the Rutherford equation.The result is marginally less than when taking co-current alone into consideration.As a result,to suppress NTM using OKCD,one only needs to align the co-current with a greater OKCD peak well with the resonant magnetic surface.The effect of its lower counter-current does not need to be considered because the location of the counter-current deviates greatly from the resonant magnetic surface.展开更多
Collaborative edge computing is a promising direction to handle the computation intensive tasks in B5G wireless networks.However,edge computing servers(ECSs)from different operators may not trust each other,and thus t...Collaborative edge computing is a promising direction to handle the computation intensive tasks in B5G wireless networks.However,edge computing servers(ECSs)from different operators may not trust each other,and thus the incentives for collaboration cannot be guaranteed.In this paper,we propose a consortium blockchain enabled collaborative edge computing framework,where users can offload computing tasks to ECSs from different operators.To minimize the total delay of users,we formulate a joint task offloading and resource optimization problem,under the constraint of the computing capability of each ECS.We apply the Tammer decomposition method and heuristic optimization algorithms to obtain the optimal solution.Finally,we propose a reputation based node selection approach to facilitate the consensus process,and also consider a completion time based primary node selection to avoid monopolization of certain edge node and enhance the security of the blockchain.Simulation results validate the effectiveness of the proposed algorithm,and the total delay can be reduced by up to 40%compared with the non-cooperative case.展开更多
In a network environment composed of different types of computing centers that can be divided into different layers(clod,edge layer,and others),the interconnection between them offers the possibility of peer-to-peer t...In a network environment composed of different types of computing centers that can be divided into different layers(clod,edge layer,and others),the interconnection between them offers the possibility of peer-to-peer task offloading.For many resource-constrained devices,the computation of many types of tasks is not feasible because they cannot support such computations as they do not have enough available memory and processing capacity.In this scenario,it is worth considering transferring these tasks to resource-rich platforms,such as Edge Data Centers or remote cloud servers.For different reasons,it is more exciting and appropriate to download various tasks to specific download destinations depending on the properties and state of the environment and the nature of the functions.At the same time,establishing an optimal offloading policy,which ensures that all tasks are executed within the required latency and avoids excessive workload on specific computing centers is not easy.This study presents two alternatives to solve the offloading decision paradigm by introducing two well-known algorithms,Graph Neural Networks(GNN)and Deep Q-Network(DQN).It applies the alternatives on a well-known Edge Computing simulator called PureEdgeSimand compares them with the two defaultmethods,Trade-Off and Round Robin.Experiments showed that variants offer a slight improvement in task success rate and workload distribution.In terms of energy efficiency,they provided similar results.Finally,the success rates of different computing centers are tested,and the lack of capacity of remote cloud servers to respond to applications in real-time is demonstrated.These novel ways of finding a download strategy in a local networking environment are unique as they emulate the state and structure of the environment innovatively,considering the quality of its connections and constant updates.The download score defined in this research is a crucial feature for determining the quality of a download path in the GNN training process and has not previously been proposed.Simultaneously,the suitability of Reinforcement Learning(RL)techniques is demonstrated due to the dynamism of the network environment,considering all the key factors that affect the decision to offload a given task,including the actual state of all devices.展开更多
The Multi-access Edge Cloud(MEC) networks extend cloud computing services and capabilities to the edge of the networks. By bringing computation and storage capabilities closer to end-users and connected devices, MEC n...The Multi-access Edge Cloud(MEC) networks extend cloud computing services and capabilities to the edge of the networks. By bringing computation and storage capabilities closer to end-users and connected devices, MEC networks can support a wide range of applications. MEC networks can also leverage various types of resources, including computation resources, network resources, radio resources,and location-based resources, to provide multidimensional resources for intelligent applications in 5/6G.However, tasks generated by users often consist of multiple subtasks that require different types of resources. It is a challenging problem to offload multiresource task requests to the edge cloud aiming at maximizing benefits due to the heterogeneity of resources provided by devices. To address this issue,we mathematically model the task requests with multiple subtasks. Then, the problem of task offloading of multi-resource task requests is proved to be NP-hard. Furthermore, we propose a novel Dual-Agent Deep Reinforcement Learning algorithm with Node First and Link features(NF_L_DA_DRL) based on the policy network, to optimize the benefits generated by offloading multi-resource task requests in MEC networks. Finally, simulation results show that the proposed algorithm can effectively improve the benefit of task offloading with higher resource utilization compared with baseline algorithms.展开更多
The Internet of Medical Things(Io MT) is regarded as a critical technology for intelligent healthcare in the foreseeable 6G era. Nevertheless, due to the limited computing power capability of edge devices and task-rel...The Internet of Medical Things(Io MT) is regarded as a critical technology for intelligent healthcare in the foreseeable 6G era. Nevertheless, due to the limited computing power capability of edge devices and task-related coupling relationships, Io MT faces unprecedented challenges. Considering the associative connections among tasks, this paper proposes a computing offloading policy for multiple-user devices(UDs) considering device-to-device(D2D) communication and a multi-access edge computing(MEC)technique under the scenario of Io MT. Specifically,to minimize the total delay and energy consumption concerning the requirement of Io MT, we first analyze and model the detailed local execution, MEC execution, D2D execution, and associated tasks offloading exchange model. Consequently, the associated tasks’ offloading scheme of multi-UDs is formulated as a mixed-integer nonconvex optimization problem. Considering the advantages of deep reinforcement learning(DRL) in processing tasks related to coupling relationships, a Double DQN based associative tasks computing offloading(DDATO) algorithm is then proposed to obtain the optimal solution, which can make the best offloading decision under the condition that tasks of UDs are associative. Furthermore, to reduce the complexity of the DDATO algorithm, the cacheaided procedure is intentionally introduced before the data training process. This avoids redundant offloading and computing procedures concerning tasks that previously have already been cached by other UDs. In addition, we use a dynamic ε-greedy strategy in the action selection section of the algorithm, thus preventing the algorithm from falling into a locally optimal solution. Simulation results demonstrate that compared with other existing methods for associative task models concerning different structures in the Io MT network, the proposed algorithm can lower the total cost more effectively and efficiently while also providing a tradeoff between delay and energy consumption tolerance.展开更多
The widespread use of advanced electronic devices has led to the emergence of spatial crowdsourcing,a method that taps into collective efforts to perform real-world tasks like environmental monitoring and traffic surv...The widespread use of advanced electronic devices has led to the emergence of spatial crowdsourcing,a method that taps into collective efforts to perform real-world tasks like environmental monitoring and traffic surveillance.Our research focuses on a specific type of spatial crowdsourcing that involves ongoing,collaborative efforts for continuous spatial data acquisition.However,due to limited budgets and workforce availability,the collected data often lacks completeness,posing a data deficiency problem.To address this,we propose a reciprocal framework to optimize task assignments by leveraging the mutual benefits of spatiotemporal subtask execution.We introduce an entropy-based quality metric to capture the combined effects of incomplete data acquisition and interpolation imprecision.Building on this,we explore a quality-aware task assignment method,corresponding to spatiotemporal assignment strategies.Since the assignment problem is NP-hard,we develop a polynomial-time algorithm with the guaranteed approximation ratio.Novel indexing and pruning techniques are proposed to further enhance performance.Extensive experiments conducted on datasets validate the effectiveness of our methods.展开更多
Thoracic diseases pose significant risks to an individual's chest health and are among the most perilous medical diseases. They can impact either one or both lungs, which leads to a severe impairment of a person’...Thoracic diseases pose significant risks to an individual's chest health and are among the most perilous medical diseases. They can impact either one or both lungs, which leads to a severe impairment of a person’s ability to breathe normally. Some notable examples of such diseases encompass pneumonia, lung cancer, coronavirus disease 2019 (COVID-19), tuberculosis, and chronic obstructive pulmonary disease (COPD). Consequently, early and precise detection of these diseases is paramount during the diagnostic process. Traditionally, the primary methods employed for the detection involve the use of X-ray imaging or computed tomography (CT) scans. Nevertheless, due to the scarcity of proficient radiologists and the inherent similarities between these diseases, the accuracy of detection can be compromised, leading to imprecise or erroneous results. To address this challenge, scientists have turned to computer-based solutions, aiming for swift and accurate diagnoses. The primary objective of this study is to develop two machine learning models, utilizing single-task and multi-task learning frameworks, to enhance classification accuracy. Within the multi-task learning architecture, two principal approaches exist soft parameter sharing and hard parameter sharing. Consequently, this research adopts a multi-task deep learning approach that leverages CNNs to achieve improved classification performance for the specified tasks. These tasks, focusing on pneumonia and COVID-19, are processed and learned simultaneously within a multi-task model. To assess the effectiveness of the trained model, it is rigorously validated using three different real-world datasets for training and testing.展开更多
In the era of the Internet of Things(IoT),the crowdsourcing process is driven by data collected by devices that interact with each other and with the physical world.As a part of the IoT ecosystem,task assignment has b...In the era of the Internet of Things(IoT),the crowdsourcing process is driven by data collected by devices that interact with each other and with the physical world.As a part of the IoT ecosystem,task assignment has become an important goal of the research community.Existing task assignment algorithms can be categorized as offline(performs better with datasets but struggles to achieve good real-life results)or online(works well with real-life input but is difficult to optimize regarding in-depth assignments).This paper proposes a Cross-regional Online Task(CROT)assignment problem based on the online assignment model.Given the CROT problem,an Online Task Assignment across Regions based on Prediction(OTARP)algorithm is proposed.OTARP is a two-stage graphics-driven bilateral assignment strategy that uses edge cloud and graph embedding to complete task assignments.The first stage uses historical data to make offline predictions,with a graph-driven method for offline bipartite graph matching.The second stage uses a bipartite graph to complete the online task assignment process.This paper proposes accelerating the task assignment process through multiple assignment rounds and optimizing the process by combining offline guidance and online assignment strategies.To encourage crowd workers to complete crowd tasks across regions,an incentive strategy is designed to encourage crowd workers’movement.To avoid the idle problem in the process of crowd worker movement,a drop-by-rider problem is used to help crowd workers accept more crowd tasks,optimize the number of assignments,and increase utility.Finally,through comparison experiments on real datasets,the performance of the proposed algorithm on crowd worker utility value and the matching number is evaluated.展开更多
Battlefield environment simulation process is an important part of battlefield environment information support, which needs to be built around the task process. At present, the interoperability between battlefield env...Battlefield environment simulation process is an important part of battlefield environment information support, which needs to be built around the task process. At present, the interoperability between battlefield environment simulation system and command and control system is still imperfect, and the traditional simulation data model cannot meet war fighters’ high-efficient and accurate understanding and analysis on battlefield environment’s information. Therefore, a kind of task-orientated battlefield environment simulation process model needs to be construed to effectively analyze the key information demands of the command and control system. The structured characteristics of tasks and simulation process are analyzed, and the simulation process concept model is constructed with the method of object-orientated. The data model and formal syntax of GeoBML are analyzed, and the logical model of simulation process is constructed with formal language. The object data structure of simulation process is defined and the object model of simulation process which maps tasks is constructed. In the end, the battlefield environment simulation platform modules are designed and applied based on this model, verifying that the model can effectively express the real-time dynamic correlation between battlefield environment simulation data and operational tasks.展开更多
In the present context of increasing social demands for natural science education,increasing people s awareness of environmental biodiversity protection,and ecological civilization lifting to the state strategy,it is ...In the present context of increasing social demands for natural science education,increasing people s awareness of environmental biodiversity protection,and ecological civilization lifting to the state strategy,it is just the time to explore a new botany field practice model.The attempt of a new task-driven model for botany field practice will greatly enhance students thinking about plants and nature,plants and environment,and plant and ecological civilization,and will inevitably enhance students initiative awareness and practical ability to protect and rationally utilize plant resources.展开更多
This paper studies the coordinated planning of transmission tasks in the heterogeneous space networks to enable efficient sharing of ground stations cross satellite systems.Specifically,we first formulate the coordina...This paper studies the coordinated planning of transmission tasks in the heterogeneous space networks to enable efficient sharing of ground stations cross satellite systems.Specifically,we first formulate the coordinated planning problem into a mixed integer liner programming(MILP)problem based on time expanded graph.Then,the problem is transferred and reformulated into a consensus optimization framework which can be solved by satellite systems parallelly.With alternating direction method of multipliers(ADMM),a semi-distributed coordinated transmission task planning algorithm is proposed,in which each satellite system plans its own tasks based on local information and limited communication with the coordination center.Simulation results demonstrate that compared with the centralized and fully-distributed methods,the proposed semi-distributed coordinated method can strike a better balance among task complete rate,complexity,and the amount of information required to be exchanged.展开更多
This paper proposes new methods and strategies for Multi-UAVs cooperative attacks with safety and time constraints in a complex environment.Delaunay triangle is designed to construct a map of the complex flight enviro...This paper proposes new methods and strategies for Multi-UAVs cooperative attacks with safety and time constraints in a complex environment.Delaunay triangle is designed to construct a map of the complex flight environment for aerial vehicles.Delaunay-Map,Safe Flight Corridor(SFC),and Relative Safe Flight Corridor(RSFC)are applied to ensure each UAV flight trajectory's safety.By using such techniques,it is possible to avoid the collision with obstacles and collision between UAVs.Bezier-curve is further developed to ensure that multi-UAVs can simultaneously reach the target at the specified time,and the trajectory is within the flight corridor.The trajectory tracking controller is also designed based on model predictive control to track the planned trajectory accurately.The simulation and experiment results are presented to verifying developed strategies of Multi-UAV cooperative attacks.展开更多
Deploying service nodes hierarchically at the edge of the network can effectively improve the service quality of offloaded task requests and increase the utilization of resources.In this paper,we study the task schedu...Deploying service nodes hierarchically at the edge of the network can effectively improve the service quality of offloaded task requests and increase the utilization of resources.In this paper,we study the task scheduling problem in the hierarchically deployed edge cloud.We first formulate the minimization of the service time of scheduled tasks in edge cloud as a combinatorial optimization problem,blue and then prove the NP-hardness of the problem.Different from the existing work that mostly designs heuristic approximation-based algorithms or policies to make scheduling decision,we propose a newly designed scheduling policy,named Joint Neural Network and Heuristic Scheduling(JNNHSP),which combines a neural network-based method with a heuristic based solution.JNNHSP takes the Sequence-to-Sequence(Seq2Seq)model trained by Reinforcement Learning(RL)as the primary policy and adopts the heuristic algorithm as the auxiliary policy to obtain the scheduling solution,thereby achieving a good balance between the quality and the efficiency of the scheduling solution.In-depth experiments show that compared with a variety of related policies and optimization solvers,JNNHSP can achieve better performance in terms of scheduling error ratio,the degree to which the policy is affected by re-sources limitations,average service latency,and execution efficiency in a typical hierarchical edge cloud.展开更多
Well organized datacentres with interconnected servers constitute the cloud computing infrastructure.User requests are submitted through an interface to these servers that provide service to them in an on-demand basis...Well organized datacentres with interconnected servers constitute the cloud computing infrastructure.User requests are submitted through an interface to these servers that provide service to them in an on-demand basis.The scientific applications that get executed at cloud by making use of the heterogeneous resources being allocated to them in a dynamic manner are grouped under NP hard problem category.Task scheduling in cloud poses numerous challenges impacting the cloud performance.If not handled properly,user satisfaction becomes questionable.More recently researchers had come up with meta-heuristic type of solutions for enriching the task scheduling activity in the cloud environment.The prime aim of task scheduling is to utilize the resources available in an optimal manner and reduce the time span of task execution.An improvised seagull optimization algorithm which combines the features of the Cuckoo search(CS)and seagull optimization algorithm(SOA)had been proposed in this work to enhance the performance of the scheduling activity inside the cloud computing environment.The proposed algorithm aims to minimize the cost and time parameters that are spent during task scheduling in the heterogeneous cloud environment.Performance evaluation of the proposed algorithm had been performed using the Cloudsim 3.0 toolkit by comparing it with Multi objective-Ant Colony Optimization(MO-ACO),ACO and Min-Min algorithms.The proposed SOA-CS technique had produced an improvement of 1.06%,4.2%,and 2.4%for makespan and had reduced the overall cost to the extent of 1.74%,3.93%and 2.77%when compared with PSO,ACO,IDEA algorithms respectively when 300 vms are considered.The comparative simulation results obtained had shown that the proposed improvised seagull optimization algorithm fares better than other contemporaries.展开更多
Cloud computing has taken over the high-performance distributed computing area,and it currently provides on-demand services and resource polling over the web.As a result of constantly changing user service demand,the ...Cloud computing has taken over the high-performance distributed computing area,and it currently provides on-demand services and resource polling over the web.As a result of constantly changing user service demand,the task scheduling problem has emerged as a critical analytical topic in cloud computing.The primary goal of scheduling tasks is to distribute tasks to available processors to construct the shortest possible schedule without breaching precedence restrictions.Assignments and schedules of tasks substantially influence system operation in a heterogeneous multiprocessor system.The diverse processes inside the heuristic-based task scheduling method will result in varying makespan in the heterogeneous computing system.As a result,an intelligent scheduling algorithm should efficiently determine the priority of every subtask based on the resources necessary to lower the makespan.This research introduced a novel efficient scheduling task method in cloud computing systems based on the cooperation search algorithm to tackle an essential task and schedule a heterogeneous cloud computing problem.The basic idea of thismethod is to use the advantages of meta-heuristic algorithms to get the optimal solution.We assess our algorithm’s performance by running it through three scenarios with varying numbers of tasks.The findings demonstrate that the suggested technique beats existingmethods NewGenetic Algorithm(NGA),Genetic Algorithm(GA),Whale Optimization Algorithm(WOA),Gravitational Search Algorithm(GSA),and Hybrid Heuristic and Genetic(HHG)by 7.9%,2.1%,8.8%,7.7%,3.4%respectively according to makespan.展开更多
With the arrival of 5G,latency-sensitive applications are becoming increasingly diverse.Mobile Edge Computing(MEC)technology has the characteristics of high bandwidth,low latency and low energy consumption,and has att...With the arrival of 5G,latency-sensitive applications are becoming increasingly diverse.Mobile Edge Computing(MEC)technology has the characteristics of high bandwidth,low latency and low energy consumption,and has attracted much attention among researchers.To improve the Quality of Service(QoS),this study focuses on computation offloading in MEC.We consider the QoS from the perspective of computational cost,dimensional disaster,user privacy and catastrophic forgetting of new users.The QoS model is established based on the delay and energy consumption and is based on DDQN and a Federated Learning(FL)adaptive task offloading algorithm in MEC.The proposed algorithm combines the QoS model and deep reinforcement learning algorithm to obtain an optimal offloading policy according to the local link and node state information in the channel coherence time to address the problem of time-varying transmission channels and reduce the computing energy consumption and task processing delay.To solve the problems of privacy and catastrophic forgetting,we use FL to make distributed use of multiple users’data to obtain the decision model,protect data privacy and improve the model universality.In the process of FL iteration,the communication delay of individual devices is too large,which affects the overall delay cost.Therefore,we adopt a communication delay optimization algorithm based on the unary outlier detection mechanism to reduce the communication delay of FL.The simulation results indicate that compared with existing schemes,the proposed method significantly reduces the computation cost on a device and improves the QoS when handling complex tasks.展开更多
The most critical obstacle for four-wheel independently driven electric vehicles(4WID-EVs)is the driving range.Being the actuators of 4WID-EVs,motors account for its major power consumption.In this sense,by properly d...The most critical obstacle for four-wheel independently driven electric vehicles(4WID-EVs)is the driving range.Being the actuators of 4WID-EVs,motors account for its major power consumption.In this sense,by properly distributing torques to minimize the power consumption,the driving range of 4WID-EV can be effectively improved.This paper proposes a model predictive control(MPC)-based torque distribution scheme,which minimizes the power consumption of 4WID-EVs while guaranteeing its tracking performance of planar motions.By incorporating the motor model considering iron losses,the optimal torque distribution can be achieved without an additional torque controller.Also,for this reason,the proposed control scheme is computationally efficient,since the power consumption term to be optimized,which is expressed as the product of the motor voltages and currents,is much simpler than that derived from the efficiency map.With reasonable simplification and linearization,the MPC problem is converted to a quadratic programming problem,which can be solved efficiently.The simulation results in MATLAB and CarSim co-simulation environments demonstrate that the proposed scheme effectively reduces power consumption with guaranteed tracking performance.展开更多
文摘Brain tissue is one of the softest parts of the human body,composed of white matter and grey matter.The mechanical behavior of the brain tissue plays an essential role in regulating brain morphology and brain function.Besides,traumatic brain injury(TBI)and various brain diseases are also greatly influenced by the brain's mechanical properties.Whether white matter or grey matter,brain tissue contains multiscale structures composed of neurons,glial cells,fibers,blood vessels,etc.,each with different mechanical properties.As such,brain tissue exhibits complex mechanical behavior,usually with strong nonlinearity,heterogeneity,and directional dependence.Building a constitutive law for multiscale brain tissue using traditional function-based approaches can be very challenging.Instead,this paper proposes a data-driven approach to establish the desired mechanical model of brain tissue.We focus on blood vessels with internal pressure embedded in a white or grey matter matrix material to demonstrate our approach.The matrix is described by an isotropic or anisotropic nonlinear elastic model.A representative unit cell(RUC)with blood vessels is built,which is used to generate the stress-strain data under different internal blood pressure and various proportional displacement loading paths.The generated stress-strain data is then used to train a mechanical law using artificial neural networks to predict the macroscopic mechanical response of brain tissue under different internal pressures.Finally,the trained material model is implemented into finite element software to predict the mechanical behavior of a whole brain under intracranial pressure and distributed body forces.Compared with a direct numerical simulation that employs a reference material model,our proposed approach greatly reduces the computational cost and improves modeling efficiency.The predictions made by our trained model demonstrate sufficient accuracy.Specifically,we find that the level of internal blood pressure can greatly influence stress distribution and determine the possible related damage behaviors.
基金Project supported by the National Key R&D Program of China(Grant Nos.2022YFE03070000 and 2022YFE03070003)the National Natural Science Foundation of China(Grant Nos.12375220 and 12075114)+3 种基金the Hunan Provincial Natural Science Foundation(Grant No.2021JJ30569)the Doctoral Initiation Fund Project of University of South China(Grant No.190XQD114)the Hunan Nuclear Fusion International Science and Technology Innovation Cooperation Base(Grant No.2018WK4009)the Hengyang Key Laboratory of Magnetic Confinement Nuclear Fusion Research(Grant No.2018KJ108)。
文摘Through theoretical analysis,we construct a physical model that includes the influence of counter-external driven current opposite to the plasma current direction in the neoclassical tearing mode(NTM).The equation is used with this model to obtain the modified Rutherford equation with co-current and counter-current contributions.Consistent with the reported experimental results,numerical simulations have shown that the localized counter external current can only partially suppress NTM when it is far from the resonant magnetic surface.Under some circumstances,the Ohkawa mechanism dominated current drive(OKCD)by electron cyclotron waves can concurrently create both co-current and counter-current.In this instance,the minimal electron cyclotron wave power that suppresses a particular NTM was calculated by the Rutherford equation.The result is marginally less than when taking co-current alone into consideration.As a result,to suppress NTM using OKCD,one only needs to align the co-current with a greater OKCD peak well with the resonant magnetic surface.The effect of its lower counter-current does not need to be considered because the location of the counter-current deviates greatly from the resonant magnetic surface.
基金supported in part by the National Key R&D Program of China under Grant 2020YFB1005900the National Natural Science Foundation of China under Grant 62001220+3 种基金the Jiangsu Provincial Key Research and Development Program under Grants BE2022068the Natural Science Foundation of Jiangsu Province under Grants BK20200440the Future Network Scientific Research Fund Project FNSRFP-2021-YB-03the Young Elite Scientist Sponsorship Program,China Association for Science and Technology.
文摘Collaborative edge computing is a promising direction to handle the computation intensive tasks in B5G wireless networks.However,edge computing servers(ECSs)from different operators may not trust each other,and thus the incentives for collaboration cannot be guaranteed.In this paper,we propose a consortium blockchain enabled collaborative edge computing framework,where users can offload computing tasks to ECSs from different operators.To minimize the total delay of users,we formulate a joint task offloading and resource optimization problem,under the constraint of the computing capability of each ECS.We apply the Tammer decomposition method and heuristic optimization algorithms to obtain the optimal solution.Finally,we propose a reputation based node selection approach to facilitate the consensus process,and also consider a completion time based primary node selection to avoid monopolization of certain edge node and enhance the security of the blockchain.Simulation results validate the effectiveness of the proposed algorithm,and the total delay can be reduced by up to 40%compared with the non-cooperative case.
基金funding from TECNALIA,Basque Research and Technology Alliance(BRTA)supported by the project aOptimization of Deep Learning algorithms for Edge IoT devices for sensorization and control in Buildings and Infrastructures(EMBED)funded by the Gipuzkoa Provincial Council and approved under the 2023 call of the Guipuzcoan Network of Science,Technology and Innovation Program with File Number 2023-CIEN-000051-01.
文摘In a network environment composed of different types of computing centers that can be divided into different layers(clod,edge layer,and others),the interconnection between them offers the possibility of peer-to-peer task offloading.For many resource-constrained devices,the computation of many types of tasks is not feasible because they cannot support such computations as they do not have enough available memory and processing capacity.In this scenario,it is worth considering transferring these tasks to resource-rich platforms,such as Edge Data Centers or remote cloud servers.For different reasons,it is more exciting and appropriate to download various tasks to specific download destinations depending on the properties and state of the environment and the nature of the functions.At the same time,establishing an optimal offloading policy,which ensures that all tasks are executed within the required latency and avoids excessive workload on specific computing centers is not easy.This study presents two alternatives to solve the offloading decision paradigm by introducing two well-known algorithms,Graph Neural Networks(GNN)and Deep Q-Network(DQN).It applies the alternatives on a well-known Edge Computing simulator called PureEdgeSimand compares them with the two defaultmethods,Trade-Off and Round Robin.Experiments showed that variants offer a slight improvement in task success rate and workload distribution.In terms of energy efficiency,they provided similar results.Finally,the success rates of different computing centers are tested,and the lack of capacity of remote cloud servers to respond to applications in real-time is demonstrated.These novel ways of finding a download strategy in a local networking environment are unique as they emulate the state and structure of the environment innovatively,considering the quality of its connections and constant updates.The download score defined in this research is a crucial feature for determining the quality of a download path in the GNN training process and has not previously been proposed.Simultaneously,the suitability of Reinforcement Learning(RL)techniques is demonstrated due to the dynamism of the network environment,considering all the key factors that affect the decision to offload a given task,including the actual state of all devices.
基金supported in part by the National Natural Science Foundation of China under Grants 62201105,62331017,and 62075024in part by the Natural Science Foundation of Chongqing under Grant cstc2021jcyj-msxmX0404+1 种基金in part by the Chongqing Municipal Education Commission under Grant KJQN202100643in part by Guangdong Basic and Applied Basic Research Foundation under Grant 2022A1515110056.
文摘The Multi-access Edge Cloud(MEC) networks extend cloud computing services and capabilities to the edge of the networks. By bringing computation and storage capabilities closer to end-users and connected devices, MEC networks can support a wide range of applications. MEC networks can also leverage various types of resources, including computation resources, network resources, radio resources,and location-based resources, to provide multidimensional resources for intelligent applications in 5/6G.However, tasks generated by users often consist of multiple subtasks that require different types of resources. It is a challenging problem to offload multiresource task requests to the edge cloud aiming at maximizing benefits due to the heterogeneity of resources provided by devices. To address this issue,we mathematically model the task requests with multiple subtasks. Then, the problem of task offloading of multi-resource task requests is proved to be NP-hard. Furthermore, we propose a novel Dual-Agent Deep Reinforcement Learning algorithm with Node First and Link features(NF_L_DA_DRL) based on the policy network, to optimize the benefits generated by offloading multi-resource task requests in MEC networks. Finally, simulation results show that the proposed algorithm can effectively improve the benefit of task offloading with higher resource utilization compared with baseline algorithms.
基金supported by National Natural Science Foundation of China(Grant No.62071377,62101442,62201456)Natural Science Foundation of Shaanxi Province(Grant No.2023-YBGY-036,2022JQ-687)The Graduate Student Innovation Foundation Project of Xi’an University of Posts and Telecommunications under Grant CXJJDL2022003.
文摘The Internet of Medical Things(Io MT) is regarded as a critical technology for intelligent healthcare in the foreseeable 6G era. Nevertheless, due to the limited computing power capability of edge devices and task-related coupling relationships, Io MT faces unprecedented challenges. Considering the associative connections among tasks, this paper proposes a computing offloading policy for multiple-user devices(UDs) considering device-to-device(D2D) communication and a multi-access edge computing(MEC)technique under the scenario of Io MT. Specifically,to minimize the total delay and energy consumption concerning the requirement of Io MT, we first analyze and model the detailed local execution, MEC execution, D2D execution, and associated tasks offloading exchange model. Consequently, the associated tasks’ offloading scheme of multi-UDs is formulated as a mixed-integer nonconvex optimization problem. Considering the advantages of deep reinforcement learning(DRL) in processing tasks related to coupling relationships, a Double DQN based associative tasks computing offloading(DDATO) algorithm is then proposed to obtain the optimal solution, which can make the best offloading decision under the condition that tasks of UDs are associative. Furthermore, to reduce the complexity of the DDATO algorithm, the cacheaided procedure is intentionally introduced before the data training process. This avoids redundant offloading and computing procedures concerning tasks that previously have already been cached by other UDs. In addition, we use a dynamic ε-greedy strategy in the action selection section of the algorithm, thus preventing the algorithm from falling into a locally optimal solution. Simulation results demonstrate that compared with other existing methods for associative task models concerning different structures in the Io MT network, the proposed algorithm can lower the total cost more effectively and efficiently while also providing a tradeoff between delay and energy consumption tolerance.
文摘The widespread use of advanced electronic devices has led to the emergence of spatial crowdsourcing,a method that taps into collective efforts to perform real-world tasks like environmental monitoring and traffic surveillance.Our research focuses on a specific type of spatial crowdsourcing that involves ongoing,collaborative efforts for continuous spatial data acquisition.However,due to limited budgets and workforce availability,the collected data often lacks completeness,posing a data deficiency problem.To address this,we propose a reciprocal framework to optimize task assignments by leveraging the mutual benefits of spatiotemporal subtask execution.We introduce an entropy-based quality metric to capture the combined effects of incomplete data acquisition and interpolation imprecision.Building on this,we explore a quality-aware task assignment method,corresponding to spatiotemporal assignment strategies.Since the assignment problem is NP-hard,we develop a polynomial-time algorithm with the guaranteed approximation ratio.Novel indexing and pruning techniques are proposed to further enhance performance.Extensive experiments conducted on datasets validate the effectiveness of our methods.
文摘Thoracic diseases pose significant risks to an individual's chest health and are among the most perilous medical diseases. They can impact either one or both lungs, which leads to a severe impairment of a person’s ability to breathe normally. Some notable examples of such diseases encompass pneumonia, lung cancer, coronavirus disease 2019 (COVID-19), tuberculosis, and chronic obstructive pulmonary disease (COPD). Consequently, early and precise detection of these diseases is paramount during the diagnostic process. Traditionally, the primary methods employed for the detection involve the use of X-ray imaging or computed tomography (CT) scans. Nevertheless, due to the scarcity of proficient radiologists and the inherent similarities between these diseases, the accuracy of detection can be compromised, leading to imprecise or erroneous results. To address this challenge, scientists have turned to computer-based solutions, aiming for swift and accurate diagnoses. The primary objective of this study is to develop two machine learning models, utilizing single-task and multi-task learning frameworks, to enhance classification accuracy. Within the multi-task learning architecture, two principal approaches exist soft parameter sharing and hard parameter sharing. Consequently, this research adopts a multi-task deep learning approach that leverages CNNs to achieve improved classification performance for the specified tasks. These tasks, focusing on pneumonia and COVID-19, are processed and learned simultaneously within a multi-task model. To assess the effectiveness of the trained model, it is rigorously validated using three different real-world datasets for training and testing.
基金supported in part by the National Natural Science Foundation of China under Grant 62072392,Grant 61822602,Grant 61772207,Grant 61802331,Grant 61602399,Grant 61702439,Grant 61773331,and Grant 62062034the China Postdoctoral Science Foundation under Grant 2019T120732 and Grant 2017M622691+2 种基金the Natural Science Foundation of Shandong Province under Grant ZR2016FM42the Major scientific and technological innovation projects of Shandong Province under Grant 2019JZZY020131the Key projects of Shandong Natural Science Foundation under Grant ZR2020KF019.
文摘In the era of the Internet of Things(IoT),the crowdsourcing process is driven by data collected by devices that interact with each other and with the physical world.As a part of the IoT ecosystem,task assignment has become an important goal of the research community.Existing task assignment algorithms can be categorized as offline(performs better with datasets but struggles to achieve good real-life results)or online(works well with real-life input but is difficult to optimize regarding in-depth assignments).This paper proposes a Cross-regional Online Task(CROT)assignment problem based on the online assignment model.Given the CROT problem,an Online Task Assignment across Regions based on Prediction(OTARP)algorithm is proposed.OTARP is a two-stage graphics-driven bilateral assignment strategy that uses edge cloud and graph embedding to complete task assignments.The first stage uses historical data to make offline predictions,with a graph-driven method for offline bipartite graph matching.The second stage uses a bipartite graph to complete the online task assignment process.This paper proposes accelerating the task assignment process through multiple assignment rounds and optimizing the process by combining offline guidance and online assignment strategies.To encourage crowd workers to complete crowd tasks across regions,an incentive strategy is designed to encourage crowd workers’movement.To avoid the idle problem in the process of crowd worker movement,a drop-by-rider problem is used to help crowd workers accept more crowd tasks,optimize the number of assignments,and increase utility.Finally,through comparison experiments on real datasets,the performance of the proposed algorithm on crowd worker utility value and the matching number is evaluated.
基金The National Natural Science Foundation of China(41271393).
文摘Battlefield environment simulation process is an important part of battlefield environment information support, which needs to be built around the task process. At present, the interoperability between battlefield environment simulation system and command and control system is still imperfect, and the traditional simulation data model cannot meet war fighters’ high-efficient and accurate understanding and analysis on battlefield environment’s information. Therefore, a kind of task-orientated battlefield environment simulation process model needs to be construed to effectively analyze the key information demands of the command and control system. The structured characteristics of tasks and simulation process are analyzed, and the simulation process concept model is constructed with the method of object-orientated. The data model and formal syntax of GeoBML are analyzed, and the logical model of simulation process is constructed with formal language. The object data structure of simulation process is defined and the object model of simulation process which maps tasks is constructed. In the end, the battlefield environment simulation platform modules are designed and applied based on this model, verifying that the model can effectively express the real-time dynamic correlation between battlefield environment simulation data and operational tasks.
基金Supported by Special Fund for Reform of Teaching Model of Huanggang Normal University(2016CK06,2018CE42)
文摘In the present context of increasing social demands for natural science education,increasing people s awareness of environmental biodiversity protection,and ecological civilization lifting to the state strategy,it is just the time to explore a new botany field practice model.The attempt of a new task-driven model for botany field practice will greatly enhance students thinking about plants and nature,plants and environment,and plant and ecological civilization,and will inevitably enhance students initiative awareness and practical ability to protect and rationally utilize plant resources.
基金supported in part by the NSF China under Grant(61701365,61801365,62001347)in part by Natural Science Foundation of Shaanxi Province(2020JQ-686)+4 种基金in part by the China Postdoctoral Science Foundation under Grant(2018M643581,2019TQ0210,2019TQ0241,2020M673344)in part by Young Talent fund of University Association for Science and Technology in Shaanxi,China(20200112)in part by Key Research and Development Program in Shaanxi Province of China(2021GY066)in part by Postdoctoral Foundation in Shaanxi Province of China(2018BSHEDZZ47)the Fundamental Research Funds for the Central Universities。
文摘This paper studies the coordinated planning of transmission tasks in the heterogeneous space networks to enable efficient sharing of ground stations cross satellite systems.Specifically,we first formulate the coordinated planning problem into a mixed integer liner programming(MILP)problem based on time expanded graph.Then,the problem is transferred and reformulated into a consensus optimization framework which can be solved by satellite systems parallelly.With alternating direction method of multipliers(ADMM),a semi-distributed coordinated transmission task planning algorithm is proposed,in which each satellite system plans its own tasks based on local information and limited communication with the coordination center.Simulation results demonstrate that compared with the centralized and fully-distributed methods,the proposed semi-distributed coordinated method can strike a better balance among task complete rate,complexity,and the amount of information required to be exchanged.
基金National Natural Science Foundation of China(No.61903350)Beijing Institute of Technology Research Fund Program for Young Scholars。
文摘This paper proposes new methods and strategies for Multi-UAVs cooperative attacks with safety and time constraints in a complex environment.Delaunay triangle is designed to construct a map of the complex flight environment for aerial vehicles.Delaunay-Map,Safe Flight Corridor(SFC),and Relative Safe Flight Corridor(RSFC)are applied to ensure each UAV flight trajectory's safety.By using such techniques,it is possible to avoid the collision with obstacles and collision between UAVs.Bezier-curve is further developed to ensure that multi-UAVs can simultaneously reach the target at the specified time,and the trajectory is within the flight corridor.The trajectory tracking controller is also designed based on model predictive control to track the planned trajectory accurately.The simulation and experiment results are presented to verifying developed strategies of Multi-UAV cooperative attacks.
基金Supported by Scientific and Technological Innovation Project of Chongqing(No.cstc2021jxjl20010)The Graduate Student Innovation Program of Chongqing University of Technology(No.clgycx-20203166,No.gzlcx20222061,No.gzlcx20223229)。
文摘Deploying service nodes hierarchically at the edge of the network can effectively improve the service quality of offloaded task requests and increase the utilization of resources.In this paper,we study the task scheduling problem in the hierarchically deployed edge cloud.We first formulate the minimization of the service time of scheduled tasks in edge cloud as a combinatorial optimization problem,blue and then prove the NP-hardness of the problem.Different from the existing work that mostly designs heuristic approximation-based algorithms or policies to make scheduling decision,we propose a newly designed scheduling policy,named Joint Neural Network and Heuristic Scheduling(JNNHSP),which combines a neural network-based method with a heuristic based solution.JNNHSP takes the Sequence-to-Sequence(Seq2Seq)model trained by Reinforcement Learning(RL)as the primary policy and adopts the heuristic algorithm as the auxiliary policy to obtain the scheduling solution,thereby achieving a good balance between the quality and the efficiency of the scheduling solution.In-depth experiments show that compared with a variety of related policies and optimization solvers,JNNHSP can achieve better performance in terms of scheduling error ratio,the degree to which the policy is affected by re-sources limitations,average service latency,and execution efficiency in a typical hierarchical edge cloud.
文摘Well organized datacentres with interconnected servers constitute the cloud computing infrastructure.User requests are submitted through an interface to these servers that provide service to them in an on-demand basis.The scientific applications that get executed at cloud by making use of the heterogeneous resources being allocated to them in a dynamic manner are grouped under NP hard problem category.Task scheduling in cloud poses numerous challenges impacting the cloud performance.If not handled properly,user satisfaction becomes questionable.More recently researchers had come up with meta-heuristic type of solutions for enriching the task scheduling activity in the cloud environment.The prime aim of task scheduling is to utilize the resources available in an optimal manner and reduce the time span of task execution.An improvised seagull optimization algorithm which combines the features of the Cuckoo search(CS)and seagull optimization algorithm(SOA)had been proposed in this work to enhance the performance of the scheduling activity inside the cloud computing environment.The proposed algorithm aims to minimize the cost and time parameters that are spent during task scheduling in the heterogeneous cloud environment.Performance evaluation of the proposed algorithm had been performed using the Cloudsim 3.0 toolkit by comparing it with Multi objective-Ant Colony Optimization(MO-ACO),ACO and Min-Min algorithms.The proposed SOA-CS technique had produced an improvement of 1.06%,4.2%,and 2.4%for makespan and had reduced the overall cost to the extent of 1.74%,3.93%and 2.77%when compared with PSO,ACO,IDEA algorithms respectively when 300 vms are considered.The comparative simulation results obtained had shown that the proposed improvised seagull optimization algorithm fares better than other contemporaries.
文摘Cloud computing has taken over the high-performance distributed computing area,and it currently provides on-demand services and resource polling over the web.As a result of constantly changing user service demand,the task scheduling problem has emerged as a critical analytical topic in cloud computing.The primary goal of scheduling tasks is to distribute tasks to available processors to construct the shortest possible schedule without breaching precedence restrictions.Assignments and schedules of tasks substantially influence system operation in a heterogeneous multiprocessor system.The diverse processes inside the heuristic-based task scheduling method will result in varying makespan in the heterogeneous computing system.As a result,an intelligent scheduling algorithm should efficiently determine the priority of every subtask based on the resources necessary to lower the makespan.This research introduced a novel efficient scheduling task method in cloud computing systems based on the cooperation search algorithm to tackle an essential task and schedule a heterogeneous cloud computing problem.The basic idea of thismethod is to use the advantages of meta-heuristic algorithms to get the optimal solution.We assess our algorithm’s performance by running it through three scenarios with varying numbers of tasks.The findings demonstrate that the suggested technique beats existingmethods NewGenetic Algorithm(NGA),Genetic Algorithm(GA),Whale Optimization Algorithm(WOA),Gravitational Search Algorithm(GSA),and Hybrid Heuristic and Genetic(HHG)by 7.9%,2.1%,8.8%,7.7%,3.4%respectively according to makespan.
基金supported by the National Natural Science Foundation of China(62032013,62072094Liaoning Province Science and Technology Fund Project(2020MS086)+1 种基金Shenyang Science and Technology Plan Project(20206424)the Fundamental Research Funds for the Central Universities(N2116014,N180101028)CERNET Innovation Project(NGII20190504).
文摘With the arrival of 5G,latency-sensitive applications are becoming increasingly diverse.Mobile Edge Computing(MEC)technology has the characteristics of high bandwidth,low latency and low energy consumption,and has attracted much attention among researchers.To improve the Quality of Service(QoS),this study focuses on computation offloading in MEC.We consider the QoS from the perspective of computational cost,dimensional disaster,user privacy and catastrophic forgetting of new users.The QoS model is established based on the delay and energy consumption and is based on DDQN and a Federated Learning(FL)adaptive task offloading algorithm in MEC.The proposed algorithm combines the QoS model and deep reinforcement learning algorithm to obtain an optimal offloading policy according to the local link and node state information in the channel coherence time to address the problem of time-varying transmission channels and reduce the computing energy consumption and task processing delay.To solve the problems of privacy and catastrophic forgetting,we use FL to make distributed use of multiple users’data to obtain the decision model,protect data privacy and improve the model universality.In the process of FL iteration,the communication delay of individual devices is too large,which affects the overall delay cost.Therefore,we adopt a communication delay optimization algorithm based on the unary outlier detection mechanism to reduce the communication delay of FL.The simulation results indicate that compared with existing schemes,the proposed method significantly reduces the computation cost on a device and improves the QoS when handling complex tasks.
基金supported in part by National Natural Science Foundation of China(NSFC)under Project No.51737010.
文摘The most critical obstacle for four-wheel independently driven electric vehicles(4WID-EVs)is the driving range.Being the actuators of 4WID-EVs,motors account for its major power consumption.In this sense,by properly distributing torques to minimize the power consumption,the driving range of 4WID-EV can be effectively improved.This paper proposes a model predictive control(MPC)-based torque distribution scheme,which minimizes the power consumption of 4WID-EVs while guaranteeing its tracking performance of planar motions.By incorporating the motor model considering iron losses,the optimal torque distribution can be achieved without an additional torque controller.Also,for this reason,the proposed control scheme is computationally efficient,since the power consumption term to be optimized,which is expressed as the product of the motor voltages and currents,is much simpler than that derived from the efficiency map.With reasonable simplification and linearization,the MPC problem is converted to a quadratic programming problem,which can be solved efficiently.The simulation results in MATLAB and CarSim co-simulation environments demonstrate that the proposed scheme effectively reduces power consumption with guaranteed tracking performance.