As an advanced combat weapon,Unmanned Aerial Vehicles(UAVs)have been widely used in military wars.In this paper,we formulated the Autonomous Navigation Control(ANC)problem of UAVs as a Markov Decision Process(MDP)and ...As an advanced combat weapon,Unmanned Aerial Vehicles(UAVs)have been widely used in military wars.In this paper,we formulated the Autonomous Navigation Control(ANC)problem of UAVs as a Markov Decision Process(MDP)and proposed a novel Deep Reinforcement Learning(DRL)method to allow UAVs to perform dynamic target tracking tasks in large-scale unknown environments.To solve the problem of limited training experience,the proposed Imaginary Filtered Hindsight Experience Replay(IFHER)generates successful episodes by reasonably imagining the target trajectory in the failed episode to augment the experiences.The welldesigned goal,episode,and quality filtering strategies ensure that only high-quality augmented experiences can be stored,while the sampling filtering strategy of IFHER ensures that these stored augmented experiences can be fully learned according to their high priorities.By training in a complex environment constructed based on the parameters of a real UAV,the proposed IFHER algorithm improves the convergence speed by 28.99%and the convergence result by 11.57%compared to the state-of-the-art Twin Delayed Deep Deterministic Policy Gradient(TD3)algorithm.The testing experiments carried out in environments with different complexities demonstrate the strong robustness and generalization ability of the IFHER agent.Moreover,the flight trajectory of the IFHER agent shows the superiority of the learned policy and the practical application value of the algorithm.展开更多
Unmanned Aerial Vehicles(UAVs)play a vital role in military warfare.In a variety of battlefield mission scenarios,UAVs are required to safely fly to designated locations without human intervention.Therefore,finding a ...Unmanned Aerial Vehicles(UAVs)play a vital role in military warfare.In a variety of battlefield mission scenarios,UAVs are required to safely fly to designated locations without human intervention.Therefore,finding a suitable method to solve the UAV Autonomous Motion Planning(AMP)problem can improve the success rate of UAV missions to a certain extent.In recent years,many studies have used Deep Reinforcement Learning(DRL)methods to address the AMP problem and have achieved good results.From the perspective of sampling,this paper designs a sampling method with double-screening,combines it with the Deep Deterministic Policy Gradient(DDPG)algorithm,and proposes the Relevant Experience Learning-DDPG(REL-DDPG)algorithm.The REL-DDPG algorithm uses a Prioritized Experience Replay(PER)mechanism to break the correlation of continuous experiences in the experience pool,finds the experiences most similar to the current state to learn according to the theory in human education,and expands the influence of the learning process on action selection at the current state.All experiments are applied in a complex unknown simulation environment constructed based on the parameters of a real UAV.The training experiments show that REL-DDPG improves the convergence speed and the convergence result compared to the state-of-the-art DDPG algorithm,while the testing experiments show the applicability of the algorithm and investigate the performance under different parameter conditions.展开更多
This paper presents a smart energy community management approach which is capable of implementing P2P trading and managing household energy storage systems.A smart residential community concept is proposed consisting ...This paper presents a smart energy community management approach which is capable of implementing P2P trading and managing household energy storage systems.A smart residential community concept is proposed consisting of domestic users and a local energy pool,in which users are free to trade with the local energy pool and enjoy cheap renewable energy while avoiding the installation of new energy generation equipment.The local energy pool could harvest surplus energy from users and renewable resources,at the same time it sells energy at a higher price than Feed-in-Tariff(FIT)but lower than the retail price.In order to encourage the participation in local energy trading,the electricity price of the energy pool is determined by a real-time demand/supply ratio.Under this pricing mechanism,retail price,users and renewable energy could all affect the electricity price which leads to higher consumers’profits and more optimized utilization of renewable energy.The proposed energy trading process was modeled as a Markov Decision Process(MDP)and a reinforcement learning algorithm was adopted to find the optimal decision in the MDP because of its excellent performance in on-going and model-free tasks.In addition,the fuzzy inference system makes it possible to use Q-learning in continuous state-space problems(Fuzzy Q-learning)considering the infinite possibilities in the energy trading process.To evaluate the performance of the proposed demand side management system,a numerical analysis is conducted in a community comparing the electricity costs before and after using the proposed energy management system.展开更多
In multicellular and even single-celled organisms,individual components are interconnected at multiscale levels to produce enormously complex biological networks that help these systems maintain homeostasis for develo...In multicellular and even single-celled organisms,individual components are interconnected at multiscale levels to produce enormously complex biological networks that help these systems maintain homeostasis for development and environmental adaptation.Systems biology studies initially adopted network analysis to explore how relationships between individual components give rise to complex biological processes.Network analysis has been applied to dissect the complex connectivity of mammalian brains across different scales in time and space in The Human Brain Project.In plant science,network analysis has similarly been applied to study the connectivity of plant components at the molecular,subcellular,cellular,organic,and organism levels.Analysis of these multiscale networks contributes to our understanding of how genotype determines phenotype.In this review,we summarized the theoretical framework of plant multiscale networks and introduced studies investigating plant networks by various experimental and computational modalities.We next discussed the currently available analytic methodologies and multi-level imaging techniques used to map multiscale networks in plants.Finally,we highlighted some of the technical challenges and key questions remaining to be addressed in this emerging field.展开更多
基金co-supported by the National Natural Science Foundation of China(Nos.62003267 and 61573285)the Natural Science Basic Research Plan in Shaanxi Province of China(No.2020JQ-220)+1 种基金the Open Project of Science and Technology on Electronic Information Control Laboratory,China(No.JS20201100339)the Open Project of Science and Technology on Electromagnetic Space Operations and Applications Laboratory,China(No.JS20210586512).
文摘As an advanced combat weapon,Unmanned Aerial Vehicles(UAVs)have been widely used in military wars.In this paper,we formulated the Autonomous Navigation Control(ANC)problem of UAVs as a Markov Decision Process(MDP)and proposed a novel Deep Reinforcement Learning(DRL)method to allow UAVs to perform dynamic target tracking tasks in large-scale unknown environments.To solve the problem of limited training experience,the proposed Imaginary Filtered Hindsight Experience Replay(IFHER)generates successful episodes by reasonably imagining the target trajectory in the failed episode to augment the experiences.The welldesigned goal,episode,and quality filtering strategies ensure that only high-quality augmented experiences can be stored,while the sampling filtering strategy of IFHER ensures that these stored augmented experiences can be fully learned according to their high priorities.By training in a complex environment constructed based on the parameters of a real UAV,the proposed IFHER algorithm improves the convergence speed by 28.99%and the convergence result by 11.57%compared to the state-of-the-art Twin Delayed Deep Deterministic Policy Gradient(TD3)algorithm.The testing experiments carried out in environments with different complexities demonstrate the strong robustness and generalization ability of the IFHER agent.Moreover,the flight trajectory of the IFHER agent shows the superiority of the learned policy and the practical application value of the algorithm.
基金co-supported by the National Natural Science Foundation of China(Nos.62003267,61573285)the Aeronautical Science Foundation of China(ASFC)(No.20175553027)Natural Science Basic Research Plan in Shaanxi Province of China(No.2020JQ-220)。
文摘Unmanned Aerial Vehicles(UAVs)play a vital role in military warfare.In a variety of battlefield mission scenarios,UAVs are required to safely fly to designated locations without human intervention.Therefore,finding a suitable method to solve the UAV Autonomous Motion Planning(AMP)problem can improve the success rate of UAV missions to a certain extent.In recent years,many studies have used Deep Reinforcement Learning(DRL)methods to address the AMP problem and have achieved good results.From the perspective of sampling,this paper designs a sampling method with double-screening,combines it with the Deep Deterministic Policy Gradient(DDPG)algorithm,and proposes the Relevant Experience Learning-DDPG(REL-DDPG)algorithm.The REL-DDPG algorithm uses a Prioritized Experience Replay(PER)mechanism to break the correlation of continuous experiences in the experience pool,finds the experiences most similar to the current state to learn according to the theory in human education,and expands the influence of the learning process on action selection at the current state.All experiments are applied in a complex unknown simulation environment constructed based on the parameters of a real UAV.The training experiments show that REL-DDPG improves the convergence speed and the convergence result compared to the state-of-the-art DDPG algorithm,while the testing experiments show the applicability of the algorithm and investigate the performance under different parameter conditions.
基金This work was supported by the National Natural Science Foundation of China(No.51807024).
文摘This paper presents a smart energy community management approach which is capable of implementing P2P trading and managing household energy storage systems.A smart residential community concept is proposed consisting of domestic users and a local energy pool,in which users are free to trade with the local energy pool and enjoy cheap renewable energy while avoiding the installation of new energy generation equipment.The local energy pool could harvest surplus energy from users and renewable resources,at the same time it sells energy at a higher price than Feed-in-Tariff(FIT)but lower than the retail price.In order to encourage the participation in local energy trading,the electricity price of the energy pool is determined by a real-time demand/supply ratio.Under this pricing mechanism,retail price,users and renewable energy could all affect the electricity price which leads to higher consumers’profits and more optimized utilization of renewable energy.The proposed energy trading process was modeled as a Markov Decision Process(MDP)and a reinforcement learning algorithm was adopted to find the optimal decision in the MDP because of its excellent performance in on-going and model-free tasks.In addition,the fuzzy inference system makes it possible to use Q-learning in continuous state-space problems(Fuzzy Q-learning)considering the infinite possibilities in the energy trading process.To evaluate the performance of the proposed demand side management system,a numerical analysis is conducted in a community comparing the electricity costs before and after using the proposed energy management system.
基金supported by the National Natural Science Foundation of China(31530084,32000558,32000483,and31800504)the Programme of Introducing Talents of Discipline to Universities(111 project,B13007)the China Postdoctoral Science Foundation Grant(2019M660494)。
文摘In multicellular and even single-celled organisms,individual components are interconnected at multiscale levels to produce enormously complex biological networks that help these systems maintain homeostasis for development and environmental adaptation.Systems biology studies initially adopted network analysis to explore how relationships between individual components give rise to complex biological processes.Network analysis has been applied to dissect the complex connectivity of mammalian brains across different scales in time and space in The Human Brain Project.In plant science,network analysis has similarly been applied to study the connectivity of plant components at the molecular,subcellular,cellular,organic,and organism levels.Analysis of these multiscale networks contributes to our understanding of how genotype determines phenotype.In this review,we summarized the theoretical framework of plant multiscale networks and introduced studies investigating plant networks by various experimental and computational modalities.We next discussed the currently available analytic methodologies and multi-level imaging techniques used to map multiscale networks in plants.Finally,we highlighted some of the technical challenges and key questions remaining to be addressed in this emerging field.