摘要
基于深度强化学习方法对未知环境的无人机区域覆盖路径规划进行研究,通过搭建栅格环境模型,在环境中随机部署无人机和禁飞区位置,利用双深度Q网络(DDQN)训练无人机的覆盖策略,得到了一套基于DDQN的无人机未知区域覆盖路径规划框架。仿真实验表明,设计的无人机未知区域覆盖路径规划框架在无禁飞区的环境下可以实现完全覆盖,在含有未知数量的禁飞区下也能比较好的完成区域覆盖任务,与DQN方法比较,其平均覆盖率能够在相同训练条件和训练次数下高出2%,与Q-Learning方法和Sarsa方法对比,在无禁飞区的环境中分别高出4%和3%。
The path planning of UAV area coverage in unknown environment is studied based on deep reinforcement learning method.By building agrid environment model,randomly deploying UAV and no-fly zone in the environment,and using a double deep Q-network(DDQN)to train the coverage strategy of UAV,a set of UAV coverage path planning framework base on DDQN is obtained.The simulation experiment shows that the designed UAV unknown area coverage path planning framework can achieve full coverage in the environment without no fly zone,and can also better complete the area coverage task in the environment with an unknown number of no fly zones.Compared with DQN method,its average coverage rate can be 2%higher under the same training conditions and training rounds,higher than Q Learning method and Sarsa method in the environment without no fly zone.
作者
沈骁
赵彤洲
Shen Xiao;Zhao Tongzhou(School of Computer Science&Engineering Artificial Intelligence,Wuhan Institute of Technology,Wuhan 430205,China;Hubei Key Laboratory of Intelligent Robot,Wuhan Institute of Technology,Wuhan 430205,China)
出处
《电子测量技术》
北大核心
2023年第14期30-36,共7页
Electronic Measurement Technology
基金
国家重点研发计划(2016YFC0801003)项目资助
关键词
未知环境
区域覆盖
深度强化学习
路径规划
unknown environments
area coverage
deep reinforcement learning
path planning