期刊文献+

基于分层深度强化学习的移动机器人导航方法 被引量:7

Navigation method for mobile robot based on hierarchical deep reinforcement learning
原文传递
导出
摘要 针对现有基于深度强化学习(deep reinforcement learning,DRL)的分层导航方法在包含长廊、死角等结构的复杂环境下导航效果不佳的问题,提出一种基于option-based分层深度强化学习(hierarchical deep reinforcement learning,HDRL)的移动机器人导航方法.该方法的模型框架分为高层和低层两部分,其中低层的避障和目标驱动控制模型分别实现避障和目标接近两种行为策略,高层的行为选择模型可自动学习稳定、可靠的行为选择策略,从而有效避免对人为设计调控规则的依赖.此外,所提出方法通过对避障控制模型进行优化训练,使学习到的避障策略更加适用于复杂环境下的导航任务.在与现有DRL方法的对比实验中,所提出方法在全部仿真测试环境中均取得最高的导航成功率,同时在其他指标上也具有整体优势,表明所提出方法可有效解决复杂环境下导航效果不佳的问题,且具有较强的泛化能力.此外,真实环境下的测试进一步验证了所提出方法的潜在应用价值. In order to solve the problem that existing hierarchical navigation methods based on deep reinforcement learning(DRL)perform poorly in complex environments including the structures like long corridors and dead corners,we propose a navigation method for mobile robots based on option-based hierarchical deep reinforcement learning(HDRL).The framework of the proposed method consists of two level control models:a low level model is to obtain policies for avoiding obstacles and reaching the goal respectively,and a high-level behavior selection model is for automatically learning stable and reliable behavior selection policy,which does not rely on manually designed control rules.In addition,a training method for optimizing the obstacle avoidance control model is proposed,which makes the learned obstacle avoidance policy more suitable for the navigation task in complex environments.In comparison with existing DRL-based navigation methods,the proposed method achieves the highest navigation success rate in all simulated test environments used in this paper and shows better overall performance on other metrics,which demonstrates the proposed method can effectively solve the problem of poor navigation performance in complex environments and has strong generalization ability.Moreover,experiments in real-world environment also verify the potential application value of the proposed method.
作者 王童 李骜 宋海荦 刘伟 王明会 WANG Tong;LI Ao;SONG Hai-luo;LIU Wei;WANG Ming-hui(School of Information Science and Technology,University of Science and Technology of China,Hefei 230027,China)
出处 《控制与决策》 EI CSCD 北大核心 2022年第11期2799-2807,共9页 Control and Decision
基金 中国科学技术大学优秀引进人才基金项目(KY2100000021) 国家自然科学基金项目(61971393,61871361)。
关键词 深度强化学习 分层深度强化学习 移动机器人 导航 避障 策略学习 deep reinforcement learning hierarchical deep reinforcement learning mobile robot navigation obstacle avoidance policy learning
  • 相关文献

参考文献6

二级参考文献35

  • 1Althaus P,Christensen H I.Smooth task switching through behaviour competition[J].Robotics and Autonomous Systems,2003,44(3/4):241-249.
  • 2Beom H R,Cho H S.A sensor-based navigation for a mobile robot using fuzzy logic and reinforcement learning[J].IEEE Trans on Systems,Man and Cybernetics,1995,25(3):464-477.
  • 3Ranganathan P,Hayet J B,Devy M,et al.Topological navigation and qualitative localization for indoor environment using multi-sensory perception[J].Robotics and Autonomous Systems,2002,41:137-144.
  • 4Sutton R S,Barto A G.Reinforcement learning:an introduction[M].Cambridge,MA:MIT Press,1998:3-9.
  • 5Ye C,Yung N H C,Wang D W.A fuzzy controller with supervised learning assisted reinforcement learning algorithm for obstacle avoidance[J].IEEE Transactions on Systems,Man and Cybernetics,2003,33(1):17-27.
  • 6Kondo T,Ito K.A reinforcement learning with evolutionary state recruitment strategy for autonomous mobile robots control[J].Robotics and Autonomous Systems,2004,46(2):111-124.
  • 7Barto A G,Mahanevan S.Recent advances in hierarchical reinforcement learning[J].Discrete Event Dynamic Systems:Theory and Applications,2003,13(4):41-77.
  • 8Sutton R,Precup D,Singh S.Between MDPs and semi-MDPs:a framework for temporal abstraction in reinforcement learning[J].Artificial Intelligence,1999,112(1/2):181-211.
  • 9Dietterich T G.Hierarchical reinforcement learning with the MAXQ value function decomposition[J].Journal of Artificial Intelligence Research,2000,13(1):227-303.
  • 10Thrun S.Learning metric-topological maps for indoor mobile robot navigation[J].Artificial Intelligence,1998,99(1):21-71.

共引文献207

同被引文献58

引证文献7

二级引证文献7

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部