期刊文献+

一种基于双经验池优先采样的深度强化学习算法

A Deep Reinforcement Learning Algorithm Based on Double Experience Memory and Prioritized Experience Replay
下载PDF
导出
摘要 智能体在游戏、机器人控制、自动驾驶和自然语言处理等领域有着广泛应用。然而,稀疏奖励问题成为智能体学习和探索的困难之一。文章提出了改进算法,采用双经验池存储经验样本,并融入优先经验采样以提高采样效率。同时,对奖励函数进行重构,细分为多段奖励,以引导智能体学习。实验结果表明,改进算法优于传统DQN(Deep Q-Network)算法和同策略的A2C(Advantage Actor-Critic)算法,有效应对了稀疏奖励问题,并提高了智能体的学习效率。在经典Cartpole游戏环境中进行的实验验证了改进算法的优越性。 Intelligent agents have been widely applied in various fields,such as gaming,robot control,autonomous driving,and natural language processing.However,the problem of sparse rewards has become a challenging issue for learning and explora-tion in these domains.This paper proposes an improved algorithm that utilizes dual experience replay buffers and incorporates prioritized experience sampling to enhance the efficiency of data sampling.Furthermore,the reward function is restructured into multiple segments,providing guiding rewards to facilitate the learning process of the intelligent agent.Experimental results demonstrate that the proposed algorithm outperforms traditional Deep Q-Network(DQN)and Advantage Actor-Critic(A2C)al-gorithms,effectively addressing the challenges posed by sparse reward settings and significantly improving the learning effi-ciency of the intelligent agent.The effectiveness of the improved algorithm is validated through experiments conducted in the classic Cartpole gaming environment.
作者 李思博 臧兆祥 吕相霖 LI Sibo;ZANG Zhaoxiang;LYU Xianglin(Hubei Key Laboratory of Itelligegnt Visual Monitoring for Hydropower Engineering,Three Gorges University,Yichang 443002,China;School of Computer and Information,Three Gorges University,Yichang 443002,China)
出处 《长江信息通信》 2023年第11期73-76,共4页 Changjiang Information & Communications
基金 国家自然科学基金(No.61502274) 湖北省自然科学基金(No.2015CFB336)资助项目。
关键词 稀疏奖励 双经验池 优先经验回放 奖励函数 深度强化学习 sparse reward double experience replay memory prioritized experience replay reward function Deep reinforcem-ent learning
  • 相关文献

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部