访谈时间:2022年5月9日15:00~17:00访谈地点:线上访谈受访者:宋庆华(北京市东城区社区参与行动服务中心创办人、主任)访谈人:王欢(北京航空航天大学公共管理学院博士研究生)、韩怡宁(北京航空航天大学公共管理学院硕士研究生)【北京市...访谈时间:2022年5月9日15:00~17:00访谈地点:线上访谈受访者:宋庆华(北京市东城区社区参与行动服务中心创办人、主任)访谈人:王欢(北京航空航天大学公共管理学院博士研究生)、韩怡宁(北京航空航天大学公共管理学院硕士研究生)【北京市东城区社区参与行动服务中心简介】北京市东城区社区参与行动服务中心又名“北京灿雨石信息咨询中心”(Shining Stone Community Action Center,以下简称“服务中心”)。展开更多
This paper addresses the issue of safety in reinforcement learning(RL)with disturbances and its application in the safety-constrained motion control of autonomous robots.To tackle this problem,a robust Lyapunov value ...This paper addresses the issue of safety in reinforcement learning(RL)with disturbances and its application in the safety-constrained motion control of autonomous robots.To tackle this problem,a robust Lyapunov value function(rLVF)is proposed.The rLVF is obtained by introducing a data-based LVF under the worst-case disturbance of the observed state.Using the rLVF,a uniformly ultimate boundedness criterion is established.This criterion is desired to ensure that the cost function,which serves as a safety criterion,ultimately converges to a range via the policy to be designed.Moreover,to mitigate the drastic variation of the rLVF caused by differences in states,a smoothing regularization of the rLVF is introduced.To train policies with safety guarantees under the worst disturbances of the observed states,an off-policy robust RL algorithm is proposed.The proposed algorithm is applied to motion control tasks of an autonomous vehicle and a cartpole,which involve external disturbances and variations of the model parameters,respectively.The experimental results demonstrate the effectiveness of the theoretical findings and the advantages of the proposed algorithm in terms of robustness and safety.展开更多
文摘访谈时间:2022年5月9日15:00~17:00访谈地点:线上访谈受访者:宋庆华(北京市东城区社区参与行动服务中心创办人、主任)访谈人:王欢(北京航空航天大学公共管理学院博士研究生)、韩怡宁(北京航空航天大学公共管理学院硕士研究生)【北京市东城区社区参与行动服务中心简介】北京市东城区社区参与行动服务中心又名“北京灿雨石信息咨询中心”(Shining Stone Community Action Center,以下简称“服务中心”)。
基金supported by the National Natural Science Foundation of China(Grant Nos.62225305 and 12072088)the Fundamental Research Funds for the Central Universities,China(Grant Nos.HIT.BRET.2022004,HIT.OCEF.2022047,and HIT.DZIJ.2023049)+1 种基金the Grant JCKY2022603C016,State Key Laboratory of Robotics and System(HIT)the Heilongjiang Touyan Team。
文摘This paper addresses the issue of safety in reinforcement learning(RL)with disturbances and its application in the safety-constrained motion control of autonomous robots.To tackle this problem,a robust Lyapunov value function(rLVF)is proposed.The rLVF is obtained by introducing a data-based LVF under the worst-case disturbance of the observed state.Using the rLVF,a uniformly ultimate boundedness criterion is established.This criterion is desired to ensure that the cost function,which serves as a safety criterion,ultimately converges to a range via the policy to be designed.Moreover,to mitigate the drastic variation of the rLVF caused by differences in states,a smoothing regularization of the rLVF is introduced.To train policies with safety guarantees under the worst disturbances of the observed states,an off-policy robust RL algorithm is proposed.The proposed algorithm is applied to motion control tasks of an autonomous vehicle and a cartpole,which involve external disturbances and variations of the model parameters,respectively.The experimental results demonstrate the effectiveness of the theoretical findings and the advantages of the proposed algorithm in terms of robustness and safety.