期刊文献+
共找到4篇文章
< 1 >
每页显示 20 50 100
Curricular Robust Reinforcement Learning via GAN-Based Perturbation Through Continuously Scheduled Task Sequence
1
作者 Yike Li Yunzhe Tian +5 位作者 Endong Tong wenjia niu Yingxiao Xiang Tong Chen Yalun Wu Jiqiang Liu 《Tsinghua Science and Technology》 SCIE EI CAS CSCD 2023年第1期27-38,共12页
Reinforcement learning(RL),one of three branches of machine learning,aims for autonomous learning and is now greatly driving the artificial intelligence development,especially in autonomous distributed systems,such as... Reinforcement learning(RL),one of three branches of machine learning,aims for autonomous learning and is now greatly driving the artificial intelligence development,especially in autonomous distributed systems,such as cooperative Boston Dynamics robots.However,robust RL has been a challenging problem of reliable aspects due to the gap between laboratory simulation and real world.Existing efforts have been made to approach this problem,such as performing random environmental perturbations in the learning process.However,one cannot guarantee to train with a positive perturbation as bad ones might bring failures to RL.In this work,we treat robust RL as a multi-task RL problem,and propose a curricular robust RL approach.We first present a generative adversarial network(GAN)based task generation model to iteratively output new tasks at the appropriate level of difficulty for the current policy.Furthermore,with these progressive tasks,we can realize curricular learning and finally obtain a robust policy.Extensive experiments in multiple environments demonstrate that our method improves the training stability and is robust to differences in training/test conditions. 展开更多
关键词 robust reinforcement learning generative adversarial network(GAN)based model curricular learning
原文传递
Boosting imbalanced data learning with Wiener process oversampling 被引量:1
2
作者 Qian LI Gang LI +4 位作者 wenjia niu Yanan CAO Liang CHANG Jianlong TAN Li GUO 《Frontiers of Computer Science》 SCIE EI CSCD 2017年第5期836-851,共16页
Learning from imbalanced data is a challenging task in a wide range of applications, which attracts significant research efforts from machine learning and data mining community. As a natural approach to this issue, ov... Learning from imbalanced data is a challenging task in a wide range of applications, which attracts significant research efforts from machine learning and data mining community. As a natural approach to this issue, oversampling balances the training samples through replicating existing samples or synthesizing new samples. In general, synthesization outperforms replication by supplying additional information on the minority class. However, the additional information needs to follow the same normal distribution of the training set, which further constrains the new samples within the predefined range of training set. In this paper, we present the Wiener process oversampling (WPO) technique that brings the physics phenomena into sample synthesization. WPO constructs a robust decision region by expanding the attribute ranges in training set while keeping the same normal distribution. The satisfactory performance of WPO can be achieved with much lower computing complexity. In addition, by integrating WPO with ensemble learning, the WPOBoost algorithm outperforms many prevalent imbalance learning solutions. 展开更多
关键词 imbalanced-data learning OVERSAMPLING ensemble learning Wiener process ADABOOST
原文传递
Adversarial attack and defense in reinforcement learning-from AI security view
3
作者 Tong Chen Jiqiang Liu +3 位作者 Yingxiao Xiang wenjia niu Endong Tong Zhen Han 《Cybersecurity》 CSCD 2019年第1期167-188,共22页
Reinforcement learning is a core technology for modern artificial intelligence,and it has become a workhorse for AI applications ranging from Atrai Game to Connected and Automated Vehicle System(CAV).Therefore,a relia... Reinforcement learning is a core technology for modern artificial intelligence,and it has become a workhorse for AI applications ranging from Atrai Game to Connected and Automated Vehicle System(CAV).Therefore,a reliable RL system is the foundation for the security critical applications in AI,which has attracted a concern that is more critical than ever.However,recent studies discover that the interesting attack mode adversarial attack also be effective when targeting neural network policies in the context of reinforcement learning,which has inspired innovative researches in this direction.Hence,in this paper,we give the very first attempt to conduct a comprehensive survey on adversarial attacks in reinforcement learning under AI security.Moreover,we give briefly introduction on the most representative defense technologies against existing adversarial attacks. 展开更多
关键词 Reinforcement learning Artificial intelligence SECURITY Adversarial attack Adversarial example DEFENSE
原文传递
Adversarial attack and defense in reinforcement learning-from AI security view
4
作者 Tong Chen Jiqiang Liu +3 位作者 Yingxiao Xiang wenjia niu Endong Tong Zhen Han 《Cybersecurity》 2018年第1期442-463,共22页
Reinforcement learning is a core technology for modern artificial intelligence,and it has become a workhorse for AI applications ranging from Atrai Game to Connected and Automated Vehicle System(CAV).Therefore,a relia... Reinforcement learning is a core technology for modern artificial intelligence,and it has become a workhorse for AI applications ranging from Atrai Game to Connected and Automated Vehicle System(CAV).Therefore,a reliable RL system is the foundation for the security critical applications in AI,which has attracted a concern that is more critical than ever.However,recent studies discover that the interesting attack mode adversarial attack also be effective when targeting neural network policies in the context of reinforcement learning,which has inspired innovative researches in this direction.Hence,in this paper,we give the very first attempt to conduct a comprehensive survey on adversarial attacks in reinforcement learning under AI security.Moreover,we give briefly introduction on the most representative defense technologies against existing adversarial attacks. 展开更多
关键词 Reinforcement learning Artificial intelligence SECURITY Adversarial attack Adversarial example DEFENSE
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部