期刊文献+
共找到1篇文章
< 1 >
每页显示 20 50 100
Hierarchical state-abstracted and socially augmented Q-Learning for reducing complexity in agent-based learning 被引量:2
1
作者 laura ray 《控制理论与应用(英文版)》 EI 2011年第3期440-450,共11页
A primary challenge of agent-based policy learning in complex and uncertain environments is escalating computational complexity with the size of the task space(action choices and world states) and the number of agents... A primary challenge of agent-based policy learning in complex and uncertain environments is escalating computational complexity with the size of the task space(action choices and world states) and the number of agents.Nonetheless,there is ample evidence in the natural world that high-functioning social mammals learn to solve complex problems with ease,both individually and cooperatively.This ability to solve computationally intractable problems stems from both brain circuits for hierarchical representation of state and action spaces and learned policies as well as constraints imposed by social cognition.Using biologically derived mechanisms for state representation and mammalian social intelligence,we constrain state-action choices in reinforcement learning in order to improve learning efficiency.Analysis results bound the reduction in computational complexity due to stateion,hierarchical representation,and socially constrained action selection in agent-based learning problems that can be described as variants of Markov decision processes.Investigation of two task domains,single-robot herding and multirobot foraging,shows that theoretical bounds hold and that acceptable policies emerge,which reduce task completion time,computational cost,and/or memory resources compared to learning without hierarchical representations and with no social knowledge. 展开更多
关键词 Decentralized Markov decision process Reinforcement learning Multiagent systems
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部