Learning from interaction is the primary way that biological agents acquire knowledge about their environment and themselves.Modern deep reinforcement learning(DRL)explores a computational approach to learning from in...Learning from interaction is the primary way that biological agents acquire knowledge about their environment and themselves.Modern deep reinforcement learning(DRL)explores a computational approach to learning from interaction and has made significant progress in solving various tasks.However,despite its power,DRL still falls short of biological agents in terms of energy efficiency.Although the underlying mechanisms are not fully understood,we believe that the integration of spiking communication between neurons and biologically-plausible synaptic plasticity plays a prominent role in achieving greater energy efficiency.Following this biological intuition,we optimized a spiking policy network(SPN)using a genetic algorithm as an energy-efficient alternative to DRL.Our SPN mimics the sensorimotor neuron pathway of insects and communicates through event-based spikes.Inspired by biological research showing that the brain forms memories by creating new synaptic connections and rewiring these connections based on new experiences,we tuned the synaptic connections instead of weights in the SPN to solve given tasks.Experimental results on several robotic control tasks demonstrate that our method can achieve the same level of performance as mainstream DRL methods while exhibiting significantly higher energy efficiency.展开更多
基金supported by the Beijing Nova Program,China(No.20230484369)the Strategic Priority Research Program of Chinese Academy of Sciences,China(No.XDA27010404)the Shanghai Municipal Science and Technology Major Project,China(No.2021SHZDZX),the Youth Innovation Promotion Association of the Chinese Academy of Sciences,China.
文摘Learning from interaction is the primary way that biological agents acquire knowledge about their environment and themselves.Modern deep reinforcement learning(DRL)explores a computational approach to learning from interaction and has made significant progress in solving various tasks.However,despite its power,DRL still falls short of biological agents in terms of energy efficiency.Although the underlying mechanisms are not fully understood,we believe that the integration of spiking communication between neurons and biologically-plausible synaptic plasticity plays a prominent role in achieving greater energy efficiency.Following this biological intuition,we optimized a spiking policy network(SPN)using a genetic algorithm as an energy-efficient alternative to DRL.Our SPN mimics the sensorimotor neuron pathway of insects and communicates through event-based spikes.Inspired by biological research showing that the brain forms memories by creating new synaptic connections and rewiring these connections based on new experiences,we tuned the synaptic connections instead of weights in the SPN to solve given tasks.Experimental results on several robotic control tasks demonstrate that our method can achieve the same level of performance as mainstream DRL methods while exhibiting significantly higher energy efficiency.