Nowadays,deep neural networks(DNNs)have been equipped with powerful representation capabilities.The deep convolutional neural networks(CNNs)that draw inspiration from the visual processing mechanism of the primate ear...Nowadays,deep neural networks(DNNs)have been equipped with powerful representation capabilities.The deep convolutional neural networks(CNNs)that draw inspiration from the visual processing mechanism of the primate early visual cortex have outperformed humans on object categorization and have been found to possess many brain-like properties.Recently,vision transformers(ViTs)have been striking paradigms of DNNs and have achieved remarkable improvements on many vision tasks compared to CNNs.It is natural to ask how the brain-like properties of ViTs are.Beyond the model paradigm,we are also interested in the effects of factors,such as model size,multimodality,and temporality,on the ability of networks to model the human visual pathway,especially when considering that existing research has been limited to CNNs.In this paper,we systematically evaluate the brain-like properties of 30 kinds of computer vision models varying from CNNs and ViTs to their hybrids from the perspective of explaining brain activities of the human visual cortex triggered by dynamic stimuli.Experiments on two neural datasets demonstrate that neither CNN nor transformer is the optimal model paradigm for modelling the human visual pathway.ViTs reveal hierarchical correspondences to the visual pathway as CNNs do.Moreover,we find that multi-modal and temporal networks can better explain the neural activities of large parts of the visual cortex,whereas a larger model size is not a sufficient condition for bridging the gap between human vision and artificial networks.Our study sheds light on the design principles for more brain-like networks.The code is available at https://github.com/QYiZhou/LWNeuralEncoding.展开更多
Expanding wearable technologies to artificial tactile perception will be of significance for intelligent human-machine interface,as neuromorphic sensing devices are promising candidates due to their low energy consump...Expanding wearable technologies to artificial tactile perception will be of significance for intelligent human-machine interface,as neuromorphic sensing devices are promising candidates due to their low energy consumption and highly effective operating properties.Skin-compatible and conformable features are required for the purpose of realizing wearable artificial tactile perception.Here,we report an intrinsically stretchable,skin-integrated neuromorphic system with triboelectric nanogenerators as tactile sensing and organic electrochemical transistors as information processing.The integrated system provides desired sensing,synaptic,and mechanical characteristics,such as sensitive response(~0.04 kPa^(-1))to low-pressure,short-and long-term synaptic plasticity,great switching endurance(>10000 pulses),symmetric weight update,together with high stretchability of 100%strain.With neural encoding,demonstrations are capable of recognizing,extracting,and encoding features of tactile information.This work provides a feasible approach to wearable,skin-conformable neuromorphic sensing system with great application prospects in intelligent robotics and replacement prosthetics.展开更多
Cognitive neuroscience aims to develop computational models that can accurately predict and explain neural responses to sensory inputs in the cortex. Recent studies attempt to leverage the representation power of deep...Cognitive neuroscience aims to develop computational models that can accurately predict and explain neural responses to sensory inputs in the cortex. Recent studies attempt to leverage the representation power of deep neural networks(DNNs) to predict the brain response and suggest a correspondence between artificial and biological neural networks in their feature representations. However, typical voxel-wise encoding models tend to rely on specific networks designed for computer vision tasks, leading to suboptimal brain-wide correspondence during cognitive tasks. To address this challenge, this work proposes a novel approach that upgrades voxel-wise encoding models through multi-level integration of features from DNNs and information from brain networks. Our approach combines DNN feature-level ensemble learning and brain atlas-level model integration, resulting in significant improvements in predicting whole-brain neural activity during naturalistic video perception. Furthermore, this multi-level integration framework enables a deeper understanding of the brain's neural representation mechanism, accurately predicting the neural response to complex visual concepts. We demonstrate that neural encoding models can be optimized by leveraging a framework that integrates both data-driven approaches and theoretical insights into the functional structure of the cortical networks.展开更多
基金supported by National Natural Science Foundation of China(Nos.61976209 and 62020106015)the CAS International Collaboration Key Project,China(No.173211KYSB20190024)the Strategic Priority Research Program of CAS,China(No.XDB32040000)。
文摘Nowadays,deep neural networks(DNNs)have been equipped with powerful representation capabilities.The deep convolutional neural networks(CNNs)that draw inspiration from the visual processing mechanism of the primate early visual cortex have outperformed humans on object categorization and have been found to possess many brain-like properties.Recently,vision transformers(ViTs)have been striking paradigms of DNNs and have achieved remarkable improvements on many vision tasks compared to CNNs.It is natural to ask how the brain-like properties of ViTs are.Beyond the model paradigm,we are also interested in the effects of factors,such as model size,multimodality,and temporality,on the ability of networks to model the human visual pathway,especially when considering that existing research has been limited to CNNs.In this paper,we systematically evaluate the brain-like properties of 30 kinds of computer vision models varying from CNNs and ViTs to their hybrids from the perspective of explaining brain activities of the human visual cortex triggered by dynamic stimuli.Experiments on two neural datasets demonstrate that neither CNN nor transformer is the optimal model paradigm for modelling the human visual pathway.ViTs reveal hierarchical correspondences to the visual pathway as CNNs do.Moreover,we find that multi-modal and temporal networks can better explain the neural activities of large parts of the visual cortex,whereas a larger model size is not a sufficient condition for bridging the gap between human vision and artificial networks.Our study sheds light on the design principles for more brain-like networks.The code is available at https://github.com/QYiZhou/LWNeuralEncoding.
基金The Foundation of National Natural Science Foundation of China,Grant/Award Number:61421002City University of Hong Kong,Grant/Award Numbers:9678274,9667221,9680322+5 种基金Research Grants Council of Hong Kong Special Administrative Region,Grant/Award Numbers:21210820,11213721,11215722Regional Joint Fund of the National Science Foundation of China,Grant/Award Number:U21A20492The Sichuan Science and Technology Program,Grant/Award Numbers:2022YFH0081,2022YFG0012,2022YFG0013The Sichuan Province Key Laboratory of Display Science and TechnologyInnoHK Project on Project 2.2—AI-based 3D ultrasound imaging algorithm at Hong Kong Centre for Cerebro-Cardiovascular Health Engineering(COCHE)RGC Senior Research Fellow Scheme,Grant/Award Number:SRFS2122-5S04.
文摘Expanding wearable technologies to artificial tactile perception will be of significance for intelligent human-machine interface,as neuromorphic sensing devices are promising candidates due to their low energy consumption and highly effective operating properties.Skin-compatible and conformable features are required for the purpose of realizing wearable artificial tactile perception.Here,we report an intrinsically stretchable,skin-integrated neuromorphic system with triboelectric nanogenerators as tactile sensing and organic electrochemical transistors as information processing.The integrated system provides desired sensing,synaptic,and mechanical characteristics,such as sensitive response(~0.04 kPa^(-1))to low-pressure,short-and long-term synaptic plasticity,great switching endurance(>10000 pulses),symmetric weight update,together with high stretchability of 100%strain.With neural encoding,demonstrations are capable of recognizing,extracting,and encoding features of tactile information.This work provides a feasible approach to wearable,skin-conformable neuromorphic sensing system with great application prospects in intelligent robotics and replacement prosthetics.
基金supported by the National Natural Science Foundation of China (62236009,61876032,32371154)Shenzhen Science and Technology Program (JCYJ20210324140807019)Shanghai Pujiang Program (22PJ1410500)。
文摘Cognitive neuroscience aims to develop computational models that can accurately predict and explain neural responses to sensory inputs in the cortex. Recent studies attempt to leverage the representation power of deep neural networks(DNNs) to predict the brain response and suggest a correspondence between artificial and biological neural networks in their feature representations. However, typical voxel-wise encoding models tend to rely on specific networks designed for computer vision tasks, leading to suboptimal brain-wide correspondence during cognitive tasks. To address this challenge, this work proposes a novel approach that upgrades voxel-wise encoding models through multi-level integration of features from DNNs and information from brain networks. Our approach combines DNN feature-level ensemble learning and brain atlas-level model integration, resulting in significant improvements in predicting whole-brain neural activity during naturalistic video perception. Furthermore, this multi-level integration framework enables a deeper understanding of the brain's neural representation mechanism, accurately predicting the neural response to complex visual concepts. We demonstrate that neural encoding models can be optimized by leveraging a framework that integrates both data-driven approaches and theoretical insights into the functional structure of the cortical networks.