With the rapid development of machine learning,the demand for high-efficient computing becomes more and more urgent.To break the bottleneck of the traditional Von Neumann architecture,computing-in-memory(CIM)has attra...With the rapid development of machine learning,the demand for high-efficient computing becomes more and more urgent.To break the bottleneck of the traditional Von Neumann architecture,computing-in-memory(CIM)has attracted increasing attention in recent years.In this work,to provide a feasible CIM solution for the large-scale neural networks(NN)requiring continuous weight updating in online training,a flash-based computing-in-memory with high endurance(10^(9) cycles)and ultrafast programming speed is investigated.On the one hand,the proposed programming scheme of channel hot electron injection(CHEI)and hot hole injection(HHI)demonstrate high linearity,symmetric potentiation,and a depression process,which help to improve the training speed and accuracy.On the other hand,the low-damage programming scheme and memory window(MW)optimizations can suppress cell degradation effectively with improved computing accuracy.Even after 109 cycles,the leakage current(I_(off))of cells remains sub-10pA,ensuring the large-scale computing ability of memory.Further characterizations are done on read disturb to demonstrate its robust reliabilities.By processing CIFAR-10 tasks,it is evident that~90%accuracy can be achieved after 109 cycles in both ResNet50 and VGG16 NN.Our results suggest that flash-based CIM has great potential to overcome the limitations of traditional Von Neumann architectures and enable high-performance NN online training,which pave the way for further development of artificial intelligence(AI)accelerators.展开更多
The growth of data and Internet of Things challenges traditional hardware,which encounters efficiency and power issues owing to separate functional units for sensors,memory,and computation.In this study,we designed an...The growth of data and Internet of Things challenges traditional hardware,which encounters efficiency and power issues owing to separate functional units for sensors,memory,and computation.In this study,we designed an a-phase indium selenide(a-In_(2)Se_(3))transistor,which is a two-dimensional ferroelectric semiconductor as the channel material,to create artificial optic-neural and electro-neural synapses,enabling cutting-edge processing-in-sensor(PIS)and computing-in-memory(CIM)functionalities.As an optic-neural synapse for low-level sensory processing,the a-In_(2)Se_(3)transistor exhibits a high photoresponsivity(2855 A/W)and detectivity(2.91×10^(14)Jones),facilitating efficient feature extraction.For high-level processing tasks as an electro-neural synapse,it offers a fast program/erase speed of 40 ns/50μs and ultralow energy consumption of 0.37 aJ/spike.An AI vision system using a-In_(2)Se_(3)transistors has been demonstrated.It achieved an impressive recognition accuracy of 92.63%within 12 epochs owing to the synergistic combination of the PIS and CIM functionalities.This study demonstrates the potential of the a-In_(2)Se_(3)transistor in future vision hardware,enhancing processing,power efficiency,and AI applications.展开更多
基金This work was supported by the National Natural Science Foundation of China(Nos.62034006,92264201,and 91964105)the Natural Science Foundation of Shandong Province(Nos.ZR2020JQ28 and ZR2020KF016)the Program of Qilu Young Scholars of Shandong University.
文摘With the rapid development of machine learning,the demand for high-efficient computing becomes more and more urgent.To break the bottleneck of the traditional Von Neumann architecture,computing-in-memory(CIM)has attracted increasing attention in recent years.In this work,to provide a feasible CIM solution for the large-scale neural networks(NN)requiring continuous weight updating in online training,a flash-based computing-in-memory with high endurance(10^(9) cycles)and ultrafast programming speed is investigated.On the one hand,the proposed programming scheme of channel hot electron injection(CHEI)and hot hole injection(HHI)demonstrate high linearity,symmetric potentiation,and a depression process,which help to improve the training speed and accuracy.On the other hand,the low-damage programming scheme and memory window(MW)optimizations can suppress cell degradation effectively with improved computing accuracy.Even after 109 cycles,the leakage current(I_(off))of cells remains sub-10pA,ensuring the large-scale computing ability of memory.Further characterizations are done on read disturb to demonstrate its robust reliabilities.By processing CIFAR-10 tasks,it is evident that~90%accuracy can be achieved after 109 cycles in both ResNet50 and VGG16 NN.Our results suggest that flash-based CIM has great potential to overcome the limitations of traditional Von Neumann architectures and enable high-performance NN online training,which pave the way for further development of artificial intelligence(AI)accelerators.
基金supported by the National Natural Science Foundation of China(62104066,52221001,62090035,U19A2090,U22A20138,52372146,and 62101181)the National Key R&D Program of China(2022YFA1402501,2022YFA1204300)+6 种基金the Natural Science Foundation of Hunan Province(2021JJ20016)the Science and Technology Innovation Program of Hunan Province(2021RC3061)the Key Program of Science and Technology Department of Hunan Province(2019XK2001,2020XK2001)the Open Project Program of Wuhan National Laboratory for Optoelectronics(2020WNLOKF016)the Open Project Program of Key Laboratory of Nanodevices and Applications,Suzhou Institute of Nano-Tech and Nano-Bionics,Chinese Academy of Sciences(22ZS01)the Project funded by China Postdoctoral Science Foundation(2023TQ0110)the Innovation Project of Optics Valley Laboratory(OVL2023ZD002).
文摘The growth of data and Internet of Things challenges traditional hardware,which encounters efficiency and power issues owing to separate functional units for sensors,memory,and computation.In this study,we designed an a-phase indium selenide(a-In_(2)Se_(3))transistor,which is a two-dimensional ferroelectric semiconductor as the channel material,to create artificial optic-neural and electro-neural synapses,enabling cutting-edge processing-in-sensor(PIS)and computing-in-memory(CIM)functionalities.As an optic-neural synapse for low-level sensory processing,the a-In_(2)Se_(3)transistor exhibits a high photoresponsivity(2855 A/W)and detectivity(2.91×10^(14)Jones),facilitating efficient feature extraction.For high-level processing tasks as an electro-neural synapse,it offers a fast program/erase speed of 40 ns/50μs and ultralow energy consumption of 0.37 aJ/spike.An AI vision system using a-In_(2)Se_(3)transistors has been demonstrated.It achieved an impressive recognition accuracy of 92.63%within 12 epochs owing to the synergistic combination of the PIS and CIM functionalities.This study demonstrates the potential of the a-In_(2)Se_(3)transistor in future vision hardware,enhancing processing,power efficiency,and AI applications.