期刊文献+

Technological Exploration of RRAM Crossbar Array for Matrix-Vector Multiplication 被引量:4

Technological Exploration of RRAM Crossbar Array for Matrix-Vector Multiplication
原文传递
导出
摘要 Matrix-vector multiplication is the key operation for many computationally intensive algorithms. The emerging metal oxide resistive switching random access memory (RRAM) device and RRAM crossbar array have demonstrated a promising hardware realization of the analog matrix-vector multiplication with ultra-high energy efficiency. In this paper, we analyze the impact of both device level and circuit level non-ideal factors, including the nonlinear current-voltage relationship of RRAM devices, the variation of device fabrication and write operation, and the interconnect resistance as well as other crossbar array parameters. On top of that, we propose a technological exploration flow for device parameter configuration to overcome the impact of non-ideal factors and achieve a better trade-off among performance, energy, and reliability for each specific application. Our simulation results of a support vector machine (SVM) and Mixed National Institute of Standards and Technology (MNIST) pattern recognition dataset show that RRAM crossbar array based SVM is robust to input signal fluctuation but sensitive to tunneling gap deviation. A further resistance resolution test presents that a 6-bit RRAM device is able to realize a recognition accuracy around 90%, indicating the physical feasibility of RRAM crossbar array based SVM. In addition, the proposed technological exploration flow is able to achieve 10.98% improvement of recognition accuracy on the MNIST dataset and 26.4% energy savings compared with previous work. Experimental results also show that more than 84.4% power saving can be achieved at the cost of little accuracy reduction. Matrix-vector multiplication is the key operation for many computationally intensive algorithms. The emerging metal oxide resistive switching random access memory (RRAM) device and RRAM crossbar array have demonstrated a promising hardware realization of the analog matrix-vector multiplication with ultra-high energy efficiency. In this paper, we analyze the impact of both device level and circuit level non-ideal factors, including the nonlinear current-voltage relationship of RRAM devices, the variation of device fabrication and write operation, and the interconnect resistance as well as other crossbar array parameters. On top of that, we propose a technological exploration flow for device parameter configuration to overcome the impact of non-ideal factors and achieve a better trade-off among performance, energy, and reliability for each specific application. Our simulation results of a support vector machine (SVM) and Mixed National Institute of Standards and Technology (MNIST) pattern recognition dataset show that RRAM crossbar array based SVM is robust to input signal fluctuation but sensitive to tunneling gap deviation. A further resistance resolution test presents that a 6-bit RRAM device is able to realize a recognition accuracy around 90%, indicating the physical feasibility of RRAM crossbar array based SVM. In addition, the proposed technological exploration flow is able to achieve 10.98% improvement of recognition accuracy on the MNIST dataset and 26.4% energy savings compared with previous work. Experimental results also show that more than 84.4% power saving can be achieved at the cost of little accuracy reduction.
出处 《Journal of Computer Science & Technology》 SCIE EI CSCD 2016年第1期3-19,共17页 计算机科学技术学报(英文版)
基金 This work was supported by the National Basic Research 973 Program of China under Grant No. 2013CB329000, the National Natural Science Foundation of China under Grant Nos. 61373026, 61261160501, the Brain Inspired Computing Research of Tsinghua University under Grant No. 20141080934, Tsinghua University Initiative Scientific Research Program, and the Importation and Development of High-Caliber Talents Project of Beijing Municipal Institutions.
关键词 resistive switching random access memory (RRAM) machine learning electronic design automation matrixvector multiplication non-ideal factor resistive switching random access memory (RRAM), machine learning, electronic design automation, matrixvector multiplication, non-ideal factor
  • 相关文献

参考文献46

  • 1Franklin J. The elements of statistical learning: Data min- ing, inference and prediction. The Mathematical Intelli- gencer, 2005, 27(2): 83-85.
  • 2Jang J W, Choi S B, Prasanna V K. Energy- and time- efficient matrix multiplication on FPGAs. IEEE Trans- actions on Very Large Scale Integration (VLSI) Systems, 2005, 13(11): 1305-1319.
  • 3Williams S, Oliker L, Vuduc R et al. Optimization of sparse matrix-vector multiplication on emerging multieore plat- forms. Parallel Computing, 2009, 35(3): 178-194.
  • 4Catanzaro B, Sundaram N, Keutzer K. Fast support vec- tor machine training and classification on graphics proces- sors. In Proe, the 25th International Conference on Ma- chine Learning, July 2008, pp.104-111.
  • 5Dean J, Corrado G, Monga R et al. Large scale distributed deep networks. In Advances in Neural Information Process- ing Systems 25, Pereira F, Burges C, Bottou L, Weinberger K (eds.), Curran Associates, Inc., 2012, pp.1232-1241.
  • 6Xu C, Dong X, Jouppi N Pet al. Design implications of memristor-based RRAM cross-point structures. In Proc. Design, Automation and Test in Europe Conference and Exhibition (DATE), March 2011.
  • 7Wang Y, Li B, Luo R et al. Energy efficient neural net- works for big data analytics. In Proc. Design, Automation and Test in Europe Conference and Exhibition (DATE), March 2014.
  • 8Hu M, Li H, Wu Q et al. Hardware realization of BSB re- call function using memristor crossbar arrays. In Proc. the49th Annual Design Automation Conference, June 2012, pp.498-503.
  • 9Li B, Shan Y, Hu M et al. Memristor-based approximated computation. In Proc. the International Symposium on Low Power Electronics and Design, September 2013, pp.242- 247.
  • 10Hu M, Li H, Chen Yet al. Memristor crossbar-based neu- romorphic computing system: A case study. IEEE Trans- actions on Neural Networks and Learning Systems, 2014, 25(10): 1864-1878.

同被引文献2

引证文献4

二级引证文献12

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部