期刊文献+

Space Efficient Quantization for Deep Convolutional Neural Networks

原文传递
导出
摘要 Deep convolutional neural networks(DCNNs)have shown outstanding performance in the fields of computer vision,natural language processing,and complex system analysis.With the improvement of performance with deeper layers,DCNNs incur higher computational complexity and larger storage requirement,making it extremely difficult to deploy DCNNs on resource-limited embedded systems(such as mobile devices or Internet of Things devices).Network quantization efficiently reduces storage space required by DCNNs.However,the performance of DCNNs often drops rapidly as the quantization bit reduces.In this article,we propose a space efficient quantization scheme which uses eight or less bits to represent the original 32-bit weights.We adopt singular value decomposition(SVD)method to decrease the parameter size of fully-connected layers for further compression.Additionally,we propose a weight clipping method based on dynamic boundary to improve the performance when using lower precision.Experimental results demonstrate that our approach can achieve up to approximately 14x compression while preserving almost the same accuracy compared with the full-precision models.The proposed weight clipping method can also significantly improve the performance of DCNNs when lower precision is required.
出处 《Journal of Computer Science & Technology》 SCIE EI CSCD 2019年第2期305-317,共13页 计算机科学技术学报(英文版)
基金 the National Natural Science Foundation of China(NSFC)under Grant Nos.61772077 and 61370192 Beijing Natural Science Foundation of China under Grant No.4192051 NSFC under Grant Nos.61428203 and 61572347.
  • 相关文献

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部