期刊文献+

深度网络模型压缩综述 被引量:42

Survey of Deep Neural Network Model Compression
下载PDF
导出
摘要 深度网络近年来在计算机视觉任务上不断刷新传统模型的性能,已逐渐成为研究热点.深度模型尽管性能强大,然而由于参数数量庞大、存储和计算代价高,依然难以部署在受限的硬件平台上(如移动设备).模型的参数在一定程度上能够表达其复杂性,相关研究表明,并不是所有的参数都在模型中发挥作用,部分参数作用有限、表达冗余,甚至会降低模型的性能.首先,对国内外学者在深度模型压缩上取得的成果进行了分类整理,依此归纳了基于网络剪枝、网络精馏和网络分解的方法;随后,总结了相关方法在多种公开深度模型上的压缩效果;最后,对未来的研究可能的方向和挑战进行了展望. Deep neural networks have continually surpassed traditional methods on a variety of computer vision tasks. Though deep neural networks are very powerful, the large number of weights consumes considerable storage and calculation time, making it hard to deploy on resource-constrained hardware platforms such as mobile system. The number of weights in deep neural networks represents the complexity to an extent, but not all the weights contribute to the performance according to recent researches. Specifically, some weights are redundant and even decrease the performance. This survey offers a systematic summarization of existing research achievements of the domestic and foreign researchers in recent years in the aspects of network pruning, network distillation, and network decomposition. Furthermore, comparisons of compression performance are provided on several public deep neural networks. Finally, a perspective of future work and challenges in this research area are discussed.
出处 《软件学报》 EI CSCD 北大核心 2018年第2期251-266,共16页 Journal of Software
基金 国家自然科学基金(61572428 U1509206)~~
关键词 深度神经网络 网络压缩 网络剪枝 网络精馏 网络分解 deep neural network network compression network pruning network distillation network decomposition
  • 相关文献

参考文献3

二级参考文献45

  • 1闫雷鸣,孙志挥.一种基于二次互信息的双聚类算法[J].计算机工程与应用,2006,42(22):158-160. 被引量:4
  • 2Quinlan J R. Comparing connectionist and symbolic learning methods. In Computational Learning Theory and Natural Learning Systems, Rivest R L (Ed.), Vol.1,Cambridge, MA, MIT Press, 1994, pp.445-456.
  • 3Chalup S, Hayward R, Diederich J. Rule extraction from artificial neural networks trained on elementary number classification tasks. In Proc. the 9th Australian Conference on Neural Networks, Brisbane, Australia, 1998,pp.265-270.
  • 4Maire F. Rule-extraction by backpropagation of polyhedra. Neural Networks, 1999, 12(4-5): 717-725.
  • 5Bologna G. Rule extraction from a multi layer perceptron with staircase activation functions. In Proc. the IEEE-INNS-ENNS International Joint Conference on Neural Networks, Como, Italy, 2000, 3: 419-424.
  • 6Vahed A, Omlin C W. Rule extraction from recurrent neural networks using a symbolic machine learning algorithm. In Proc. the 6th International Conference on Neural Information Processing, Dunedin, New Zealand,1999, pp.712-717.
  • 7Golea M. On the complexity of rule extraction from neural networks and network querying. In Proc. theAISB'96 Workshop on Rule Eztraction from TrainedNeural Networks, Brighton, UK, 1996, pp.51-59.
  • 8Roy A. On connectionism, rule extraction, and brainlike learning. IEEE Trans. Fuzzy Systems, 2000, 8(2):222-227.
  • 9Duch W, Adamczak R, Grabczewski K. A new methodology of extraction, optimization and application of crisp and fuzzy logical rules. IEEE Trans. Neural Networks, 2001, 12(2): 277-306.
  • 10Zhou Z H, Jiang Y, Chen S F, Extracting symbolic rules from trained neural network ensembles. AI Communications, 2003, 16(1): 3-15.

共引文献35

同被引文献155

引证文献42

二级引证文献329

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部