期刊文献+

深度神经网络压缩与加速综述 被引量:54

Deep Neural Network Compression and Acceleration:A Review
下载PDF
导出
摘要 深度神经网络在人工智能的应用中,包括计算机视觉、语音识别、自然语言处理方面,取得了巨大成功.但这些深度神经网络需要巨大的计算开销和内存存储,阻碍了在资源有限环境下的使用,如移动或嵌入式设备端.为解决此问题,在近年来产生大量关于深度神经网络压缩与加速的研究工作.对现有代表性的深度神经网络压缩与加速方法进行回顾与总结,这些方法包括了参数剪枝、参数共享、低秩分解、紧性滤波设计及知识蒸馏.具体地,将概述一些经典深度神经网络模型,详细描述深度神经网络压缩与加速方法,并强调这些方法的特性及优缺点.此外,总结了深度神经网络压缩与加速的评测方法及广泛使用的数据集,同时讨论分析一些代表性方法的性能表现.最后,根据不同任务的需要,讨论了如何选择不同的压缩与加速方法,并对压缩与加速方法未来发展趋势进行展望. In recent years, deep neural networks (D N Ns) have achieved remarkable success in manyartificial intelligence (AI ) applications, including computer vision, speech recognition and natural language processing. However, such DNNs have been accompanied by significant increase incomputational costs and storage services, which prohibits the usages of DNNs on resource-limited environments such as mobile or embedded devices. To this end, the studies of DNN compression and acceleration have recently become more emerging. In this paper, we provide a review on the existingrepresentative DNN compression and acceleration methods, including parameter pruning , parameter sharing, low-rank decomposition, compact filter designed, and knowledge distillation . Specifically, this paper provides an overview of DNNs , describes the details of dif ferent DNN compression and acceleration methods, and highlights the propert ies, advantages and drawbacks. Furthermore , we summarize the evaluation criteria and datasets widely used in DNN compression and acceleration, andalso discuss the performance of the representative methods. In the end, we discuss how to choosedifferent compression and acceleration methods to meet the needs of dif ferent tasks, and envisionfuture directions on this topic.
作者 纪荣嵘 林绍辉 晁飞 吴永坚 黄飞跃 Ji Rongrong;Lin Shaohui;Chao Fei;Wu Yongjian;Huang Feiyue(School of Information Science and Engineering,Xiamen University,Xiamen,Fujian,361005;Fujian Key Laboratory of Sensing and Computing for Smart City(Xiamen University),Xiamen.Fujian,361005;Bestlmage Laboratory,Tencent Technology(Shanghai)Co.,Lid,Shanghai,200233)
出处 《计算机研究与发展》 EI CSCD 北大核心 2018年第9期1871-1888,共18页 Journal of Computer Research and Development
基金 国家重点研发计划项目(2017YFC0113000 2016YFB10015032) 国家自然科学基金项目(U1705262 61772443 61402388 61572410) 国家自然科学基金优秀青年科学基金项目(61422210) 福建省自然科学基金项目(2017J01125)~~
关键词 深度神经网络压缩 深度神经网络加速 参数剪枝 参数共享 低秩分解 知识蒸馏 DNN compression DNN acceleration parameter pruning parameter sharing low-rank decomposition knowledge distillation
  • 相关文献

参考文献2

二级参考文献2

共引文献141

同被引文献261

引证文献54

二级引证文献204

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部