期刊文献+

Exploring compression and parallelization techniques for distribution of deep neural networks over Edge-Fog continuum-a review

原文传递
导出
摘要 Purpose-The trend of“Deep Learning for Internet of Things(IoT)”has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud as their resource giant.But this picture leads to underutilization of ever-increasing device pool of IoT that has already passed 15 billion mark in 2015.Thus,it is high time to explore a different approach to tackle this issue,keeping in view the characteristics and needs of the two fields.Processing at the Edge can boost applications with realtime deadlines while complementing security.Design/methodology/approach-This review paper contributes towards three cardinal directions of research in the field of DL for IoT.The first section covers the categories of IoT devices and how Fog can aid in overcoming the underutilization of millions of devices,forming the realm of the things for IoT.The second direction handles the issue of immense computational requirements of DL models by uncovering specific compression techniques.An appropriate combination of these techniques,including regularization,quantization,and pruning,can aid in building an effective compression pipeline for establishing DL models for IoT use-cases.The third direction incorporates both these views and introduces a novel approach of parallelization for setting up a distributed systems view of DL for IoT.Findings-DL models are growing deeper with every passing year.Well-coordinated distributed execution of such models using Fog displays a promising future for the IoT application realm.It is realized that a vertically partitioned compressed deep model can handle the trade-off between size,accuracy,communication overhead,bandwidth utilization,and latency but at the expense of an additionally considerable memory footprint.To reduce the memory budget,we propose to exploit Hashed Nets as potentially favorable candidates for distributed frameworks.However,the critical point between accuracy and size for such models needs further investigation.Originality/value-To the best of our knowledge,no study has explored the inherent parallelism in deep neural network architectures for their efficient distribution over the Edge-Fog continuum.Besides covering techniques and frameworks that have tried to bring inference to the Edge,the review uncovers significant issues and possible future directions for endorsing deep models as processing engines for real-time IoT.The study is directed to both researchers and industrialists to take on various applications to the Edge for better user experience.
出处 《International Journal of Intelligent Computing and Cybernetics》 EI 2020年第3期331-364,共34页 智能计算与控制论国际期刊(英文)
基金 supported by Technical Education Quality Improvement Programme(TEQIP-III) The project is implemented by NPIU which is a unit of MHRD,Govt of India for implementation of World Bank Assisted Projects in Technical Education。
  • 相关文献

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部