期刊文献+

大数据分析的无限深度神经网络方法 被引量:77

Big Data Analysis by Infinite Deep Neural Networks
下载PDF
导出
摘要 深度神经网络(deep neural networks,DNNs)及其学习算法,作为成功的大数据分析方法,已为学术界和工业界所熟知.与传统方法相比,深度学习方法以数据驱动、能自动地从数据中提取特征(知识),对于分析非结构化、模式不明多变、跨领域的大数据具有显著优势.目前,在大数据分析中使用的深度神经网络主要是前馈神经网络(feedforward neural networks,FNNs),这种网络擅长提取静态数据的相关关系,适用于基于分类的数据应用场景.但是受到自身结构本质的限制,它提取数据时序特征的能力有限.无限深度神经网络(infinite deep neural networks)是一种具有反馈连接的回复式神经网络(recurrent neural networks,RNNs),本质上是一个动力学系统,网络状态随时间演化是这种网络的本质属性,它耦合了"时间参数",更加适用于提取数据的时序特征,从而进行大数据的预测.将这种网络的反馈结构在时间维度展开,随着时间的运行,这种网络可以"无限深",故称之为无限深度神经网络.重点介绍这种网络的拓扑结构和若干学习算法及其在语音识别和图像理解领域的成功实例. Deep neural networks (DNNs) and their learning algorithms are well known in the academic community and industry as the most successful methods for big data analysis. Compared with traditional methods, deep learning methods use data-driven and can extract features (knowledge) automatically from data. Deep learning methods have significant advantages in analyzing unstructured, unknown and varied model and cross field big data. At present, the most widely used deep neural networks in big data analysis are feedforward neural networks (FNNs). They work well in extracting the correlation from static data and suiting for data application scenarios based on classification. But limited by its intrinsic structure, the ability of feedforward neural networks to extract time sequence features is weak. Infinite deep neural networks, i.e. recurrent neural networks (RNNs) are dynamical systems essentially. Their essential character is that the states of the networks change with time and couple the time parameter. Hence they are very suit for extracting time sequence features. It means that infinite deep neural networks can perform the prediction of big data. If extending recurrent structure of recurrent neural networks in the time dimension, the depth of networks can be infinite with time running, so they are called infinite deep neural networks. In thispaper, we focus on the topology and some learning algorithms of intinite deep neural networks, and introduce some successful applications in speech recognition and image understanding.
作者 张蕾 章毅
出处 《计算机研究与发展》 EI CSCD 北大核心 2016年第1期68-79,共12页 Journal of Computer Research and Development
基金 国家自然科学基金项目(61322203 61332002 61432012)~~
关键词 深度神经网络 无限深度神经网络 前馈神经网络 回复式神经网络 大数据 deep neural networks (DNNs) infinite deep neural networks feedforward neuralnetworks (FNNs) recurrent neural networks (RNNs) big data
  • 相关文献

参考文献43

  • 1工信部电信研究院大数据白皮书(2014)[M].北京:工业和信息化部电信研究院,2014.
  • 2Intel IT Center.大数据101:非结构化数据分析[M/OL].[2015-07-01 ]. http://www. intel. cn/content/dam/www/public/cn/zh/pdfvs/Big-data-101-Unstructured-data-Analytics.pdf, 2012.
  • 3Hinton G E, Salakhutdinov R R. Reducing the dimensionalityof data with neural networks [J], Science, 2006, 313(5786): 504-507.
  • 4余凯,贾磊,陈雨强,徐伟.深度学习的昨天、今天和明天[J].计算机研究与发展,2013,50(9):1799-1804. 被引量:590
  • 5Rosenblatt F. The perceptron: A probabilistic model forinformation storage and organization in the brain. [J],Psychological Review, 1958,65(6) : 386-408.
  • 6Rumelhart D E,Hinton G E, Williams R J. Learningrepresentations by back-propagating errors [J]. Nature,1986, 323(6088): 533-536.
  • 7Huang G B, Chen L,Siew C K. Universal approximationusing incremental constructive feedforward networks withrandom hidden nodes [J]. IEEE Trans on Neural Networks,2006, 17(4): 879-892.
  • 8Vincent P,Larochelle H, Lajoie I,et al. Stacked denoisingautoencoders : Learning useful representations in a deepnetwork with a local denoising criterion [J]. Journal ofMachine Learning Research, 2010,11: 3371-3408.
  • 9LeCun Y, Bottou L, Bengio Y. Gradient-based learningapplied to document recognition [J]. Proceedings of theIEEE, 1998, 86(11): 2278-2324.
  • 10Williams R J,Zipser D. Gradient-based learning algorithmsfor recurrent connectionist networks, NU-CCS-90-9 [R].Boston,MA: Northeastern University, 1990.

二级参考文献14

  • 1MarkoffJ. How many computers to identify a cat?[NJ The New York Times, 2012-06-25.
  • 2MarkoffJ. Scientists see promise in deep-learning programs[NJ. The New York Times, 2012-11-23.
  • 3李彦宏.2012百度年会主题报告:相信技术的力量[R].北京:百度,2013.
  • 410 Breakthrough Technologies 2013[N]. MIT Technology Review, 2013-04-23.
  • 5Rumelhart D, Hinton G, Williams R. Learning representations by back-propagating errors[J]. Nature. 1986, 323(6088): 533-536.
  • 6Hinton G, Salakhutdinov R. Reducing the dimensionality of data with neural networks[J]. Science. 2006, 313(504). Doi: 10. 1l26/science. 1127647.
  • 7Dahl G. Yu Dong, Deng u, et a1. Context-dependent pre?trained deep neural networks for large vocabulary speech recognition[J]. IEEE Trans on Audio, Speech, and Language Processing. 2012, 20 (1): 30-42.
  • 8Jaitly N. Nguyen P, Nguyen A, et a1. Application of pretrained deep neural networks to large vocabulary speech recognition[CJ //Proc of Interspeech , Grenoble, France: International Speech Communication Association, 2012.
  • 9LeCun y, Boser B, DenkerJ S. et a1. Backpropagation applied to handwritten zip code recognition[J]. Neural Computation, 1989, I: 541-551.
  • 10Large Scale Visual Recognition Challenge 2012 (ILSVRC2012)[OLJ.[2013-08-01J. http://www. image?net.org/challenges/LSVRC/2012/.

共引文献589

同被引文献642

引证文献77

二级引证文献667

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部