期刊文献+

基于卷积神经网络的视差图生成技术 被引量:10

Disparity map generation technology based on convolutional neural network
下载PDF
导出
摘要 针对裸眼三维中视差图生成过程中存在的高成本、长耗时以及容易出现背景空洞的问题,提出了一种基于卷积神经网络(CNN)学习预测的算法。首先通过对数据集的训练学习,掌握数据集中的变化规律;然后对输入卷积神经网络中的左视图进行特征提取和预测,得到深度值连续的深度图像;其次将预测所得到的每一个深度图和原图进行卷积,将生成的多个立体图像对进行叠加,最终形成右视图。仿真结果表明:该算法的像素重构尺寸误差相比基于水平视差的三维显示算法和深度图像视点绘制的算法降低了12.82%和10.52%,且背景空洞、背景粘连等问题都得到了明显改善。实验结果表明,卷积神经网络能提高视差图生成的图像质量。 Focusing on the issues such as high cost, long time consumption and background holes in the disparity map in naked-eye 3D applications, learning and prediction algorithm based on Convolutional Neural Network (CNN) was introduced. Firstly, the change rules of a dataset could be mastered through training and learning the dataset. Secondly, the depth map with continuous lasting depth value was attained by extracting and predicting the features of the left view in the input CNN. Finally, the right view was produced by the superposition of diverse stereo pairs after folding the predicted depth and original maps. The simulation results show that the pixel-wise reconstruction error of the proposed algorithm is 12.82% and 10.52% lower than that of 3D horizontal disparity algorithm and depth image-based rendering algorithm. In addition, the problems of background hole and background adhesion have been greatly improved. The experimental results show that CNN can improve the image quality of disparity maps.
出处 《计算机应用》 CSCD 北大核心 2018年第1期255-259,289,共6页 journal of Computer Applications
基金 装备学院校级基础研究项目(DXZT-JC-ZZ-2013-009)~~
关键词 裸眼三维 视差图 背景空洞 特征提取 卷积神经网络 naked-eye 3D disparity map background hole feature extraction Convolutional Neural Network (CNN)
  • 相关文献

参考文献3

二级参考文献229

  • 1HINTON G E, OSINDERO S, TEH Y W. A fast learning algorithm for deep belief nets [ J ]. Neural Computation, 2006,18 ( 7 ) : 1527- 1554.
  • 2BENGIO Y, LAMBLIN P, POPOVICI D, et al. Greedy layer-wise training of deep networks [ C ]//Advances in Neural Information Pro- cessing Systems. Cambridge:MIT Press,2007:153-160.
  • 3VINCENT P, LAROCHELLE H, BENBIO Y, et al. Extracting and composing robust features with denoising autoencoders [ C ]//Proc of the 25th International Conference on Machine [.earning. New York: ACM Press ,2008 : 1096-1103.
  • 4LAROCHELLE H, BENGIO Y, LOURADOUR J, et al. Exploring strategies for training deep neural networks[ J]. Journal of Machine Learning Research,2009,10 (12) : 1-40.
  • 5TAYLOR G, HINTON G E. Factored conditional restricted Bohzmann machines for modeling motion style [ C ]//Proc of the 26th Annual In- ternational Conference on Machine Learning. New York:ACM Press, 2009 : 1025-1032.
  • 6SALAKHUTDINOV R, HINTON G E. Deep Boltzmann machines [ C ]//Proe of the 12th International Conference on Artificial Intelli- gence and Statistics. 2009:448-455.
  • 7TAYLOR G, SIGAL L, FLEET D J, et al. Dynamical binary latent variable models for 3D human pose tracking[ C ]//Proe of IEEE Con- ferenee on Computer Vision and Pattern Recognition. 2010:631-638.
  • 8JARRETY K, KAVUKCUOGLU K, RANZATO M, et al. What is the best multi-stage architecture for object recognition? [ C ]//Pine of the 12th International Conference on Computer Vision. 2009:2146-2153.
  • 9LEE H, GROSSE R, RANGANATH R, et al. Convolutional deep belief networks for scalable unsupervised learning of hierarchical rep- resentations [ C ]//Proc of the 26th International Conference on Ma- chine Learning. New York : ACM Press ,2009:609-616.
  • 10LEE H, PHAM P, LARGMAN Y, et at. Unsupervised feature learn- ing for audio classification using convolutional deep belief networks [C ]//Advances in Neural Information Processing Systems. Cam- bridge : MIT Press ,2009 : 1096-1104.

共引文献1298

同被引文献66

引证文献10

二级引证文献40

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部