期刊文献+

基于BP网络的权值更新快速收敛算法 被引量:6

Rapid convergence algorithms for weight values updating based on BP network
下载PDF
导出
摘要 针对标准BP网络学习算法收敛慢的问题,提出了两种权值更新的快速收敛算法,即基于梯度变化率的快速传递算法和基于梯度方向的弹性传递算法,并在煤矿事故救援游戏式训练系统中进行仿真和比较,让游戏角色根据井下空气成分学习判断危险程度,以便受训人员或仿生机器人采取相应的措施。仿真结果表明,所提算法的收敛时间比标准算法有一定改善。 To solve the slow convergence of standard learning algorithm in BP network, two rapid convergence algorithms were suggested for weight values updating. One is rapid transmission algorithm based on gradient change rate. The other is flexible transmission algorithm based on gradient orientation. The two algorithms were simulated and compared in Game Style Training System for Mine Accident Rescuing. Here the algorithms would help game roles learn to estimate the danger degree according to ingredients of mine air, and then help trainees or biorobots take corresponding actions. The simulating results show that shorter convergence time is taken for the two algorithms than the standard algorithm.
出处 《计算机应用》 CSCD 北大核心 2006年第8期1940-1942,共3页 journal of Computer Applications
基金 山西省自然科学基金资助项目(20041043) 山西省留学回国人员科研资助项目(200336)
关键词 快速收敛算法 游戏式训练 BP人工神经网络 quick convergence algorithm game style training BP artificial neural network
  • 相关文献

参考文献9

  • 1余雪丽,孙承意,冯秀芳,等.神经网络与学习实例[M].北京:中国铁道出版社,1996.
  • 2[美]CHAMPANDARD AJ.Artificial Intelligence from Theory to Fun![EB/OL].http://AiGameDev.com/,2006.
  • 3[美]ROLLINGS A,ADAMS E.游戏设计技术[M].金名,张长富,译.北京:北京希望电子出版社,2004.
  • 4MADEIRA C,et al.Bootstrapping the Learning process for the semiautomated Design of a challenging Game AI[A].Proceeding of AAAI04[C].2004.
  • 5MENDEZ G,HERRERO P,DE ANTONIO A.Intelligent virtual environments for training in nuclear power plants[A].Proceeding of the 6th International Conference on Enterprise Information Systems(ICES 2004)[C].Proto,Portugal,2004.
  • 6MANTOVANI F,GIANLUCA C,GAGGIOLI A,et al.Virtual Reality Training for Health-Care Professionals[J].Cyberpsychology &Behavior,2003,6(4).
  • 7YU XL.Studying on Granularity of Reinforcement Learning Agents[A].IFPS Proceeding of ICAMT/2002[C].2002.
  • 8杨源杰,黄道.人工神经网络算法研究及应用[J].华东理工大学学报(自然科学版),2002,28(5):551-554. 被引量:28
  • 9徐昕,贺汉根.神经网络增强学习的梯度算法研究[J].计算机学报,2003,26(2):227-233. 被引量:21

二级参考文献21

  • 1Mills P M, Zomaya A Y, Tade M O. Adapative model-based control using neural networks control[J]. 1994, 60:32-35.
  • 2Frier A, Karlton P, Kocher P. The SSL 3.0 Protocol[M]. USA:Netscape, 1996.
  • 3Rivest R. RFC 1321, The MD5 Message Digest Algorithm[S].
  • 4George Apostolopoulos, Vincd Peris, Prashant Pradhan, et al. Securing electronic commerce reducing the SSL overhead[J]. IEEE Network, 2000,(7-8):8-16.
  • 5Baird L C. Residual algorithms: Reinforcement learning with function approximation. In: Proceedings of the 12th International Conference on Machine Learning (ICML95), Tahoe City, California, USA, 1995. 30~37
  • 6Rumelhart D E et al. Learning internal representations by error propagation. In: Rumelhart D E et al, eds. Parallel Distributed Processing: Explorations in the Microstructure of Cognition, Vol.1,Cambridge, MA: MIT Press,1986. 318~362
  • 7Cybenko G. Approximation by superpositions of a sigmoidal function. Mathematics of Control, Signals, and Systems, 1989, 2: 303~314
  • 8Baird L C, Moore A. Gradient descent for general reinforcement learning. In: Kearns M S, Solla S A, Cohn D A eds. Advances in Neural Information Processing Systems 11, Cambrige, MA: MIT Press, 1999. 968~974
  • 9Bertsekas D P, Tsitsiklis J N. Gradient convergence in gradient methods with errors. SIAM Journal on Optimization, 2000, 10(3): 627~642
  • 10Heger M. The loss from imperfect value functions in expectation-based and minimax-based tasks. Machine Learning, 1996, 22(1): 197~225

共引文献47

同被引文献48

引证文献6

二级引证文献41

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部