摘要
本文提出多层前馈网络的一种新的学习和综合算法──FP算法,并证明由此算法得到的网络作为通用联想记忆器时,具有如下优点:(1)每个样本都是吸引中心;(2)每个样本的吸引半径达到最大值;(3)网络没有假吸引中心;(4)网络具有最少的元件个数;(5)学习的复杂性达到最优(就其复杂性的阶而言).故此网络在性能、结构、计算复杂性等方面均达到很好状态.
A new learning algorithm-forward propagation (FP) of multilayered feed-forward neural networks is presented in this paper. The authors show that as an associative memory the network constructed by the FP algorithm has several advantages. (1)Each training sample is an attractive center. (2) The attractive radius of each training sample reaches the maximum. (3) There is no spurious attractive center in the network.(4) The network has minimal number of elements. (5) The order of its learning complexity is optimal. The FP learning algorithm is also an effective synthesis tool, i. e., the network architecture can be constructed during its learning process.
出处
《软件学报》
EI
CSCD
北大核心
1995年第7期440-448,共9页
Journal of Software
基金
攀登计划
863高技术计划的资助
关键词
神经网络
多层前馈网络
FP算法
Neural network, multilayered feed forward network, attractive center, attractive radius, learning algorithm.