期刊文献+

高斯核函数选择的广义核极化准则 被引量:9

Generalized Kernel Polarization Criterion for Optimizing Gaussian Kernel
下载PDF
导出
摘要 核函数及其参数的选择是核方法研究中的一个基本却很困难的问题,高斯核是目前各类核方法中最常使用的一种核函数.关于高斯核参数的优化已有很多研究,然而这些方法大多存在时间复杂度高,或是算法实现困难,或是样本数据需服从多元正态分布的前提假设等不足.提出的广义核极化准则可用来解决分类问题中的高斯核参数优化,该准则通过保持类内局部结构信息及中心化核矩阵以更准确地刻画特征空间中类别间的分离度,进而获得更好的高斯核参数来提高分类性能.给出了广义核极化准则对应目标函数的近似最优解的存在唯一性证明,且由于该准则独立于学习算法,因此可用许多成熟的优化算法来寻找最优参数.此外,还补充了已有文献提出的局部核极化准则对应目标函数近似最优解的存在唯一性证明,并且指出该准则是所提出的广义核极化准则的一个特例.针对多分类问题,分别给出广义核极化准则及局部核极化准则的多分类拓展形式.在标准数据集上的实验结果表明所提准则的有效性. The choice of kernel function is a basic and challenging problem in researches on kernel methods. Gaussian kernel is a popular and widely used one in various kernel methods, and many universal kernel selection methods have been derived for Gaussian kernel. However, these methods may have some disadvantages, such as heavy computational complexity, the difficulty of algorithm implement, and the requirement of the classes generated from underlying multivariate normal distributions. To remedy these problems, generalized kernel polarization criterion has been proposed to tune the parameter of Gaussian kernel for classification tasks. By taking the within-class local structure into account and centering the kernel matrix, the criterion does better in maximizing the class separability in the feature space. And the final optimized kernel parameter leads to a substantial improvement in the performance. Furthermore, the criterion function can be proved to have a determined approximate global minimum point. This good characteristic, coupled with its independence of the actual learning machine, makes the optimal parameter easier to find by many algorithms. Besides this, local kernel polarization criterion function, a special case of generalized kernel polarization criterion function, can also be proved to have a determined approximate global minimum point. The extensions of generalized kernel polarization criterion and local kernel polarization criterion to the multiclass domain have been proposed. Experimental results show the effectiveness and efficiency of our proposed criteria.
作者 田萌 王文剑
出处 《计算机研究与发展》 EI CSCD 北大核心 2015年第8期1722-1734,共13页 Journal of Computer Research and Development
基金 国家自然科学基金项目(61273291) 山西省回国留学人员科研资助项目(2012-008)
关键词 核方法 核选择 分类 核极化准则 广义核极化准则 kernel method kernel selection classification kernel polarization criterion generalizedkernel polarization criterion
  • 相关文献

参考文献28

  • 1Shawe-Taylor J, Crislianini N. Kernel Methods for Pattern Analysis [M]. Cambridge, USA: Cambridge University Press, 200,1:25-46.
  • 2Vapnik V N. "Fhe Nature of Statistical l,earning Theory [M]. Berlin: Springer, 1998: 493-520.
  • 3Xiong Huilin, Swamy M N S, Ahmad M O. Optimizing the kernel in the empirical feature space [J].IEEE Trans on Neural Networks, 2005, 16(2) : .460-474.
  • 4Daoud E A, Turabieh H. New empirical nonparametric kernels for support vector machine e ass f cat [J]. Applied Soil Computing, 2013, 13(4): 1759-1765.
  • 5Tsuda K, Kawanabe M. R3itsch G, et al. A discriminative kernel front prohabilistic models [J]. Neural Computation, 2002, 14(10) : 2397-2414.
  • 6廖士中,贾磊.一类新的球面核函数的构造[J].计算机研究与发展,2007,44(z2):398-402. 被引量:4
  • 7Chapelle O, Vapnik V N, Bousquei O, et al. Choosing multiple parameters for support vector machines [J]. Machine I.earning. 2002, 46(1): 131-159.
  • 8Joachirns T. Estimating the generalization performance of a SVM efficiently [C] //Proe of the 17th Int Conf on Machine Learning. San Francisco: Morgan Kaufmann, 2000:431-4:38.
  • 9Wang Lei, Xue Ping, Chan K L. Two criteria for model selection in multiclass support vector machines [J]. IEEE Trans on Systems, Man, and Cybernetics --Part B= Cybernetics, 2008, 38(6): 1432-1448.
  • 10Wu Kuo Ping, Wang Sheng-de. Choosing the kernel parameters for support vector machine by the inter-cluster distance in the feature space[J]. Pattern Recognition, 2009, 42(5) : 710-717.

二级参考文献17

  • 1[3]J C Burges.A tutorial on support vector machine for pattern recognition.Data Mining and Knowledge Discovery,1998,2(2):121-167
  • 2[5]S Amari,S Wu.Improving support vector machine classifier by modifying kernel function.Neural Networks,1999,12(6):783-789
  • 3[6]C Soares,P B Brazdil,P Kuba.A meta-learning method to select the kernel width in support vector regression.Machine Learning,2004,54(33):195-209
  • 4[7]O Chapelle,V Vapnik.Model selection for support vector machines.In:Advances in Neural Information Processing Systems.Cambridge,MA:MIT Press,2000.230-236
  • 5[8]N Cristianini,C Chapelle,J Shawe-Taylor.Dynamically adapting kernels in support vector machines.In:Advances in Neural Information Processing Systems.Cambridge,MA:MIT Press,1998.204-210
  • 6[9]O Chapelle,V Vapnik,O Bousquet,et al.Choosing multiple parameters for support vector machines.Machine Learning,2002,46(11):131-159
  • 7[1]J Shawe-Taylor,N Cristianini.Kernel Methods for Pattern Analysis.Beijing:China Machine Press,2005
  • 8[2]V Vapnik.The Nature of Statistical Learning Theory.Berlin:Springer-Verlag,1995
  • 9V.N. Vapnik. The Nature of Statistical Learning Theory. New York: Springer-Verlag, 1995.
  • 10V. Cherkassky, F. Mulier. Learning from Data: Concept,Theory and Method. NY: John Viley & Sons, 1997.

共引文献47

同被引文献69

引证文献9

二级引证文献55

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部