期刊文献+

面向机器学习的相对变换 被引量:10

Relative Transformation for Machine Learning
下载PDF
导出
摘要 机器学习常常面临数据稀疏和数据噪音问题.根据认知的相对性规律提出了相对变换方法,证明了相对变换是非线性的放大变换,可提高数据之间的可区分性.同时在一定条件下相对变换还能抑制噪音,并使稀疏的数据变得相对密集.通过相对变换将数据的原始空间变换到相对空间后,在相对空间中度量数据的相似性或距离更加符合人们的直觉,从而提高机器学习的性能.理论分析和实践验证了所提方法的普适性和有效性. Recently developed machine learning approaches such as manifold learning and the support vector machine learning work well on the clean data sets even if these data sets are highly folded, twisted, or curved. However, they are much sensitive to noises or outliers contained in the data set, as these noises or outliers easily distort the real topological structure of the underlying data manifold. To solve the problem, the relative transformation on the original data space is proposed by modeling the cognitive relative laws. It is proved that the relative transformation is a kind of nonlinear enlarging transformation so that it makes the transformed data more distinguishable. Meanwhile, the relative transformation can weaken the influence of noise on data and make data relative denser. To measure the similarity and distance between data points in relative space is more consistent with the intuition of people, which can be then applied to improve the machine learning approach. The relative transformation is simple, general and easy to implement. It also has clear physical meaning and does not add any parameter. The theoretical analysis and conducted experiments validate the proposed approach.
作者 文贵华
出处 《计算机研究与发展》 EI CSCD 北大核心 2008年第4期612-618,共7页 Journal of Computer Research and Development
基金 广东省科技攻关基金项目(2007B030803006) 教育部留学回国人员科研启动基金项目
关键词 机器学习 认知规律 相对变换 噪音数据 稀疏数据 machine learning cognitive laws relative transformation noisy data sparse data
  • 相关文献

参考文献15

  • 1J B Tenenbaum, V de Silva, J C Langford. A global geometric framework for nonlinear dimensionality reduction [J]. Science, 2000, 290(5500) : 2319-2323
  • 2M Balasubramanian, E L Schwartz. The ISOMAP algorithm and topological stability [J].Science, 2002, 295(5552) : 7
  • 3Heeyoul Choi, Seungjin Choi. Robust kernel ISOMAP [J]. Pattern Recognition, 2007, 40(3): 853-862
  • 4王和勇,郑杰,姚正安,李磊.基于聚类和改进距离的LLE方法在数据降维中的应用[J].计算机研究与发展,2006,43(8):1485-1490. 被引量:31
  • 5S T Roweis, L K Saul. Nonlinear dimensionality reduction by locally linear embedding [J]. Science, 2000, 290(5500): 2323 -2326
  • 6H S Sung, D D Lee. The manifold ways of perception [J]. Science, 2000, 290(5500): 2268-2269
  • 7李德毅,刘常昱,杜鹢,韩旭.不确定性人工智能[J].软件学报,2004,15(11):1583-1594. 被引量:397
  • 8Wen Guihua, Jiang Lijun, Wen Jun, et al. Generating creative ideas through patents [G]. In: LNAI 4099. Berlin: Springer, 2006. 681-690
  • 9贺玲,吴玲达,蔡益朝.高维空间中数据的相似性度量[J].数学的实践与认识,2006,36(9):189-194. 被引量:20
  • 10Wen Guihua, Jiang Liiun, Wen Jun, et al. Performing locally linear embedding with adaptive neighborhood size on manifold [G]. In: LNAI 4099. Berlin: Springer, 2006. 985-989

二级参考文献29

  • 1汪祖媛,庄镇泉,王煦法.逐维聚类的相似度索引算法[J].计算机研究与发展,2004,41(6):1003-1009. 被引量:5
  • 2李德毅,刘常昱.论正态云模型的普适性[J].中国工程科学,2004,6(8):28-34. 被引量:885
  • 3王梓坤.论混沌与随机[J].北京师范大学学报(自然科学版),1994,30(2):199-202. 被引量:16
  • 4李德毅,孟海军,史雪梅.隶属云和隶属云发生器[J].计算机研究与发展,1995,32(6):15-20. 被引量:1223
  • 5I T Jolliffe. Principal Component Analysis [M]. Berlin:Springer, 1986
  • 6B Scholkopf, A Smola, K R Muller. Nonlinear component analysis as a kernel eigenvalue problem [J]. Neural Computation, 1998, 10(5): 1299-1319
  • 7S Mika, G Ratsch, J Weston, et al. Fisher discriminant analysis with kernels [J]. Proceedings of IEEE Neural Networks for Signal Processing Workshop, 1999, 8(9) : 41-48
  • 8I Borg, P Groenen. Modern Multidimensional Scaling: Theory and Applications [ M ]. New York: Springer-Verlag, 1997
  • 9T Kohonen. The self-organizing map [J]. Proceedings of the IEEE, 1990, 78(9): 1464-1480
  • 10J B Tenenbaum, Vin de Silva, C John. A global geometric framework for nonlinear dimensionality reduction [J]. Science,2000, 290:2319-2323

共引文献496

同被引文献99

引证文献10

二级引证文献42

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部