期刊文献+
共找到2篇文章
< 1 >
每页显示 20 50 100
Efficient Preference Clustering via Random Fourier Features 被引量:1
1
作者 Jingshu Liu Li Wang Jinglei Liu 《Big Data Mining and Analytics》 2019年第3期195-204,共10页
Approximations based on random Fourier features have recently emerged as an efficient and elegant method for designing large-scale machine learning tasks.Unlike approaches using the Nystr?m method,which randomly sampl... Approximations based on random Fourier features have recently emerged as an efficient and elegant method for designing large-scale machine learning tasks.Unlike approaches using the Nystr?m method,which randomly samples the training examples,we make use of random Fourier features,whose basis functions(i.e.,cosine and sine)are sampled from a distribution independent from the training sample set,to cluster preference data which appears extensively in recommender systems.Firstly,we propose a two-stage preference clustering framework.In this framework,we make use of random Fourier features to map the preference matrix into the feature matrix,soon afterwards,utilize the traditional k-means approach to cluster preference data in the transformed feature space.Compared with traditional preference clustering,our method solves the problem of insufficient memory and greatly improves the efficiency of the operation.Experiments on movie data sets containing 100000 ratings,show that the proposed method is more effective in clustering accuracy than the Nystr?m and k-means,while also achieving better performance than these clustering approaches. 展开更多
关键词 random fourier features matrix decomposition similarity matrix Nystrom method preference clustering
原文传递
Fast Maximum Entropy Machine for Big Imbalanced Datasets
2
作者 Feng Yin Shuqing Lin +1 位作者 Chuxin Piao Shuguang(Robert)Cui 《Journal of Communications and Information Networks》 2018年第3期20-30,共11页
Driven by the need of a plethora of machine learning applications,several attempts have been made at improving the performance of classifiers applied to imbalanced datasets.In this paper,we present a fast maximum entr... Driven by the need of a plethora of machine learning applications,several attempts have been made at improving the performance of classifiers applied to imbalanced datasets.In this paper,we present a fast maximum entropy machine(MEM)combined with a synthetic minority over-sampling technique for handling binary classification problems with high imbalance ratios,large numbers of data samples,and medium/large numbers of features.A random Fourier feature representation of kernel functions and primal estimated sub-gradient solver for support vector machine(PEGASOS)are applied to speed up the classic MEM.Experiments have been conducted using various real datasets(including two China Mobile datasets and several other standard test datasets)with various configurations.The obtained results demonstrate that the proposed algorithm has extremely low complexity but an excellent overall classification performance(in terms of several widely used evaluation metrics)as compared to the classic MEM and some other state-of-the-art methods.The proposed algorithm is particularly valuable in big data applications owing to its significantly low computational complexity. 展开更多
关键词 binary classification imbalanced datasets maximum entropy machine PEGASOS random fourier feature SMOTE
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部