A least squares version of the recently proposed weighted twin support vector machine with local information(WLTSVM) for binary classification is formulated. This formulation leads to an extremely simple and fast algo...A least squares version of the recently proposed weighted twin support vector machine with local information(WLTSVM) for binary classification is formulated. This formulation leads to an extremely simple and fast algorithm, called least squares weighted twin support vector machine with local information(LSWLTSVM), for generating binary classifiers based on two non-parallel hyperplanes. Two modified primal problems of WLTSVM are attempted to solve, instead of two dual problems usually solved. The solution of the two modified problems reduces to solving just two systems of linear equations as opposed to solving two quadratic programming problems along with two systems of linear equations in WLTSVM. Moreover, two extra modifications were proposed in LSWLTSVM to improve the generalization capability. One is that a hot kernel function, not the simple-minded definition in WLTSVM, is used to define the weight matrix of adjacency graph, which ensures that the underlying similarity information between any pair of data points in the same class can be fully reflected. The other is that the weight for each point in the contrary class is considered in constructing equality constraints, which makes LSWLTSVM less sensitive to noise points than WLTSVM. Experimental results indicate that LSWLTSVM has comparable classification accuracy to that of WLTSVM but with remarkably less computational time.展开更多
According to the aggregation method of experts' evaluation information in group decision-making,the existing methods of determining experts' weights based on cluster analysis take into account the expert's preferen...According to the aggregation method of experts' evaluation information in group decision-making,the existing methods of determining experts' weights based on cluster analysis take into account the expert's preferences and the consistency of expert's collating vectors,but they lack of the measure of information similarity.So it may occur that although the collating vector is similar to the group consensus,information uncertainty is great of a certain expert.However,it is clustered to a larger group and given a high weight.For this,a new aggregation method based on entropy and cluster analysis in group decision-making process is provided,in which the collating vectors are classified with information similarity coefficient,and the experts' weights are determined according to the result of classification,the entropy of collating vectors and the judgment matrix consistency.Finally,a numerical example shows that the method is feasible and effective.展开更多
In traditional data clustering, similarity of a cluster of objects is measured by distance between objects. Such measures are not appropriate for categorical data. A new clustering criterion to determine the similarit...In traditional data clustering, similarity of a cluster of objects is measured by distance between objects. Such measures are not appropriate for categorical data. A new clustering criterion to determine the similarity between points with categorical attributes is pre- sented. Furthermore, a new clustering algorithm for categorical attributes is addressed. A single scan of the dataset yields a good clus- tering, and more additional passes can be used to improve the quality further.展开更多
The likelihood function plays a central role in statistical analysis in relation to information, from both frequentist and Bayesian perspectives. In large samples several new properties of the likelihood in relation t...The likelihood function plays a central role in statistical analysis in relation to information, from both frequentist and Bayesian perspectives. In large samples several new properties of the likelihood in relation to information are developed here. The Arrow-Pratt absolute risk aversion measure is shown to be related to the Cramer-Rao Information bound. The derivative of the log-likelihood function is seen to provide a measure of information related stability for the Bayesian posterior density. As well, information similar prior densities can be defined reflecting the central role of likelihood in the Bayes learning paradigm.展开更多
Information networks that can be extracted from many domains are widely studied recently. Different functions for mining these networks are proposed and developed, such as ranking, community detection, and link predic...Information networks that can be extracted from many domains are widely studied recently. Different functions for mining these networks are proposed and developed, such as ranking, community detection, and link prediction. Most existing network studies are on homogeneous networks, where nodes and links are assumed from one single type. In reality, however, heterogeneous information networks can better model the real-world systems, which are typically semi-structured and typed, following a network schema. In order to mine these heterogeneous information networks directly, we propose to explore the meta structure of the information network, i.e., the network schema. The concepts of meta-paths are proposed to systematically capture numerous semantic relationships across multiple types of objects, which are defined as a path over the graph of network schema. Meta-paths can provide guidance for search and mining of the network and help analyze and understand the semantic meaning of the objects and relations in the network. Under this framework, similarity search and other mining tasks such as relationship prediction and clustering can be addressed by systematic exploration of the network meta structure. Moreover, with user's guidance or feedback, we can select the best meta-path or their weighted combination for a specific mining task.展开更多
In this paper, we analyze the 180 stocks which have the potential influence on the Shanghai Stock Exchange(SSE). First, we use the stock closing prices from January 1, 2005 to June 19, 2015 to calculate logarithmic th...In this paper, we analyze the 180 stocks which have the potential influence on the Shanghai Stock Exchange(SSE). First, we use the stock closing prices from January 1, 2005 to June 19, 2015 to calculate logarithmic the correlation coefficient and then build the stock market model by threshold method. Secondly, according to different networks under different thresholds, we find out the potential influence stocks on the basis of local structural centrality. Finally, by comparing the accuracy of similarity index of the local information and path in the link prediction method, we demonstrate that there are best similarity index to predict the probability for nodes connection in the different stock networks.展开更多
基金Project(61105057)supported by the National Natural Science Foundation of ChinaProject(13KJB520024)supported by the Natural Science Foundation of Jiangsu Higher Education Institutes of ChinaProject supported by Jiangsu Province Qing Lan Project,China
文摘A least squares version of the recently proposed weighted twin support vector machine with local information(WLTSVM) for binary classification is formulated. This formulation leads to an extremely simple and fast algorithm, called least squares weighted twin support vector machine with local information(LSWLTSVM), for generating binary classifiers based on two non-parallel hyperplanes. Two modified primal problems of WLTSVM are attempted to solve, instead of two dual problems usually solved. The solution of the two modified problems reduces to solving just two systems of linear equations as opposed to solving two quadratic programming problems along with two systems of linear equations in WLTSVM. Moreover, two extra modifications were proposed in LSWLTSVM to improve the generalization capability. One is that a hot kernel function, not the simple-minded definition in WLTSVM, is used to define the weight matrix of adjacency graph, which ensures that the underlying similarity information between any pair of data points in the same class can be fully reflected. The other is that the weight for each point in the contrary class is considered in constructing equality constraints, which makes LSWLTSVM less sensitive to noise points than WLTSVM. Experimental results indicate that LSWLTSVM has comparable classification accuracy to that of WLTSVM but with remarkably less computational time.
文摘According to the aggregation method of experts' evaluation information in group decision-making,the existing methods of determining experts' weights based on cluster analysis take into account the expert's preferences and the consistency of expert's collating vectors,but they lack of the measure of information similarity.So it may occur that although the collating vector is similar to the group consensus,information uncertainty is great of a certain expert.However,it is clustered to a larger group and given a high weight.For this,a new aggregation method based on entropy and cluster analysis in group decision-making process is provided,in which the collating vectors are classified with information similarity coefficient,and the experts' weights are determined according to the result of classification,the entropy of collating vectors and the judgment matrix consistency.Finally,a numerical example shows that the method is feasible and effective.
文摘In traditional data clustering, similarity of a cluster of objects is measured by distance between objects. Such measures are not appropriate for categorical data. A new clustering criterion to determine the similarity between points with categorical attributes is pre- sented. Furthermore, a new clustering algorithm for categorical attributes is addressed. A single scan of the dataset yields a good clus- tering, and more additional passes can be used to improve the quality further.
文摘The likelihood function plays a central role in statistical analysis in relation to information, from both frequentist and Bayesian perspectives. In large samples several new properties of the likelihood in relation to information are developed here. The Arrow-Pratt absolute risk aversion measure is shown to be related to the Cramer-Rao Information bound. The derivative of the log-likelihood function is seen to provide a measure of information related stability for the Bayesian posterior density. As well, information similar prior densities can be defined reflecting the central role of likelihood in the Bayes learning paradigm.
基金supported in part by the U.S.Army Research Laboratory under Cooperative Agreement No.W911NF-09-2-0053(NS-CTA),NSF ⅡS-0905215,CNS-09-31975MIAS,a DHS-IDS Center for Multimodal Information Access and Synthesis at UIUC
文摘Information networks that can be extracted from many domains are widely studied recently. Different functions for mining these networks are proposed and developed, such as ranking, community detection, and link prediction. Most existing network studies are on homogeneous networks, where nodes and links are assumed from one single type. In reality, however, heterogeneous information networks can better model the real-world systems, which are typically semi-structured and typed, following a network schema. In order to mine these heterogeneous information networks directly, we propose to explore the meta structure of the information network, i.e., the network schema. The concepts of meta-paths are proposed to systematically capture numerous semantic relationships across multiple types of objects, which are defined as a path over the graph of network schema. Meta-paths can provide guidance for search and mining of the network and help analyze and understand the semantic meaning of the objects and relations in the network. Under this framework, similarity search and other mining tasks such as relationship prediction and clustering can be addressed by systematic exploration of the network meta structure. Moreover, with user's guidance or feedback, we can select the best meta-path or their weighted combination for a specific mining task.
基金Supported by the National Nature Science Foundation of China(71271103,71371087)
文摘In this paper, we analyze the 180 stocks which have the potential influence on the Shanghai Stock Exchange(SSE). First, we use the stock closing prices from January 1, 2005 to June 19, 2015 to calculate logarithmic the correlation coefficient and then build the stock market model by threshold method. Secondly, according to different networks under different thresholds, we find out the potential influence stocks on the basis of local structural centrality. Finally, by comparing the accuracy of similarity index of the local information and path in the link prediction method, we demonstrate that there are best similarity index to predict the probability for nodes connection in the different stock networks.