Principal Component Analysis (PCA) is a widely used technique for data analysis and dimensionality reduction, but its sensitivity to feature scale and outliers limits its applicability. Robust Principal Component Anal...Principal Component Analysis (PCA) is a widely used technique for data analysis and dimensionality reduction, but its sensitivity to feature scale and outliers limits its applicability. Robust Principal Component Analysis (RPCA) addresses these limitations by decomposing data into a low-rank matrix capturing the underlying structure and a sparse matrix identifying outliers, enhancing robustness against noise and outliers. This paper introduces a novel RPCA variant, Robust PCA Integrating Sparse and Low-rank Priors (RPCA-SL). Each prior targets a specific aspect of the data’s underlying structure and their combination allows for a more nuanced and accurate separation of the main data components from outliers and noise. Then RPCA-SL is solved by employing a proximal gradient algorithm for improved anomaly detection and data decomposition. Experimental results on simulation and real data demonstrate significant advancements.展开更多
The method of recovering a low-rank matrix with an unknown fraction whose entries are arbitrarily corrupted is known as the robust principal component analysis (RPCA). This RPCA problem, under some conditions, can b...The method of recovering a low-rank matrix with an unknown fraction whose entries are arbitrarily corrupted is known as the robust principal component analysis (RPCA). This RPCA problem, under some conditions, can be exactly solved via convex optimization by minimizing a combination of the nuclear norm and the 11 norm. In this paper, an algorithm based on the Douglas-Rachford splitting method is proposed for solving the RPCA problem. First, the convex optimization problem is solved by canceling the constraint of the variables, and ~hen the proximity operators of the objective function are computed alternately. The new algorithm can exactly recover the low-rank and sparse components simultaneously, and it is proved to be convergent. Numerical simulations demonstrate the practical utility of the proposed algorithm.展开更多
As the development of smart grid and energy internet, this leads to a significantincrease in the amount of data transmitted in real time. Due to the mismatch withcommunication networks that were not designed to carry ...As the development of smart grid and energy internet, this leads to a significantincrease in the amount of data transmitted in real time. Due to the mismatch withcommunication networks that were not designed to carry high-speed and real time data,data losses and data quality degradation may happen constantly. For this problem,according to the strong spatial and temporal correlation of electricity data which isgenerated by human’s actions and feelings, we build a low-rank electricity data matrixwhere the row is time and the column is user. Inspired by matrix decomposition, we dividethe low-rank electricity data matrix into the multiply of two small matrices and use theknown data to approximate the low-rank electricity data matrix and recover the missedelectrical data. Based on the real electricity data, we analyze the low-rankness of theelectricity data matrix and perform the Matrix Decomposition-based method on the realdata. The experimental results verify the efficiency and efficiency of the proposed scheme.展开更多
This paper considers approximately sparse signal and low-rank matrix’s recovery via truncated norm minimization minx∥xT∥q and minX∥XT∥Sq from noisy measurements.We first introduce truncated sparse approximation p...This paper considers approximately sparse signal and low-rank matrix’s recovery via truncated norm minimization minx∥xT∥q and minX∥XT∥Sq from noisy measurements.We first introduce truncated sparse approximation property,a more general robust null space property,and establish the stable recovery of signals and matrices under the truncated sparse approximation property.We also explore the relationship between the restricted isometry property and truncated sparse approximation property.And we also prove that if a measurement matrix A or linear map A satisfies truncated sparse approximation property of order k,then the first inequality in restricted isometry property of order k and of order 2k can hold for certain different constantsδk andδ2k,respectively.Last,we show that ifδs(k+|T^c|)<√(s-1)/s for some s≥4/3,then measurement matrix A and linear map A satisfy truncated sparse approximation property of order k.It should be pointed out that when Tc=Ф,our conclusion implies that sparse approximation property of order k is weaker than restricted isometry property of order sk.展开更多
Tensor robust principal component analysis has received a substantial amount of attention in various fields.Most existing methods,normally relying on tensor nuclear norm minimization,need to pay an expensive computati...Tensor robust principal component analysis has received a substantial amount of attention in various fields.Most existing methods,normally relying on tensor nuclear norm minimization,need to pay an expensive computational cost due to multiple singular value decompositions at each iteration.To overcome the drawback,we propose a scalable and efficient method,named parallel active subspace decomposition,which divides the unfolding along each mode of the tensor into a columnwise orthonormal matrix(active subspace)and another small-size matrix in parallel.Such a transformation leads to a nonconvex optimization problem in which the scale of nuclear norm minimization is generally much smaller than that in the original problem.We solve the optimization problem by an alternating direction method of multipliers and show that the iterates can be convergent within the given stopping criterion and the convergent solution is close to the global optimum solution within the prescribed bound.Experimental results are given to demonstrate that the performance of the proposed model is better than the state-of-the-art methods.展开更多
This paper aims at achieving a simultaneously sparse and low-rank estimator from the semidefinite population covariance matrices.We first benefit from a convex optimization which develops l1-norm penalty to encourage ...This paper aims at achieving a simultaneously sparse and low-rank estimator from the semidefinite population covariance matrices.We first benefit from a convex optimization which develops l1-norm penalty to encourage the sparsity and nuclear norm to favor the low-rank property.For the proposed estimator,we then prove that with high probability,the Frobenius norm of the estimation rate can be of order O(√((slgg p)/n))under a mild case,where s and p denote the number of nonzero entries and the dimension of the population covariance,respectively and n notes the sample capacity.Finally,an efficient alternating direction method of multipliers with global convergence is proposed to tackle this problem,and merits of the approach are also illustrated by practicing numerical simulations.展开更多
文摘Principal Component Analysis (PCA) is a widely used technique for data analysis and dimensionality reduction, but its sensitivity to feature scale and outliers limits its applicability. Robust Principal Component Analysis (RPCA) addresses these limitations by decomposing data into a low-rank matrix capturing the underlying structure and a sparse matrix identifying outliers, enhancing robustness against noise and outliers. This paper introduces a novel RPCA variant, Robust PCA Integrating Sparse and Low-rank Priors (RPCA-SL). Each prior targets a specific aspect of the data’s underlying structure and their combination allows for a more nuanced and accurate separation of the main data components from outliers and noise. Then RPCA-SL is solved by employing a proximal gradient algorithm for improved anomaly detection and data decomposition. Experimental results on simulation and real data demonstrate significant advancements.
基金supported by the National Natural Science Foundation of China(No.61271014)the Specialized Research Fund for the Doctoral Program of Higher Education(No.20124301110003)the Graduated Students Innovation Fund of Hunan Province(No.CX2012B238)
文摘The method of recovering a low-rank matrix with an unknown fraction whose entries are arbitrarily corrupted is known as the robust principal component analysis (RPCA). This RPCA problem, under some conditions, can be exactly solved via convex optimization by minimizing a combination of the nuclear norm and the 11 norm. In this paper, an algorithm based on the Douglas-Rachford splitting method is proposed for solving the RPCA problem. First, the convex optimization problem is solved by canceling the constraint of the variables, and ~hen the proximity operators of the objective function are computed alternately. The new algorithm can exactly recover the low-rank and sparse components simultaneously, and it is proved to be convergent. Numerical simulations demonstrate the practical utility of the proposed algorithm.
文摘As the development of smart grid and energy internet, this leads to a significantincrease in the amount of data transmitted in real time. Due to the mismatch withcommunication networks that were not designed to carry high-speed and real time data,data losses and data quality degradation may happen constantly. For this problem,according to the strong spatial and temporal correlation of electricity data which isgenerated by human’s actions and feelings, we build a low-rank electricity data matrixwhere the row is time and the column is user. Inspired by matrix decomposition, we dividethe low-rank electricity data matrix into the multiply of two small matrices and use theknown data to approximate the low-rank electricity data matrix and recover the missedelectrical data. Based on the real electricity data, we analyze the low-rankness of theelectricity data matrix and perform the Matrix Decomposition-based method on the realdata. The experimental results verify the efficiency and efficiency of the proposed scheme.
基金supported by the National Natural Science Foundation of China(11871109)NSAF(U1830107)the Science Challenge Project(TZ2018001)
文摘This paper considers approximately sparse signal and low-rank matrix’s recovery via truncated norm minimization minx∥xT∥q and minX∥XT∥Sq from noisy measurements.We first introduce truncated sparse approximation property,a more general robust null space property,and establish the stable recovery of signals and matrices under the truncated sparse approximation property.We also explore the relationship between the restricted isometry property and truncated sparse approximation property.And we also prove that if a measurement matrix A or linear map A satisfies truncated sparse approximation property of order k,then the first inequality in restricted isometry property of order k and of order 2k can hold for certain different constantsδk andδ2k,respectively.Last,we show that ifδs(k+|T^c|)<√(s-1)/s for some s≥4/3,then measurement matrix A and linear map A satisfy truncated sparse approximation property of order k.It should be pointed out that when Tc=Ф,our conclusion implies that sparse approximation property of order k is weaker than restricted isometry property of order sk.
基金the HKRGC GRF 12306616,12200317,12300218 and 12300519,and HKU Grant 104005583.
文摘Tensor robust principal component analysis has received a substantial amount of attention in various fields.Most existing methods,normally relying on tensor nuclear norm minimization,need to pay an expensive computational cost due to multiple singular value decompositions at each iteration.To overcome the drawback,we propose a scalable and efficient method,named parallel active subspace decomposition,which divides the unfolding along each mode of the tensor into a columnwise orthonormal matrix(active subspace)and another small-size matrix in parallel.Such a transformation leads to a nonconvex optimization problem in which the scale of nuclear norm minimization is generally much smaller than that in the original problem.We solve the optimization problem by an alternating direction method of multipliers and show that the iterates can be convergent within the given stopping criterion and the convergent solution is close to the global optimum solution within the prescribed bound.Experimental results are given to demonstrate that the performance of the proposed model is better than the state-of-the-art methods.
基金The work was supported in part by the National Natural Science Foundation of China(Nos.11431002,11171018,71271021,11301022).
文摘This paper aims at achieving a simultaneously sparse and low-rank estimator from the semidefinite population covariance matrices.We first benefit from a convex optimization which develops l1-norm penalty to encourage the sparsity and nuclear norm to favor the low-rank property.For the proposed estimator,we then prove that with high probability,the Frobenius norm of the estimation rate can be of order O(√((slgg p)/n))under a mild case,where s and p denote the number of nonzero entries and the dimension of the population covariance,respectively and n notes the sample capacity.Finally,an efficient alternating direction method of multipliers with global convergence is proposed to tackle this problem,and merits of the approach are also illustrated by practicing numerical simulations.