Add like
Add dislike
Add to saved papers

Robust Alternating Low-Rank Representation by joint L p - and L 2,p -norm minimization.

We propose a robust Alternating Low-Rank Representation (ALRR) model formed by an alternating forward-backward representation process. For forward representation, ALRR first recovers the low-rank PCs and random corruptions by an adaptive local Robust PCA (RPCA). Then, ALRR performs a joint Lp -norm and L2,p -norm minimization (0<p <1) based sparse LRR by taking the low-rank PCs as inputs and dictionary instead of using the original noisy data to learn the coding coefficients for subspace recovery, where the Lp -norm on the coefficients can ensure joint sparsity for subspace representation, while the L2,p -norm on the reconstruction error can handle outlier pursuit. After that, ALRR returns the coefficients as adaptive weights to local RPCA for updating PCs and dictionary in the backward representation process. Thus, ALRR is regarded as an integration of local RPCA with adaptive weights plus sparse LRR with a self-expressive low-rank dictionary. To enable ALRR to handle outside data efficiently, a projective ALRR that can extract features from data directly by embedding is also derived. To solve the L2,p -norm based minimization problem, a new iterative scheme based on the Iterative Shrinkage/Thresholding (IST) approach is presented. The relationship analysis with other related criteria show that our methods are more general. Visual and numerical results demonstrate the effectiveness of our algorithms for representation.

Full text links

We have located links that may give you full text access.
Can't access the paper?
Try logging in through your university/institutional subscription. For a smoother one-click institutional access experience, please use our mobile app.

Related Resources

For the best experience, use the Read mobile app

Mobile app image

Get seemless 1-tap access through your institution/university

For the best experience, use the Read mobile app

All material on this website is protected by copyright, Copyright © 1994-2024 by WebMD LLC.
This website also contains material copyrighted by 3rd parties.

By using this service, you agree to our terms of use and privacy policy.

Your Privacy Choices Toggle icon

You can now claim free CME credits for this literature searchClaim now

Get seemless 1-tap access through your institution/university

For the best experience, use the Read mobile app