TY - GEN
T1 - Robust low-rank subspace segmentation with semidefinite guarantees
AU - Ni, Yuzhao
AU - Sun, Ju
AU - Yuan, Xiaotong
AU - Yan, Shuicheng
AU - Cheong, Loong Fah
PY - 2010/12/1
Y1 - 2010/12/1
N2 - Recently there is a line of research work proposing to employ Spectral Clustering (SC) to segment (group)1\ high-dimensional structural data such as those (approximately) lying on subspaces2 or low-dimensional manifolds. By learning the affinity matrix in the form of sparse reconstruction, techniques proposed in this vein often considerably boost the performance in subspace settings where traditional SC can fail. Despite the success, there are fundamental problems that have been left unsolved: the spectrum property of the learned affinity matrix cannot be gauged in advance, and there is often one ugly symmetrization step that post-processes the affinity for SC input. Hence we advocate to enforce the symmetric positive semi definite constraint explicitly during learning (Low-Rank Representation with Positive Semi Definite constraint, or LRR-PSD), and show that factually it can be solved in an exquisite scheme efficiently instead of general-purpose SDP solvers that usually scale up poorly. We provide rigorous mathematical derivations to show that, in its canonical form, LRR-PSD is equivalent to the recently proposed Low-Rank Representation (LRR) scheme[1], and hence offer theoretic and practical insights to both LRR-PSD and LRR, inviting future research. As per the computational cost, our proposal is at most comparable to that of LRR, if not less. We validate our theoretic analysis and optimization scheme by experiments on both synthetic and real data sets.
AB - Recently there is a line of research work proposing to employ Spectral Clustering (SC) to segment (group)1\ high-dimensional structural data such as those (approximately) lying on subspaces2 or low-dimensional manifolds. By learning the affinity matrix in the form of sparse reconstruction, techniques proposed in this vein often considerably boost the performance in subspace settings where traditional SC can fail. Despite the success, there are fundamental problems that have been left unsolved: the spectrum property of the learned affinity matrix cannot be gauged in advance, and there is often one ugly symmetrization step that post-processes the affinity for SC input. Hence we advocate to enforce the symmetric positive semi definite constraint explicitly during learning (Low-Rank Representation with Positive Semi Definite constraint, or LRR-PSD), and show that factually it can be solved in an exquisite scheme efficiently instead of general-purpose SDP solvers that usually scale up poorly. We provide rigorous mathematical derivations to show that, in its canonical form, LRR-PSD is equivalent to the recently proposed Low-Rank Representation (LRR) scheme[1], and hence offer theoretic and practical insights to both LRR-PSD and LRR, inviting future research. As per the computational cost, our proposal is at most comparable to that of LRR, if not less. We validate our theoretic analysis and optimization scheme by experiments on both synthetic and real data sets.
KW - Affinity matrix learning
KW - Eigenvalue thresholding
KW - Rank minimization
KW - Robust estimation
KW - Spectral clustering
UR - http://www.scopus.com/inward/record.url?scp=79951739650&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=79951739650&partnerID=8YFLogxK
U2 - 10.1109/ICDMW.2010.64
DO - 10.1109/ICDMW.2010.64
M3 - Conference contribution
AN - SCOPUS:79951739650
SN - 9780769542577
T3 - Proceedings - IEEE International Conference on Data Mining, ICDM
SP - 1179
EP - 1188
BT - Proceedings - 10th IEEE International Conference on Data Mining Workshops, ICDMW 2010
T2 - 10th IEEE International Conference on Data Mining Workshops, ICDMW 2010
Y2 - 14 December 2010 through 17 December 2010
ER -