Abstract
The performance of traditional graph Laplacian methods for semi-supervised learning degrades substantially as the ratio of labeled to unlabeled data decreases, due to a degeneracy in the graph Laplacian. Several approaches have been proposed recently to address this, however we show that some of them remain ill-posed in the large-data limit. In this paper, we show a way to correctly set the weights in Laplacian regularization so that the estimator remains well posed and stable in the large-sample limit. We prove that our semi-supervised learning algorithm converges, in the infinite sample size limit, to the smooth solution of a continuum variational problem that attains the labeled values continuously. Our method is fast and easy to implement.
Original language | English (US) |
---|---|
Pages (from-to) | 1111-1159 |
Number of pages | 49 |
Journal | Applied Mathematics and Optimization |
Volume | 82 |
Issue number | 3 |
DOIs | |
State | Published - Dec 1 2020 |
Bibliographical note
Funding Information:Calder was supported by NSF Grant DMS:1713691. Slep?ev acknowledges the NSF support (Grants DMS-1516677 and DMS-1814991). He is grateful to University of Minnesota, where this project started, for hospitality. He is also grateful to the Center for Nonlinear Analysis of CMU for its support. Both authors thank the referees for valuable suggestions.
Funding Information:
Calder was supported by NSF Grant DMS:1713691. Slepčev acknowledges the NSF support (Grants DMS-1516677 and DMS-1814991). He is grateful to University of Minnesota, where this project started, for hospitality. He is also grateful to the Center for Nonlinear Analysis of CMU for its support. Both authors thank the referees for valuable suggestions.
Publisher Copyright:
© 2019, Springer Science+Business Media, LLC, part of Springer Nature.
Keywords
- Asymptotic consistency
- Gamma-convergence
- Label propagation
- PDEs on graphs
- Semi-supervised learning