Skip to main navigation Skip to search Skip to main content

Effective Semi-supervised Learning on Manifolds

Research output: Contribution to journalConference articlepeer-review

10 Scopus citations

Abstract

The abundance of unlabeled data makes semi-supervised learning (SSL) an attractive approach for improving the accuracy of learning systems. However, we are still far from a complete theoretical understanding of the benefits of this learning scenario in terms of sample complexity. In particular, for many natural learning settings it can in fact be shown that SSL does not improve sample complexity. Thus far, the only case where SSL provably helps, without compatibility assumptions, is a recent combinatorial construction of Darnstädt et al. (2013). Deriving similar theoretical guarantees for more commonly used approaches to SSL remains a challenge. Here, we provide the first analysis of manifold based SSL, where there is a provable gap between supervised learning and SSL, and this gap can be arbitrarily large. Proving the required lower bound is a technical challenge, involving tools from geometric measure theory. The algorithm we analyze is similar to subspace clustering, and thus our results demonstrate that this method can be used to improve sample complexity.

Original languageEnglish
Pages (from-to)978-1003
Number of pages26
JournalProceedings of Machine Learning Research
Volume65
StatePublished - 2017
Event30th Conference on Learning Theory, COLT 2017 - Amsterdam, Netherlands
Duration: 7 Jul 201710 Jul 2017

Bibliographical note

Publisher Copyright:
© 2017 A. Globerson, R. Livni & S. Shalev-Shwartz.

Fingerprint

Dive into the research topics of 'Effective Semi-supervised Learning on Manifolds'. Together they form a unique fingerprint.

Cite this