Abstract
We derive bounds on the sample complexity of empirical risk minimization (ERM) in the context of minimizing non-convex risks that admit the strict saddle property. Recent progress in non-convex optimization has yielded efficient algorithms for minimizing such functions. Our results imply that these efficient algorithms are statistically stable and also generalize well. In particular, we derive fast rates which resemble the bounds that are often attained in the strongly convex setting. We specify our bounds to Principal Component Analysis and Independent Component Analysis. Our results and techniques may pave the way for statistical analyses of additional strict saddle problems.
| Original language | English |
|---|---|
| Pages (from-to) | 1043-1063 |
| Number of pages | 21 |
| Journal | Proceedings of Machine Learning Research |
| Volume | 65 |
| State | Published - 2017 |
| Event | 30th Conference on Learning Theory, COLT 2017 - Amsterdam, Netherlands Duration: 7 Jul 2017 → 10 Jul 2017 |
Bibliographical note
Publisher Copyright:© 2017 A. Gonen & S. Shalev-Shwartz.
Fingerprint
Dive into the research topics of 'Fast Rates for Empirical Risk Minimization of Strict Saddle Problems'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver