Abstract
Let f : Sd-1 × Sd-1 → R be a function of the form f(x, x0) = g(hx, x0i) for g : [-1, 1] → R. We give a simple proof that shows that poly-size depth two neural networks with (exponentially) bounded weights cannot approximate f whenever g cannot be approximated by a low degree polynomial. Moreover, for many g’s, such as g(x) = sin(πd3x), the number of neurons must be 2Ω(d log(d)). Furthermore, the result holds w.r.t. the uniform distribution on Sd-1 × Sd-1. As many functions of the above form can be well approximated by poly-size depth three networks with poly-bounded weights, this establishes a separation between depth two and depth three networks w.r.t.
| Original language | English |
|---|---|
| Pages (from-to) | 690-696 |
| Number of pages | 7 |
| Journal | Proceedings of Machine Learning Research |
| Volume | 65 |
| State | Published - 2017 |
| Externally published | Yes |
| Event | 30th Conference on Learning Theory, COLT 2017 - Amsterdam, Netherlands Duration: 7 Jul 2017 → 10 Jul 2017 |
Bibliographical note
Publisher Copyright:© 2017 A. Daniely.
Keywords
- Depth Separation
- Neural Networks
- Uniform Distribution
Fingerprint
Dive into the research topics of 'Depth Separation for Neural Networks'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver