Skip to main navigation Skip to search Skip to main content

Depth Separation for Neural Networks

Research output: Contribution to journalConference articlepeer-review

39 Scopus citations

Abstract

Let f : Sd-1 × Sd-1 → R be a function of the form f(x, x0) = g(hx, x0i) for g : [-1, 1] → R. We give a simple proof that shows that poly-size depth two neural networks with (exponentially) bounded weights cannot approximate f whenever g cannot be approximated by a low degree polynomial. Moreover, for many g’s, such as g(x) = sin(πd3x), the number of neurons must be 2Ω(d log(d)). Furthermore, the result holds w.r.t. the uniform distribution on Sd-1 × Sd-1. As many functions of the above form can be well approximated by poly-size depth three networks with poly-bounded weights, this establishes a separation between depth two and depth three networks w.r.t.

Original languageEnglish
Pages (from-to)690-696
Number of pages7
JournalProceedings of Machine Learning Research
Volume65
StatePublished - 2017
Externally publishedYes
Event30th Conference on Learning Theory, COLT 2017 - Amsterdam, Netherlands
Duration: 7 Jul 201710 Jul 2017

Bibliographical note

Publisher Copyright:
© 2017 A. Daniely.

Keywords

  • Depth Separation
  • Neural Networks
  • Uniform Distribution

Fingerprint

Dive into the research topics of 'Depth Separation for Neural Networks'. Together they form a unique fingerprint.

Cite this