Function approximation by deep networks

9Citations
Citations of this article
39Readers
Mendeley users who have this article in their library.

Abstract

We show that deep networks are better than shallow networks at approximating functions that can be expressed as a composition of functions described by a directed acyclic graph, because the deep networks can be designed to have the same compositional structure, while a shallow network cannot exploit this knowledge. Thus, the blessing of compositionality mitigates the curse of dimensionality. On the other hand, a theorem called good propagation of errors allows to "lift" theorems about shallow networks to those about deep networks with an appropriate choice of norms, smoothness, etc. We illustrate this in three contexts where each channel in the deep network calculates a spherical polynomial, a non-smooth ReLU network, or another zonal function network related closely with the ReLU network.

Cite

CITATION STYLE

APA

Mhaskar, H. N., & Poggio, T. (2020). Function approximation by deep networks. Communications on Pure and Applied Analysis, 19(8), 4085–4095. https://doi.org/10.3934/cpaa.2020181

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free