Limitations of Shallow Networks

1Citations
Citations of this article
1Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Although originally biologically inspired neural networks were introduced as multilayer computational models, shallow networks have been dominant in applications till the recent renewal of interest in deep architectures. Experimental evidence and successful applications of deep networks pose theoretical questions asking: When and why are deep networks better than shallow ones? This chapter presents some probabilistic and constructive results on limitations of shallow networks. It shows implications of geometrical properties of high-dimensional spaces for probabilistic lower bounds on network complexity. The bounds depend on covering numbers of dictionaries of computational units and sizes of domains of functions to be computed. Probabilistic results are complemented by constructive ones built using Hadamard matrices and pseudo-noise sequences.

Cite

CITATION STYLE

APA

Kůrková, V. (2020). Limitations of Shallow Networks. In Studies in Computational Intelligence (Vol. 896, pp. 129–154). Springer. https://doi.org/10.1007/978-3-030-43883-8_6

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free