Although originally biologically inspired neural networks were introduced as multilayer computational models, shallow networks have been dominant in applications till the recent renewal of interest in deep architectures. Experimental evidence and successful applications of deep networks pose theoretical questions asking: When and why are deep networks better than shallow ones? This chapter presents some probabilistic and constructive results on limitations of shallow networks. It shows implications of geometrical properties of high-dimensional spaces for probabilistic lower bounds on network complexity. The bounds depend on covering numbers of dictionaries of computational units and sizes of domains of functions to be computed. Probabilistic results are complemented by constructive ones built using Hadamard matrices and pseudo-noise sequences.
CITATION STYLE
Kůrková, V. (2020). Limitations of Shallow Networks. In Studies in Computational Intelligence (Vol. 896, pp. 129–154). Springer. https://doi.org/10.1007/978-3-030-43883-8_6
Mendeley helps you to discover research relevant for your work.