Abstract
Complexity of shallow (one-hidden-layer) networks representing finitemultivariate mappings is investigated. Lower bounds are derived on growth of numbers of network units and sizes of output weights in terms of variational norms of mappings to be represented. Probability distributions of mappings whose computations require large networks are described. It is shown that due to geometrical properties of highdimensional Euclidean spaces, representation of almost any randomly chosen function on a sufficiently large domain by a shallow network with perceptrons requires untractably large network. Concrete examples of such functions are constructed using Hadamard matrices.
Author supplied keywords
Cite
CITATION STYLE
Kůrková, V. (2015). Complexity of shallow networks representing finite mappings. In Lecture Notes in Artificial Intelligence (Subseries of Lecture Notes in Computer Science) (Vol. 9119, pp. 39–48). Springer Verlag. https://doi.org/10.1007/978-3-319-19324-3_4
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.