Abstract
In this paper we prove convergence rates for the problem of approximating functions f by neural networks and similar constructions. We show that the rates are the better the smoother the activation functions are, provided that f satisfies an integral representation. We give error bounds not only in Hilbert spaces but also in general Sobolev spaces Wm,r(Ω). Finally, we apply our results to a class of perceptrons and present a sufficient smoothness condition on f guaranteeing the integral representation. © 2001 Academic Press.
Author supplied keywords
Cite
CITATION STYLE
Burger, M., & Neubauer, A. (2001). Error bounds for approximation with neural networks. Journal of Approximation Theory, 112(2), 235–250. https://doi.org/10.1006/jath.2001.3613
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.