The Universal Approximation Property: Characterization, Construction, Representation, and Existence

43Citations
Citations of this article
23Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

The universal approximation property of various machine learning models is currently only understood on a case-by-case basis, limiting the rapid development of new theoretically justified neural network architectures and blurring our understanding of our current models’ potential. This paper works towards overcoming these challenges by presenting a characterization, a representation, a construction method, and an existence result, each of which applies to any universal approximator on most function spaces of practical interest. Our characterization result is used to describe which activation functions allow the feed-forward architecture to maintain its universal approximation capabilities when multiple constraints are imposed on its final layers and its remaining layers are only sparsely connected. These include a rescaled and shifted Leaky ReLU activation function but not the ReLU activation function. Our construction and representation result is used to exhibit a simple modification of the feed-forward architecture, which can approximate any continuous function with non-pathological growth, uniformly on the entire Euclidean input space. This improves the known capabilities of the feed-forward architecture.

Cite

CITATION STYLE

APA

Kratsios, A. (2021). The Universal Approximation Property: Characterization, Construction, Representation, and Existence. Annals of Mathematics and Artificial Intelligence, 89(5–6), 435–469. https://doi.org/10.1007/s10472-020-09723-1

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free