Topological Properties of the Set of Functions Generated by Neural Networks of Fixed Size

55Citations
Citations of this article
61Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

We analyze the topological properties of the set of functions that can be implemented by neural networks of a fixed size. Surprisingly, this set has many undesirable properties. It is highly non-convex, except possibly for a few exotic activation functions. Moreover, the set is not closed with respect to Lp-norms, 0 < p< ∞, for all practically used activation functions, and also not closed with respect to the L∞-norm for all practically used activation functions except for the ReLU and the parametric ReLU. Finally, the function that maps a family of weights to the function computed by the associated network is not inverse stable for every practically used activation function. In other words, if f1, f2 are two functions realized by neural networks and if f1, f2 are close in the sense that ‖f1-f2‖L∞≤ε for ε> 0 , it is, regardless of the size of ε, usually not possible to find weights w1, w2 close together such that each fi is realized by a neural network with weights wi. Overall, our findings identify potential causes for issues in the training procedure of deep learning such as no guaranteed convergence, explosion of parameters, and slow convergence.

Cite

CITATION STYLE

APA

Petersen, P., Raslan, M., & Voigtlaender, F. (2021). Topological Properties of the Set of Functions Generated by Neural Networks of Fixed Size. Foundations of Computational Mathematics, 21(2), 375–444. https://doi.org/10.1007/s10208-020-09461-0

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free