Deep and Wide Neural Networks Covariance Estimation

1Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

It has been recently shown that a deep neural network with i.i.d. random parameters is equivalent to a Gaussian process in the limit of infinite network width. The Gaussian process associated to the neural network is fully described by a recursive covariance kernel determined by the architecture of the network, and which is expressed in terms of expectation. We give a numerically workable analytic expression of the neural network recursive covariance based on Hermite polynomials. We give explicit forms of this recursive covariance for the cases of neural networks with activation function the Heaviside, ReLU and sigmoid.

Cite

CITATION STYLE

APA

Arratia, A., Cabaña, A., & León, J. R. (2020). Deep and Wide Neural Networks Covariance Estimation. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12396 LNCS, pp. 195–206). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-61609-0_16

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free