Hermite polynomials facilitating on-line learning analysis of layered neural networks with arbitrary activation function

1Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Following the standard statistical mechanics methods we analyze the training by online stochastic gradient descent of two-layer neural networks in a student–teacher scenario. We focus on understanding the role that different activations play, in particular mismatches between the student and the teacher, in these learning scenarios. By expanding the activation functions in the Hermite polynomial basis, we are able to effectively approximate the relevant integrals with much less computational effort than naive numerical integration. Moreover, we also extend the framework to study scenarios of concept drift and weight decay also with arbitrary activation functions. All these extensions comprise relevant advances in the field, allowing us to obtain analytical results for more realistic scenarios.

Cite

CITATION STYLE

APA

Citton, O., Richert, F., Biehl, M., & Straat, M. (2025). Hermite polynomials facilitating on-line learning analysis of layered neural networks with arbitrary activation function. Neurocomputing, 655. https://doi.org/10.1016/j.neucom.2025.131328

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free