Embedding theory of reservoir computing and reducing reservoir network using time delays

22Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.

Abstract

Reservoir computing (RC), a particular form of recurrent neural network, is under explosive development due to its exceptional efficacy and high performance in reconstruction and/or prediction of complex physical systems. However, the mechanism triggering such effective applications of RC is still unclear, awaiting deep and systematic exploration. Here, combining the delayed embedding theory with the generalized embedding theory, we rigorously prove that RC is essentially a high-dimensional embedding of the original input nonlinear dynamical system. Thus, using this embedding property, we unify into a universal framework the standard RC and the time-delayed RC where we introduce time delays only into the network's output layer, and we further find a trade-off relation between the time delays and the number of neurons in RC. Based on these findings, we significantly reduce the RC's network size and promote its memory capacity in completing systems reconstruction and prediction. More surprisingly, only using a single-neuron reservoir with time delays is sometimes sufficient for achieving reconstruction and prediction tasks, while the standard RC of any large size but without time delay cannot complete them yet.

Cite

CITATION STYLE

APA

Duan, X. Y., Ying, X., Leng, S. Y., Kurths, J., Lin, W., & Ma, H. F. (2023). Embedding theory of reservoir computing and reducing reservoir network using time delays. Physical Review Research, 5(2). https://doi.org/10.1103/PhysRevResearch.5.L022041

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free