A chaos hypothesis for some large systems of random equations

19Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

This paper is about the behavior of solutions to large systems of linear algebraic and differential equations when the coefficients are random variables. We will prove a law of large numbers and a central limit theorem for the solutions of certain algebraic systems, and the weak convergence to a Gaussian process for the solution of a system of differential equations. Some of the results were surprisingly difficult to prove, but they are all easily anticipated from a "chaos hypothesis": i.e. an assumption of near independence for the components of the solutions of large systems of weakly coupled equations. © 1982 Springer-Verlag.

Cite

CITATION STYLE

APA

Geman, S., & Hwang, C. R. (1982). A chaos hypothesis for some large systems of random equations. Zeitschrift Für Wahrscheinlichkeitstheorie Und Verwandte Gebiete, 60(3), 291–314. https://doi.org/10.1007/BF00535717

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free