Algorithmic statistics revisited

5Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The mission of statistics is to provide adequate statistical hypotheses (models) for observed data. But what is an "adequate" model? To answer this question, one needs to use the notions of algorithmic information theory. It turns out that for every data string x one can naturally define a "stochasticity profile," a curve that represents a trade-off between the complexity of a model and its adequacy. This curve has four different equivalent definitions in terms of (1) randomness deficiency, (2) minimal description length, (3) position in the lists of simple strings, and (4) Kolmogorov complexity with decompression time bounded by the busy beaver function. We present a survey of the corresponding definitions and results relating them to each other.

Cite

CITATION STYLE

APA

Vereshchagin, N., & Shen, A. (2015). Algorithmic statistics revisited. In Measures of Complexity: Festschrift for Alexey Chervonenkis (pp. 235–252). Springer International Publishing. https://doi.org/10.1007/978-3-319-21852-6_17

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free