On the convergence of MDL density estimation

6Citations
Citations of this article
4Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present a general information exponential inequality that measures the statistical complexity of some deterministic and randomized density estimators. Using this inequality, we are able to improve classical results concerning the convergence of two-part code MDL in [1], Moreover, we are able to derive clean finite-sample convergence bounds that are not obtainable using previous approaches.

Cite

CITATION STYLE

APA

Zhang, T. (2004). On the convergence of MDL density estimation. In Lecture Notes in Artificial Intelligence (Subseries of Lecture Notes in Computer Science) (Vol. 3120, pp. 315–330). Springer Verlag. https://doi.org/10.1007/978-3-540-27819-1_22

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free