A characterization of entropy in terms of information loss

95Citations
Citations of this article
175Readers
Mendeley users who have this article in their library.

Abstract

There are numerous characterizations of Shannon entropy and Tsallis entropy as measures of information obeying certain properties. Using work by Faddeev and Furuichi, we derive a very simple characterization. Instead of focusing on the entropy of a probability measure on a finite set, this characterization focuses on the "information loss", or change in entropy, associated with a measure-preserving function. Information loss is a special case of conditional entropy: namely, it is the entropy of a random variable conditioned on some function of that variable. We show that Shannon entropy gives the only concept of information loss that is functorial, convex-linear and continuous. This characterization naturally generalizes to Tsallis entropy as well. © 2011 by the authors.

Cite

CITATION STYLE

APA

Baez, J. C., Fritz, T., & Leinster, T. (2011). A characterization of entropy in terms of information loss. Entropy, 13(11), 1945–1957. https://doi.org/10.3390/e13111945

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free