We regard a data pattern as a physical particle experiencing a force acting on it imposed by an overall "potential energy" of the data set, obtained via a non-parametric estimate of Renyi's entropy. The "potential energy" is called the information potential, and the forces are called information forces, due to their information-theoretic origin. We create directed trees by selecting the predecessor of a node (pattern) according to the direction of the information force acting on the pattern. Each directed tree correspond to a cluster, hence enabling us to partition the data set. The clustering metric underlying our method is thus based on entropy, which is a quantity that conveys information about the shape of a probability density, and not only it's variance, as many traditional algorithms based on mere second order statistics rely on. We demonstrate the performance of our clustering technique when applied to both artificially created data and real data, and also discuss some limitations of the proposed method. © Springer-Verlag Berlin Heidelberg 2003.
CITATION STYLE
Jenssen, R., Erdogmus, D., Hild, K. E., Principe, J. C., & Eltoft, T. (2003). Information force clustering using directed trees. Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2683, 68–82. https://doi.org/10.1007/978-3-540-45063-4_5
Mendeley helps you to discover research relevant for your work.