Abstract
We present an approach to multiscale image analysis. It hinges on an operative definition of texture that involves a "small region", where some (unknown) statistic is aggregated, and a "large region" within which it is stationary. At each point, multiple small and large regions co-exist at multiple scales, as image structures are pooled by the scaling and quantization process to form "textures" and then transitions between textures define again "structures." We present a technique to learn and agglomerate sparse bases at multiple scales. To do so efficiently, we propose an analysis of cluster statistics after a clustering step is performed, and a new clustering method with linear-time performance. In both cases, we can infer all the "small" and "large" regions at multiple scale in one shot. © 2010 Springer-Verlag.
Cite
CITATION STYLE
Boltz, S., Nielsen, F., & Soatto, S. (2010). Texture regimes for entropy-based multiscale image analysis. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 6313 LNCS, pp. 692–705). Springer Verlag. https://doi.org/10.1007/978-3-642-15558-1_50
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.