U-Attention to Textures: Hierarchical Hourglass Vision Transformer for Universal Texture Synthesis

6Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present a novel U-Attention vision Transformer for universal texture synthesis. We exploit the natural long-range dependencies enabled by the attention mechanism to allow our approach to synthesize diverse textures while preserving their structures in a single inference. We propose a hierarchical hourglass backbone that attends to the global structure and performs patch mapping at varying scales in a coarse-to-fine-to-coarse stream. Completed by skip connection and convolution designs that propagate and fuse information at different scales, our hierarchical U-Attention architecture unifies attention to features from macro structures to micro details, and progressively refines synthesis results at successive stages. Our method achieves stronger 2 × synthesis than previous work on both stochastic and structured textures while generalizing to unseen textures without fine-tuning. Ablation studies demonstrate the effectiveness of each component of our architecture.

Cite

CITATION STYLE

APA

Guo, S., Deschaintre, V., Noll, D., & Roullier, A. (2022). U-Attention to Textures: Hierarchical Hourglass Vision Transformer for Universal Texture Synthesis. In Proceedings - CVMP 2022: 19th ACM SIGGRAPH European Conference on Visual Media Production. Association for Computing Machinery, Inc. https://doi.org/10.1145/3565516.3565525

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free