UniNeXt: Exploring A Unified Architecture for Vision Recognition

12Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Vision Transformers have shown great potential in computer vision tasks. Most recent works have focused on elaborating the spatial token mixer for performance gains. However, we observe that a well-designed general architecture can significantly improve the performance of the entire backbone, regardless of which spatial token mixer is equipped. In this paper, we propose UniNeXt, an improved general architecture for the vision backbone. To verify its effectiveness, we instantiate the spatial token mixer with various typical and modern designs, including both convolution and attention modules. Compared with the architecture in which they are first proposed, our UniNeXt architecture can steadily boost the performance of all the spatial token mixers, and narrows the performance gap among them. Surprisingly, our UniNeXt equipped with naive local window attention even outperforms the previous state-of-the-art. Interestingly, the ranking of these spatial token mixers also changes under our UniNeXt, suggesting that an excellent spatial token mixer may be stifled due to a suboptimal general architecture, which further shows the importance of the study on the general architecture of vision backbone. Code is available at UniNeXt.

Cite

CITATION STYLE

APA

Lin, F., Yuan, J., Wu, S., Wang, F., & Wang, Z. (2023). UniNeXt: Exploring A Unified Architecture for Vision Recognition. In MM 2023 - Proceedings of the 31st ACM International Conference on Multimedia (pp. 3200–3208). Association for Computing Machinery, Inc. https://doi.org/10.1145/3581783.3612260

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free