Attribute-Based Progressive Fusion Network for RGBT Tracking

125Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.

Abstract

RGBT tracking usually suffers from various challenging factors of fast motion, scale variation, illumination variation, thermal crossover and occlusion, to name a few. Existing works often study fusion models to solve all challenges simultaneously, which requires fusion models complex enough and training data large enough, and are usually difficult to be constructed in real-world scenarios. In this work, we disentangle the fusion process via the challenge attributes, and thus propose a novel Attribute-Based Progressive Fusion Network (APFNet) to increase the fusion capacity with a small number of parameters while reducing the dependence on large-scale training data. In particular, we design five attribute-specific fusion branches to integrate RGB and thermal features under the challenges of thermal crossover, illumination variation, scale variation, occlusion and fast motion respectively. By disentangling the fusion process, we can use a small number of parameters for each branch to achieve robust fusion of different modalities and train each branch using the small training subset with the corresponding attribute annotation. Then, to adaptive fuse features of all branches, we design an aggregation fusion module based on SKNet. Finally, we also design an enhancement fusion transformer to strengthen the aggregated feature and modality-specific features. Experimental results on benchmark datasets demonstrate the effectiveness of our APFNet against other state-of-the-art methods.

References Powered by Scopus

Selective kernel networks

2594Citations
N/AReaders
Get full text

Learning Multi-domain Convolutional Neural Networks for Visual Tracking

2489Citations
N/AReaders
Get full text

Learning discriminative model prediction for tracking

1156Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Visual Prompt Multi-Modal Tracking

159Citations
N/AReaders
Get full text

ICAFusion: Iterative cross-attention guided feature fusion for multispectral object detection

76Citations
N/AReaders
Get full text

Bridging Search Region Interaction with Template for RGB-T Tracking

60Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Xiao, Y., Yang, M., Li, C., Liu, L., & Tang, J. (2022). Attribute-Based Progressive Fusion Network for RGBT Tracking. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022 (Vol. 36, pp. 2831–2838). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v36i3.20187

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 8

80%

Professor / Associate Prof. 1

10%

Researcher 1

10%

Readers' Discipline

Tooltip

Engineering 5

56%

Computer Science 3

33%

Neuroscience 1

11%

Save time finding and organizing research with Mendeley

Sign up for free