The visual prompts have provided an efficient manner in addressing visual cross-domain problems. In previous works, (Gan et al. 2022a) first introduces domain prompts to tackle the classification Test-Time Adaptation (TTA) problem by placing image-level prompts on the input and fine-tuning prompts for each target domain. However, since the image-level prompts mask out continuous spatial details in the prompt-allocated region, it will suffer from inaccurate contextual information and limited domain knowledge extraction, particularly when dealing with dense prediction TTA problems. To overcome these challenges, we propose a novel Sparse Visual Domain Prompts (SVDP) approach, which applies minimal trainable parameters (e.g., 0.1%) to pixels across the entire image and reserves more spatial information of the input. To better apply SVDP in extracting domain-specific knowledge, we introduce the Domain Prompt Placement (DPP) method to adaptively allocates trainable parameters of SVDP on the pixels with large distribution shifts. Furthermore, recognizing that each target domain sample exhibits a unique domain shift, we design Domain Prompt Updating (DPU) strategy to optimize prompt parameters differently for each sample, facilitating efficient adaptation to the target domain. Extensive experiments were conducted on widely-used TTA and continual TTA benchmarks, and our proposed method achieves state-of-the-art performance in both semantic segmentation and depth estimation tasks.
CITATION STYLE
Yang, S., Wu, J., Liu, J., Li, X., Zhang, Q., Pan, M., … Zhang, S. (2024). Exploring Sparse Visual Prompt for Domain Adaptive Dense Prediction. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 38, pp. 16334–16342). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v38i15.29569
Mendeley helps you to discover research relevant for your work.