Video question answering aims at answering a question about the video content by reasoning the alignment semantics within them. However, since relying heavily on human instructions, i.e., annotations or priors, current contrastive learning-based VideoQA methods remains challenging to perform fine-grained visual-linguistic alignments. In this work, we innovatively resort to game theory, which can simulate complicated relationships among multiple players with specific interaction strategies, e.g., video, question, and answer as ternary players, to achieve fine-grained alignment for VideoQA task. Specifically, we carefully design a VideoQA-specific interaction strategy to tailor the characteristics of VideoQA, which can mathematically generate the fine-grained visual-linguistic alignment label without label-intensive efforts. Our TG-VQA outperforms existing state-of-the-art by a large margin (more than 5%) on long-term and short-term VideoQA datasets, verifying its effectiveness and generalization ability. Thanks to the guidance of game-theoretic interaction, our model impressively convergences well on limited data (104 videos), surpassing most of those pre-trained on large-scale data (107 videos).
CITATION STYLE
Li, H., Jin, P., Cheng, Z., Zhang, S., Chen, K., Wang, Z., … Chen, J. (2023). TG-VQA: Ternary Game of Video Question Answering. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2023-August, pp. 1044–1052). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2023/116
Mendeley helps you to discover research relevant for your work.