XVTP3D: Cross-view Trajectory Prediction Using Shared 3D Queries for Autonomous Driving

0Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.

Abstract

Trajectory prediction with uncertainty is a critical and challenging task for autonomous driving. Nowadays, we can easily access sensor data represented in multiple views. However, cross-view consistency has not been evaluated by the existing models, which might lead to divergences between the multimodal predictions from different views. It is not practical and effective when the network does not comprehend the 3D scene, which could cause the downstream module in a dilemma. Instead, we predicts multimodal trajectories while maintaining cross-view consistency. We presented a cross-view trajectory prediction method using shared 3D Queries (XVTP3D). We employ a set of 3D queries shared across views to generate multi-goals that are cross-view consistent. We also proposed a random mask method and coarse-to-fine cross-attention to capture robust cross-view features. As far as we know, this is the first work that introduces the outstanding top-down paradigm in BEV detection field to a trajectory prediction problem. The results of experiments on two publicly available datasets show that XVTP3D achieved state-of-the-art performance with consistent cross-view predictions.

Cite

CITATION STYLE

APA

Song, Z., Bi, H., Zhang, R., Mao, T., & Wang, Z. (2023). XVTP3D: Cross-view Trajectory Prediction Using Shared 3D Queries for Autonomous Driving. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2023-August, pp. 298–308). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2023/34

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free