M2P3: Multimodal multi-pedestrian path prediction by self-driving cars with egocentric vision

30Citations
Citations of this article
40Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Accurate prediction of the future position of pedestrians in traffic scenarios is required for safe navigation of an autonomous vehicle but remains a challenge. This concerns, in particular, the effective and efficient multimodal prediction of most likely trajectories of tracked pedestrians from egocentric view of self-driving car. In this paper, we present a novel solution, named M2P3, which combines a conditional variational autoencoder with recurrent neural network encoder-decoder architecture in order to predict a set of possible future locations of each pedestrian in a traffic scene. The M2P3 system uses a sequence of RGB images delivered through an internal vehicle-mounted camera for egocentric vision. It takes as an input only two modes, that are past trajectories and scales of pedestrians, and delivers as an output the three most likely paths for each tracked pedestrian. Experimental evaluation of the proposed architecture on the JAAD and ETH/UCY datasets reveal that the M2P3 system is significantly superior to selected state-of-the-art solutions.

Cite

CITATION STYLE

APA

Poibrenski, A., Klusch, M., Vozniak, I., & Müller, C. (2020). M2P3: Multimodal multi-pedestrian path prediction by self-driving cars with egocentric vision. In Proceedings of the ACM Symposium on Applied Computing (pp. 190–197). Association for Computing Machinery. https://doi.org/10.1145/3341105.3373877

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free