Video-based self-positioning for intelligent transportation systems applications

0Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Many urban areas face traffic congestion. Automatic traffic management systems and congestion pricing are getting prominence in recent research. An important stage in such systems is lane prediction and on-road self-positioning. We introduce a novel problem of vehicle self-positioning which involves predicting the number of lanes on the road and localizing the vehicle within those lanes, using the video captured by a dashboard camera. To overcome the disadvantages of most existing low-level vision-based techniques while tackling this complex problem, we formulate a model in which the video is a key observation. The model consists of the number of lanes and vehicle position in those lanes as parameters, hence allowing the use of high-level semantic knowledge. Under this formulation, we employ a lane-width-based model and a maximum-likelihoodestimator making the method tolerant to slight viewing angle variation. The overall approach is tested on real-world videos and is found to be effective.

Cite

CITATION STYLE

APA

Chandakkar, P. S., Venkatesan, R., & Li, B. (2014). Video-based self-positioning for intelligent transportation systems applications. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8887, pp. 718–729). Springer Verlag. https://doi.org/10.1007/978-3-319-14249-4_69

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free