LDTR: Transformer-based lane detection with anchor-chain representation

9Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Despite recent advances in lane detection methods, scenarios with limited- or no-visual-clue of lanes due to factors such as lighting conditions and occlusion remain challenging and crucial for automated driving. Moreover, current lane representations require complex post-processing and struggle with specific instances. Inspired by the DETR architecture, we propose LDTR, a transformer-based model to address these issues. Lanes are modeled with a novel anchor-chain, regarding a lane as a whole from the beginning, which enables LDTR to handle special lanes inherently. To enhance lane instance perception, LDTR incorporates a novel multi-referenced deformable attention module to distribute attention around the object. Additionally, LDTR incorporates two line IoU algorithms to improve convergence efficiency and employs a Gaussian heatmap auxiliary branch to enhance model representation capability during training. To evaluate lane detection models, we rely on Fréchet distance, parameterized Fl-score, and additional synthetic metrics. Experimental results demonstrate that LDTR achieves state-of-the-art performance on well-known datasets.

Cite

CITATION STYLE

APA

Yang, Z., Shen, C., Shao, W., Xing, T., Hu, R., Xu, P., … Xue, R. (2024). LDTR: Transformer-based lane detection with anchor-chain representation. Computational Visual Media, 10(4), 753–769. https://doi.org/10.1007/s41095-024-0421-5

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free