Laneformer: Object-aware Row-Column Transformers for Lane Detection

by   Jianhua Han, et al.

We present Laneformer, a conceptually simple yet powerful transformer-based architecture tailored for lane detection that is a long-standing research topic for visual perception in autonomous driving. The dominant paradigms rely on purely CNN-based architectures which often fail in incorporating relations of long-range lane points and global contexts induced by surrounding objects (e.g., pedestrians, vehicles). Inspired by recent advances of the transformer encoder-decoder architecture in various vision tasks, we move forwards to design a new end-to-end Laneformer architecture that revolutionizes the conventional transformers into better capturing the shape and semantic characteristics of lanes, with minimal overhead in latency. First, coupling with deformable pixel-wise self-attention in the encoder, Laneformer presents two new row and column self-attention operations to efficiently mine point context along with the lane shapes. Second, motivated by the appearing objects would affect the decision of predicting lane segments, Laneformer further includes the detected object instances as extra inputs of multi-head attention blocks in the encoder and decoder to facilitate the lane point detection by sensing semantic contexts. Specifically, the bounding box locations of objects are added into Key module to provide interaction with each pixel and query while the ROI-aligned features are inserted into Value module. Extensive experiments demonstrate our Laneformer achieves state-of-the-art performances on CULane benchmark, in terms of 77.1 effective Laneformer will serve as a strong baseline for future research in self-attention models for lane detection.


page 1

page 3

page 7

page 10


Robust Lane Detection via Expanded Self Attention

The image-based lane detection algorithm is one of the key technologies ...

Bottleneck Transformers for Visual Recognition

We present BoTNet, a conceptually simple yet powerful backbone architect...

End-to-end Lane Shape Prediction with Transformers

Lane detection, the process of identifying lane markings as approximated...

DAT++: Spatially Dynamic Vision Transformer with Deformable Attention

Transformers have shown superior performance on various vision tasks. Th...

Defect Transformer: An Efficient Hybrid Transformer Architecture for Surface Defect Detection

Surface defect detection is an extremely crucial step to ensure the qual...

Lane Detection with Versatile AtrousFormer and Local Semantic Guidance

Lane detection is one of the core functions in autonomous driving and ha...

Transformer-based models and hardware acceleration analysis in autonomous driving: A survey

Transformer architectures have exhibited promising performance in variou...

Please sign up or login with your details

Forgot password? Click here to reset