TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Robust Lane Detection via Expanded Self Attention

Robust Lane Detection via Expanded Self Attention

Minhyeok Lee, Junhyeop Lee, Dogyoon Lee, Woojin Kim, Sangwon Hwang, Sangyoun Lee

2021-02-14Autonomous VehiclesLane Detection
PaperPDFCode(official)

Abstract

The image-based lane detection algorithm is one of the key technologies in autonomous vehicles. Modern deep learning methods achieve high performance in lane detection, but it is still difficult to accurately detect lanes in challenging situations such as congested roads and extreme lighting conditions. To be robust on these challenging situations, it is important to extract global contextual information even from limited visual cues. In this paper, we propose a simple but powerful self-attention mechanism optimized for lane detection called the Expanded Self Attention (ESA) module. Inspired by the simple geometric structure of lanes, the proposed method predicts the confidence of a lane along the vertical and horizontal directions in an image. The prediction of the confidence enables estimating occluded locations by extracting global contextual information. ESA module can be easily implemented and applied to any encoder-decoder-based model without increasing the inference time. The performance of our method is evaluated on three popular lane detection benchmarks (TuSimple, CULane and BDD100K). We achieve state-of-the-art performance in CULane and BDD100K and distinct improvement on TuSimple dataset. The experimental results show that our approach is robust to occlusion and extreme lighting conditions.

Results

TaskDatasetMetricValueModel
Autonomous VehiclesCULaneF1 score74.2ERFNet-H&VESA
Lane DetectionCULaneF1 score74.2ERFNet-H&VESA

Related Papers

Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16Geo-ORBIT: A Federated Digital Twin Framework for Scene-Adaptive Lane Geometry Detection2025-07-11Fast and Accurate Collision Probability Estimation for Autonomous Vehicles using Adaptive Sigma-Point Sampling2025-07-08Robustifying 3D Perception through Least-Squares Multi-Agent Graphs Object Tracking2025-07-07LLM-based Realistic Safety-Critical Driving Video Generation2025-07-02A Survey on Vision-Language-Action Models for Autonomous Driving2025-06-30Where, What, Why: Towards Explainable Driver Attention Prediction2025-06-29Coordinated Control of Autonomous Vehicles for Traffic Density Reduction at a Signalized Junction: An MPC Approach2025-06-26