TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Embracing Single Stride 3D Object Detector with Sparse Tra...

Embracing Single Stride 3D Object Detector with Sparse Transformer

Lue Fan, Ziqi Pang, Tianyuan Zhang, Yu-Xiong Wang, Hang Zhao, Feng Wang, Naiyan Wang, Zhaoxiang Zhang

2021-12-13CVPR 2022 1Autonomous DrivingPedestrian Detectionobject-detection3D Object DetectionObject Detection
PaperPDFCodeCode(official)

Abstract

In LiDAR-based 3D object detection for autonomous driving, the ratio of the object size to input scene size is significantly smaller compared to 2D detection cases. Overlooking this difference, many 3D detectors directly follow the common practice of 2D detectors, which downsample the feature maps even after quantizing the point clouds. In this paper, we start by rethinking how such multi-stride stereotype affects the LiDAR-based 3D object detectors. Our experiments point out that the downsampling operations bring few advantages, and lead to inevitable information loss. To remedy this issue, we propose Single-stride Sparse Transformer (SST) to maintain the original resolution from the beginning to the end of the network. Armed with transformers, our method addresses the problem of insufficient receptive field in single-stride architectures. It also cooperates well with the sparsity of point clouds and naturally avoids expensive computation. Eventually, our SST achieves state-of-the-art results on the large scale Waymo Open Dataset. It is worth mentioning that our method can achieve exciting performance (83.8 LEVEL 1 AP on validation split) on small object (pedestrian) detection due to the characteristic of single stride. Codes will be released at https://github.com/TuSimple/SST

Results

TaskDatasetMetricValueModel
Object Detectionwaymo cyclistAPH/L272.17SST
Object Detectionwaymo vehicleAPH/L272.74SST
Object Detectionwaymo pedestrianAPH/L273.51SST
3Dwaymo cyclistAPH/L272.17SST
3Dwaymo vehicleAPH/L272.74SST
3Dwaymo pedestrianAPH/L273.51SST
3D Object Detectionwaymo cyclistAPH/L272.17SST
3D Object Detectionwaymo vehicleAPH/L272.74SST
3D Object Detectionwaymo pedestrianAPH/L273.51SST
2D Classificationwaymo cyclistAPH/L272.17SST
2D Classificationwaymo vehicleAPH/L272.74SST
2D Classificationwaymo pedestrianAPH/L273.51SST
2D Object Detectionwaymo cyclistAPH/L272.17SST
2D Object Detectionwaymo vehicleAPH/L272.74SST
2D Object Detectionwaymo pedestrianAPH/L273.51SST
16kwaymo cyclistAPH/L272.17SST
16kwaymo vehicleAPH/L272.74SST
16kwaymo pedestrianAPH/L273.51SST

Related Papers

GEMINUS: Dual-aware Global and Scene-Adaptive Mixture-of-Experts for End-to-End Autonomous Driving2025-07-19AGENTS-LLM: Augmentative GENeration of Challenging Traffic Scenarios with an Agentic LLM Framework2025-07-18World Model-Based End-to-End Scene Generation for Accident Anticipation in Autonomous Driving2025-07-17Orbis: Overcoming Challenges of Long-Horizon Prediction in Driving World Models2025-07-17Channel-wise Motion Features for Efficient Motion Segmentation2025-07-17LaViPlan : Language-Guided Visual Path Planning with RLVR2025-07-17A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17RS-TinyNet: Stage-wise Feature Fusion Network for Detecting Tiny Objects in Remote Sensing Images2025-07-17