TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/OPV2V: An Open Benchmark Dataset and Fusion Pipeline for P...

OPV2V: An Open Benchmark Dataset and Fusion Pipeline for Perception with Vehicle-to-Vehicle Communication

Runsheng Xu, Hao Xiang, Xin Xia, Xu Han, Jinlong Li, Jiaqi Ma

2021-09-16Benchmarking3D Object Detection
PaperPDFCodeCode(official)

Abstract

Employing Vehicle-to-Vehicle communication to enhance perception performance in self-driving technology has attracted considerable attention recently; however, the absence of a suitable open dataset for benchmarking algorithms has made it difficult to develop and assess cooperative perception technologies. To this end, we present the first large-scale open simulated dataset for Vehicle-to-Vehicle perception. It contains over 70 interesting scenes, 11,464 frames, and 232,913 annotated 3D vehicle bounding boxes, collected from 8 towns in CARLA and a digital town of Culver City, Los Angeles. We then construct a comprehensive benchmark with a total of 16 implemented models to evaluate several information fusion strategies~(i.e. early, late, and intermediate fusion) with state-of-the-art LiDAR detection algorithms. Moreover, we propose a new Attentive Intermediate Fusion pipeline to aggregate information from multiple connected vehicles. Our experiments show that the proposed pipeline can be easily integrated with existing 3D LiDAR detectors and achieve outstanding performance even with large compression rates. To encourage more researchers to investigate Vehicle-to-Vehicle perception, we will release the dataset, benchmark methods, and all related codes in https://mobility-lab.seas.ucla.edu/opv2v/.

Results

TaskDatasetMetricValueModel
Object DetectionOPV2VAP@0.7@CulverCity0.735Attentive Fusion (PointPillar backbone)
Object DetectionOPV2VAP@0.7@Default0.815Attentive Fusion (PointPillar backbone)
Object DetectionOPV2VAP@0.7@CulverCity0.669Late Fusion (PointPillar backbone)
Object DetectionOPV2VAP@0.7@Default0.781Late Fusion (PointPillar backbone)
Object DetectionV2XSetAP0.5 (Noisy)0.709AttentiveFusion
Object DetectionV2XSetAP0.5 (Perfect)0.807AttentiveFusion
Object DetectionV2XSetAP0.7 (Noisy)0.487AttentiveFusion
Object DetectionV2XSetAP0.7 (Perfect)0.664AttentiveFusion
3DOPV2VAP@0.7@CulverCity0.735Attentive Fusion (PointPillar backbone)
3DOPV2VAP@0.7@Default0.815Attentive Fusion (PointPillar backbone)
3DOPV2VAP@0.7@CulverCity0.669Late Fusion (PointPillar backbone)
3DOPV2VAP@0.7@Default0.781Late Fusion (PointPillar backbone)
3DV2XSetAP0.5 (Noisy)0.709AttentiveFusion
3DV2XSetAP0.5 (Perfect)0.807AttentiveFusion
3DV2XSetAP0.7 (Noisy)0.487AttentiveFusion
3DV2XSetAP0.7 (Perfect)0.664AttentiveFusion
3D Object DetectionOPV2VAP@0.7@CulverCity0.735Attentive Fusion (PointPillar backbone)
3D Object DetectionOPV2VAP@0.7@Default0.815Attentive Fusion (PointPillar backbone)
3D Object DetectionOPV2VAP@0.7@CulverCity0.669Late Fusion (PointPillar backbone)
3D Object DetectionOPV2VAP@0.7@Default0.781Late Fusion (PointPillar backbone)
3D Object DetectionV2XSetAP0.5 (Noisy)0.709AttentiveFusion
3D Object DetectionV2XSetAP0.5 (Perfect)0.807AttentiveFusion
3D Object DetectionV2XSetAP0.7 (Noisy)0.487AttentiveFusion
3D Object DetectionV2XSetAP0.7 (Perfect)0.664AttentiveFusion
2D ClassificationOPV2VAP@0.7@CulverCity0.735Attentive Fusion (PointPillar backbone)
2D ClassificationOPV2VAP@0.7@Default0.815Attentive Fusion (PointPillar backbone)
2D ClassificationOPV2VAP@0.7@CulverCity0.669Late Fusion (PointPillar backbone)
2D ClassificationOPV2VAP@0.7@Default0.781Late Fusion (PointPillar backbone)
2D ClassificationV2XSetAP0.5 (Noisy)0.709AttentiveFusion
2D ClassificationV2XSetAP0.5 (Perfect)0.807AttentiveFusion
2D ClassificationV2XSetAP0.7 (Noisy)0.487AttentiveFusion
2D ClassificationV2XSetAP0.7 (Perfect)0.664AttentiveFusion
2D Object DetectionOPV2VAP@0.7@CulverCity0.735Attentive Fusion (PointPillar backbone)
2D Object DetectionOPV2VAP@0.7@Default0.815Attentive Fusion (PointPillar backbone)
2D Object DetectionOPV2VAP@0.7@CulverCity0.669Late Fusion (PointPillar backbone)
2D Object DetectionOPV2VAP@0.7@Default0.781Late Fusion (PointPillar backbone)
2D Object DetectionV2XSetAP0.5 (Noisy)0.709AttentiveFusion
2D Object DetectionV2XSetAP0.5 (Perfect)0.807AttentiveFusion
2D Object DetectionV2XSetAP0.7 (Noisy)0.487AttentiveFusion
2D Object DetectionV2XSetAP0.7 (Perfect)0.664AttentiveFusion
16kOPV2VAP@0.7@CulverCity0.735Attentive Fusion (PointPillar backbone)
16kOPV2VAP@0.7@Default0.815Attentive Fusion (PointPillar backbone)
16kOPV2VAP@0.7@CulverCity0.669Late Fusion (PointPillar backbone)
16kOPV2VAP@0.7@Default0.781Late Fusion (PointPillar backbone)
16kV2XSetAP0.5 (Noisy)0.709AttentiveFusion
16kV2XSetAP0.5 (Perfect)0.807AttentiveFusion
16kV2XSetAP0.7 (Noisy)0.487AttentiveFusion
16kV2XSetAP0.7 (Perfect)0.664AttentiveFusion

Related Papers

Visual Place Recognition for Large-Scale UAV Applications2025-07-20Training Transformers with Enforced Lipschitz Constants2025-07-17Disentangling coincident cell events using deep transfer learning and compressive sensing2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17Dual LiDAR-Based Traffic Movement Count Estimation at a Signalized Intersection: Deployment, Data Collection, and Preliminary Analysis2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16DCR: Quantifying Data Contamination in LLMs Evaluation2025-07-15A Multi-View High-Resolution Foot-Ankle Complex Point Cloud Dataset During Gait for Occlusion-Robust 3D Completion2025-07-15