TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/V2X-AHD:Vehicle-to-Everything Cooperation Perception via A...

V2X-AHD:Vehicle-to-Everything Cooperation Perception via Asymmetric Heterogenous Distillation Network

Caizhen He, Hai Wang, Long Chen, Tong Luo, Yingfeng Cai

2023-10-10object-detection3D Object DetectionObject Detection
PaperPDFCode(official)

Abstract

Object detection is the central issue of intelligent traffic systems, and recent advancements in single-vehicle lidar-based 3D detection indicate that it can provide accurate position information for intelligent agents to make decisions and plan. Compared with single-vehicle perception, multi-view vehicle-road cooperation perception has fundamental advantages, such as the elimination of blind spots and a broader range of perception, and has become a research hotspot. However, the current perception of cooperation focuses on improving the complexity of fusion while ignoring the fundamental problems caused by the absence of single-view outlines. We propose a multi-view vehicle-road cooperation perception system, vehicle-to-everything cooperative perception (V2X-AHD), in order to enhance the identification capability, particularly for predicting the vehicle's shape. At first, we propose an asymmetric heterogeneous distillation network fed with different training data to improve the accuracy of contour recognition, with multi-view teacher features transferring to single-view student features. While the point cloud data are sparse, we propose Spara Pillar, a spare convolutional-based plug-in feature extraction backbone, to reduce the number of parameters and improve and enhance feature extraction capabilities. Moreover, we leverage the multi-head self-attention (MSA) to fuse the single-view feature, and the lightweight design makes the fusion feature a smooth expression. The results of applying our algorithm to the massive open dataset V2Xset demonstrate that our method achieves the state-of-the-art result. The V2X-AHD can effectively improve the accuracy of 3D object detection and reduce the number of network parameters, according to this study, which serves as a benchmark for cooperative perception. The code for this article is available at https://github.com/feeling0414-lab/V2X-AHD.

Results

TaskDatasetMetricValueModel
Object DetectionV2XSetAP0.5 (Perfect)0.855V2X-AHD
Object DetectionV2XSetAP0.7 (Perfect)0.724V2X-AHD
3DV2XSetAP0.5 (Perfect)0.855V2X-AHD
3DV2XSetAP0.7 (Perfect)0.724V2X-AHD
3D Object DetectionV2XSetAP0.5 (Perfect)0.855V2X-AHD
3D Object DetectionV2XSetAP0.7 (Perfect)0.724V2X-AHD
2D ClassificationV2XSetAP0.5 (Perfect)0.855V2X-AHD
2D ClassificationV2XSetAP0.7 (Perfect)0.724V2X-AHD
2D Object DetectionV2XSetAP0.5 (Perfect)0.855V2X-AHD
2D Object DetectionV2XSetAP0.7 (Perfect)0.724V2X-AHD
16kV2XSetAP0.5 (Perfect)0.855V2X-AHD
16kV2XSetAP0.7 (Perfect)0.724V2X-AHD

Related Papers

A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17RS-TinyNet: Stage-wise Feature Fusion Network for Detecting Tiny Objects in Remote Sensing Images2025-07-17Decoupled PROB: Decoupled Query Initialization Tasks and Objectness-Class Learning for Open World Object Detection2025-07-17Dual LiDAR-Based Traffic Movement Count Estimation at a Signalized Intersection: Deployment, Data Collection, and Preliminary Analysis2025-07-17Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16Tomato Multi-Angle Multi-Pose Dataset for Fine-Grained Phenotyping2025-07-15ECORE: Energy-Conscious Optimized Routing for Deep Learning Models at the Edge2025-07-08Beyond One Shot, Beyond One Perspective: Cross-View and Long-Horizon Distillation for Better LiDAR Representations2025-07-07