TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/State Space Models for Event Cameras

State Space Models for Event Cameras

Nikola Zubić, Mathias Gehrig, Davide Scaramuzza

2024-02-23CVPR 2024 1Event-based visionObject Detection
PaperPDFCodeCode(official)

Abstract

Today, state-of-the-art deep neural networks that process event-camera data first convert a temporal window of events into dense, grid-like input representations. As such, they exhibit poor generalizability when deployed at higher inference frequencies (i.e., smaller temporal windows) than the ones they were trained on. We address this challenge by introducing state-space models (SSMs) with learnable timescale parameters to event-based vision. This design adapts to varying frequencies without the need to retrain the network at different frequencies. Additionally, we investigate two strategies to counteract aliasing effects when deploying the model at higher frequencies. We comprehensively evaluate our approach against existing methods based on RNN and Transformer architectures across various benchmarks, including Gen1 and 1 Mpx event camera datasets. Our results demonstrate that SSM-based models train 33% faster and also exhibit minimal performance degradation when tested at higher frequencies than the training input. Traditional RNN and Transformer models exhibit performance drops of more than 20 mAP, with SSMs having a drop of 3.76 mAP, highlighting the effectiveness of SSMs in event-based vision tasks.

Results

TaskDatasetMetricValueModel
Object DetectionGEN1 DetectionParams18.2S5-ViT-B
Object DetectionGEN1 DetectionmAP47.4S5-ViT-B
Object DetectionGEN1 DetectionParams16.5S4D-ViT-B
Object DetectionGEN1 DetectionmAP46.2S4D-ViT-B
3DGEN1 DetectionParams18.2S5-ViT-B
3DGEN1 DetectionmAP47.4S5-ViT-B
3DGEN1 DetectionParams16.5S4D-ViT-B
3DGEN1 DetectionmAP46.2S4D-ViT-B
2D ClassificationGEN1 DetectionParams18.2S5-ViT-B
2D ClassificationGEN1 DetectionmAP47.4S5-ViT-B
2D ClassificationGEN1 DetectionParams16.5S4D-ViT-B
2D ClassificationGEN1 DetectionmAP46.2S4D-ViT-B
2D Object DetectionGEN1 DetectionParams18.2S5-ViT-B
2D Object DetectionGEN1 DetectionmAP47.4S5-ViT-B
2D Object DetectionGEN1 DetectionParams16.5S4D-ViT-B
2D Object DetectionGEN1 DetectionmAP46.2S4D-ViT-B
16kGEN1 DetectionParams18.2S5-ViT-B
16kGEN1 DetectionmAP47.4S5-ViT-B
16kGEN1 DetectionParams16.5S4D-ViT-B
16kGEN1 DetectionmAP46.2S4D-ViT-B

Related Papers

A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17RS-TinyNet: Stage-wise Feature Fusion Network for Detecting Tiny Objects in Remote Sensing Images2025-07-17Decoupled PROB: Decoupled Query Initialization Tasks and Objectness-Class Learning for Open World Object Detection2025-07-17Dual LiDAR-Based Traffic Movement Count Estimation at a Signalized Intersection: Deployment, Data Collection, and Preliminary Analysis2025-07-17Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16Tomato Multi-Angle Multi-Pose Dataset for Fine-Grained Phenotyping2025-07-15ECORE: Energy-Conscious Optimized Routing for Deep Learning Models at the Edge2025-07-08Beyond One Shot, Beyond One Perspective: Cross-View and Long-Horizon Distillation for Better LiDAR Representations2025-07-07