TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/High Speed and High Dynamic Range Video with an Event Camera

High Speed and High Dynamic Range Video with an Event Camera

Henri Rebecq, René Ranftl, Vladlen Koltun, Davide Scaramuzza

2019-06-15Vocal Bursts Intensity PredictionEvent-based Object SegmentationVideo Reconstruction
PaperPDFCode

Abstract

Event cameras are novel sensors that report brightness changes in the form of a stream of asynchronous "events" instead of intensity frames. They offer significant advantages with respect to conventional cameras: high temporal resolution, high dynamic range, and no motion blur. While the stream of events encodes in principle the complete visual signal, the reconstruction of an intensity image from a stream of events is an ill-posed problem in practice. Existing reconstruction approaches are based on hand-crafted priors and strong assumptions about the imaging process as well as the statistics of natural images. In this work we propose to learn to reconstruct intensity images from event streams directly from data instead of relying on any hand-crafted priors. We propose a novel recurrent network to reconstruct videos from a stream of events, and train it on a large amount of simulated event data. During training we propose to use a perceptual loss to encourage reconstructions to follow natural image statistics. We further extend our approach to synthesize color images from color event streams. Our network surpasses state-of-the-art reconstruction methods by a large margin in terms of image quality (> 20%), while comfortably running in real-time. We show that the network is able to synthesize high framerate videos (> 5,000 frames per second) of high-speed phenomena (e.g. a bullet hitting an object) and is able to provide high dynamic range reconstructions in challenging lighting conditions. We also demonstrate the effectiveness of our reconstructions as an intermediate representation for event data. We show that off-the-shelf computer vision algorithms can be applied to our reconstructions for tasks such as object classification and visual-inertial odometry and that this strategy consistently outperforms algorithms that were specifically designed for event data.

Results

TaskDatasetMetricValueModel
3DMVSECLPIPS0.644E2VID
3DMVSECMean Squared Error0.225E2VID
3DEvent-Camera DatasetLPIPS0.322E2VID
3DEvent-Camera DatasetMean Squared Error0.179E2VID
Video ReconstructionMVSECLPIPS0.644E2VID
Video ReconstructionMVSECMean Squared Error0.225E2VID
Video ReconstructionEvent-Camera DatasetLPIPS0.322E2VID
Video ReconstructionEvent-Camera DatasetMean Squared Error0.179E2VID
Event-based Object SegmentationMVSEC-SEGmIoU0.35E2VID
Event-based Object SegmentationRGBE-SEGmIoU0.32E2VID

Related Papers

GSVR: 2D Gaussian-based Video Representation for 800+ FPS with Hybrid Deformation Field2025-07-08Quanta Diffusion2025-06-07Voyager: Long-Range and World-Consistent Video Diffusion for Explorable 3D Scene Generation2025-06-04Compressing Human Body Video with Interactive Semantics: A Generative Approach2025-05-22Motion Matters: Compact Gaussian Streaming for Free-Viewpoint Video Reconstruction2025-05-22V2V: Scaling Event-Based Vision through Efficient Video-to-Voxel Simulation2025-05-22Learning Adaptive and Temporally Causal Video Tokenization in a 1D Latent Space2025-05-22Few-shot Semantic Encoding and Decoding for Video Surveillance2025-05-12