TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A Deep Moving-camera Background Model

A Deep Moving-camera Background Model

Guy Erez, Ron Shapira Weber, Oren Freifeld

2022-09-16Change Detection
PaperPDFCode(official)

Abstract

In video analysis, background models have many applications such as background/foreground separation, change detection, anomaly detection, tracking, and more. However, while learning such a model in a video captured by a static camera is a fairly-solved task, in the case of a Moving-camera Background Model (MCBM), the success has been far more modest due to algorithmic and scalability challenges that arise due to the camera motion. Thus, existing MCBMs are limited in their scope and their supported camera-motion types. These hurdles also impeded the employment, in this unsupervised task, of end-to-end solutions based on deep learning (DL). Moreover, existing MCBMs usually model the background either on the domain of a typically-large panoramic image or in an online fashion. Unfortunately, the former creates several problems, including poor scalability, while the latter prevents the recognition and leveraging of cases where the camera revisits previously-seen parts of the scene. This paper proposes a new method, called DeepMCBM, that eliminates all the aforementioned issues and achieves state-of-the-art results. Concretely, first we identify the difficulties associated with joint alignment of video frames in general and in a DL setting in particular. Next, we propose a new strategy for joint alignment that lets us use a spatial transformer net with neither a regularization nor any form of specialized (and non-differentiable) initialization. Coupled with an autoencoder conditioned on unwarped robust central moments (obtained from the joint alignment), this yields an end-to-end regularization-free MCBM that supports a broad range of camera motions and scales gracefully. We demonstrate DeepMCBM's utility on a variety of videos, including ones beyond the scope of other methods. Our code is available at https://github.com/BGU-CS-VIL/DeepMCBM .

Results

TaskDatasetMetricValueModel
VideoDAVIS 2017 (stroller)AUC0.877DeepMCBM (Basic/Aff)
VideoDAVIS 2017 (dog-gooses)AUC0.984DeepMCBM (CAE/Aff)
VideoDAVIS 2017 (dog-gooses)AUC0.984DeepMCBM (CAE/Hom)
VideoDAVIS 2017 (continuousPan)AUC0.94DeepMCBM (CAE/Aff)
VideoDAVIS 2017 (zoomInZoomOut)AUC0.994DeepMCBM (CAE/Aff)
VideoDAVIS 2017 (zoomInZoomOut)AUC0.994DeepMCBM (CAE/Hom)
VideoDAVIS 2017 (stunt)AUC0.979DeepMCBM (CAE/Aff)
VideoDAVIS 2017 (tennis)AUC0.963DeepMCBM (CAE/Hom)
VideoDAVIS 2017 (swing)AUC0.897DeepMCBM (CAE/Hom)
VideoDAVIS 2017 (flamingo)AUC0.98DeepMCBM (CAE/Aff)
VideoDAVIS 2017 (flamingo)AUC0.98DeepMCBM (CAE/Hom)
VideoDAVIS 2017 (bmx-trees)AUC0.916DeepMCBM (Basic/Hom)
VideoDAVIS 2017 (horsejump-high)AUC0.943DeepMCBM (CAE/Hom)
VideoDAVIS 2017 (boxing-fisheye)AUC0.927DeepMCBM (Basic/Hom)
VideoDAVIS 2017 (sidewalk)AUC0.932DeepMCBM (CAE/Hom)
VideoDAVIS 2017 (breakdance-flare)AUC0.963DeepMCBM (CAE/Hom)

Related Papers

Precision Spatio-Temporal Feature Fusion for Robust Remote Sensing Change Detection2025-07-15Be the Change You Want to See: Revisiting Remote Sensing Change Detection Practices2025-07-04Be the Change You Want to See: Revisiting Remote Sensing Change Detection Practices2025-07-04Pushing Trade-Off Boundaries: Compact yet Effective Remote Sensing Change Detection2025-06-26CL-Splats: Continual Learning of Gaussian Splatting with Local Optimization2025-06-26HydroChronos: Forecasting Decades of Surface Water Change2025-06-17Active InSAR monitoring of building damage in Gaza during the Israel-Hamas War2025-06-17Revisiting Clustering of Neural Bandits: Selective Reinitialization for Mitigating Loss of Plasticity2025-06-14