TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Technical Report of the Video Event Reconstruction and Ana...

Technical Report of the Video Event Reconstruction and Analysis (VERA) System -- Shooter Localization, Models, Interface, and Beyond

Junwei Liang, Jay D. Aronson, Alexander Hauptmann

2019-05-26Shooter LocalizationTemporal LocalizationGunshot DetectionVideo Synchronization
PaperPDFCode(official)Code

Abstract

Every minute, hundreds of hours of video are uploaded to social media sites and the Internet from around the world. This material creates a visual record of the experiences of a significant percentage of humanity and can help illuminate how we live in the present moment. When properly analyzed, this video can also help analysts to reconstruct events of interest, including war crimes, human rights violations, and terrorist acts. Machine learning and computer vision can play a crucial role in this process. In this technical report, we describe the Video Event Reconstruction and Analysis (VERA) system. This new tool brings together a variety of capabilities we have developed over the past few years (including video synchronization and geolocation to order unstructured videos lacking metadata over time and space, and sound recognition algorithms) to enable the reconstruction and analysis of events captured on video. Among other uses, VERA enables the localization of a shooter from just a few videos that include the sound of gunshots. To demonstrate the efficacy of this suite of tools, we present the results of estimating the shooter's location of the Las Vegas Shooting in 2017 and show that VERA accurately predicts the shooter's location using only the first few gunshots. We then point out future directions that can help improve the system and further reduce unnecessary human labor in the process. All of the components of VERA run through a web interface that enables human-in-the-loop verification to ensure accurate estimations. All relevant source code, including the web interface and machine learning models, is freely available on Github. We hope that researchers and software developers will be inspired to improve and expand this system moving forward to better meet the needs of human rights and public safety.

Related Papers

Deciphering GunType Hierarchy through Acoustic Analysis of Gunshot Recordings2025-06-25Beyond Audio and Pose: A General-Purpose Framework for Video Synchronization2025-06-19Fine-Tuning Large Audio-Language Models with LoRA for Precise Temporal Localization of Prolonged Exposure Therapy Elements2025-06-11VideoMolmo: Spatio-Temporal Grounding Meets Pointing2025-06-05DisTime: Distribution-based Time Representation for Video Large Language Models2025-05-30Transforming faces into video stories -- VideoFace2.02025-05-04MINERVA: Evaluating Complex Video Reasoning2025-05-01AlignDiT: Multimodal Aligned Diffusion Transformer for Synchronized Speech Generation2025-04-29