TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/How to Train Your Energy-Based Model for Regression

How to Train Your Energy-Based Model for Regression

Fredrik K. Gustafsson, Martin Danelljan, Radu Timofte, Thomas B. Schön

2020-05-04Visual Object TrackingVisual Trackingregressionobject-detectionObject Detection
PaperPDFCode(official)Code

Abstract

Energy-based models (EBMs) have become increasingly popular within computer vision in recent years. While they are commonly employed for generative image modeling, recent work has applied EBMs also for regression tasks, achieving state-of-the-art performance on object detection and visual tracking. Training EBMs is however known to be challenging. While a variety of different techniques have been explored for generative modeling, the application of EBMs to regression is not a well-studied problem. How EBMs should be trained for best possible regression performance is thus currently unclear. We therefore accept the task of providing the first detailed study of this problem. To that end, we propose a simple yet highly effective extension of noise contrastive estimation, and carefully compare its performance to six popular methods from literature on the tasks of 1D regression and object detection. The results of this comparison suggest that our training method should be considered the go-to approach. We also apply our method to the visual tracking task, achieving state-of-the-art performance on five datasets. Notably, our tracker achieves 63.7% AUC on LaSOT and 78.7% Success on TrackingNet. Code is available at https://github.com/fregu856/ebms_regression.

Results

TaskDatasetMetricValueModel
Object TrackingUAV123AUC0.672DiMP-NCE+
Object TrackingLaSOTAUC63.7DiMP-NCE+
Object TrackingNeedForSpeedAUC0.65DiMP-NCE+
Object TrackingOTB-100AUC0.707DiMP-NCE+
Object TrackingTrackingNetAUC0.787DiMP-NCE+
Object TrackingTrackingNetNormalized Precision83.7DiMP-NCE+
Object TrackingTrackingNetPrecision73.7DiMP-NCE+
Object TrackingTrackingNetSuccess Rate0.787DiMP-NCE+
Visual Object TrackingUAV123AUC0.672DiMP-NCE+
Visual Object TrackingLaSOTAUC63.7DiMP-NCE+
Visual Object TrackingNeedForSpeedAUC0.65DiMP-NCE+
Visual Object TrackingOTB-100AUC0.707DiMP-NCE+
Visual Object TrackingTrackingNetAUC0.787DiMP-NCE+
Visual Object TrackingTrackingNetNormalized Precision83.7DiMP-NCE+
Visual Object TrackingTrackingNetPrecision73.7DiMP-NCE+
Visual Object TrackingTrackingNetSuccess Rate0.787DiMP-NCE+

Related Papers

Language Integration in Fine-Tuning Multimodal Large Language Models for Image-Based Regression2025-07-20A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17RS-TinyNet: Stage-wise Feature Fusion Network for Detecting Tiny Objects in Remote Sensing Images2025-07-17Decoupled PROB: Decoupled Query Initialization Tasks and Objectness-Class Learning for Open World Object Detection2025-07-17Dual LiDAR-Based Traffic Movement Count Estimation at a Signalized Intersection: Deployment, Data Collection, and Preliminary Analysis2025-07-17Neural Network-Guided Symbolic Regression for Interpretable Descriptor Discovery in Perovskite Catalysts2025-07-16Imbalanced Regression Pipeline Recommendation2025-07-16Second-Order Bounds for [0,1]-Valued Regression via Betting Loss2025-07-16