TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/Huber loss

Huber loss

GeneralIntroduced 200077 papers

Description

The Huber loss function describes the penalty incurred by an estimation procedure f. Huber (1964) defines the loss function piecewise by[1]

L δ ( a ) = { 1 2 a 2 for  | a | ≤ δ , δ ⋅ ( | a | − 1 2 δ ) , otherwise. {\displaystyle L_{\delta }(a)={\begin{cases}{\frac {1}{2}}{a^{2}}&{\text{for }}|a|\leq \delta ,\\\delta \cdot \left(|a|-{\frac {1}{2}}\delta \right),&{\text{otherwise.}}\end{cases}}}

This function is quadratic for small values of a, and linear for large values, with equal values and slopes of the different sections at the two points where | a | = δ |a|=\delta . The variable a often refers to the residuals, that is to the difference between the observed and predicted values a = y − f ( x ) a=y-f(x), so the former can be expanded to[2]

L δ ( y , f ( x ) ) = { 1 2 ( y − f ( x ) ) 2 for  | y − f ( x ) | ≤ δ , δ   ⋅ ( | y − f ( x ) | − 1 2 δ ) , otherwise. {\displaystyle L_{\delta }(y,f(x))={\begin{cases}{\frac {1}{2}}(y-f(x))^{2}&{\text{for }}|y-f(x)|\leq \delta ,\\\delta \ \cdot \left(|y-f(x)|-{\frac {1}{2}}\delta \right),&{\text{otherwise.}}\end{cases}}}

The Huber loss is the convolution of the absolute value function with the rectangular function, scaled and translated. Thus it "smoothens out" the former's corner at the origin.

.. math:: \ell(x, y) = L = {l_1, ..., l_N}^T

with

.. math::
    l_n = \begin{cases}
    0.5 (x_n - y_n)^2, & \text{if } |x_n - y_n| < delta \\
    delta * (|x_n - y_n| - 0.5 * delta), & \text{otherwise }
    \end{cases}

Papers Using This Method

SHORE: A Long-term User Lifetime Value Prediction Model in Digital Games2025-06-12Heavy Lasso: sparse penalized regression under heavy-tailed noise via data-augmented soft-thresholding2025-06-09Differentially Private Sparse Linear Regression with Heavy-tailed Responses2025-06-07A Surrogate Model for the Forward Design of Multi-layered Metasurface-based Radar Absorbing Structures2025-05-14A Piecewise Lyapunov Analysis of Sub-quadratic SGD: Applications to Robust and Quantile Regression2025-04-11Asymptotics of Non-Convex Generalized Linear Models in High-Dimensions: A proof of the replica formula2025-02-27Explainable AI based System for Supply Air Temperature Forecast2025-01-09Robust Sparse Regression with Non-Isotropic Designs2024-10-31Residual-based Adaptive Huber Loss (RAHL) -- Design of an improved Huber loss for CQI prediction in 5G networks2024-08-27A Huber Loss Minimization Approach to Mean Estimation under User-level Differential Privacy2024-05-22Robust deep learning from weakly dependent data2024-05-08$H$-Consistency Guarantees for Regression2024-03-28A Unified Framework for Center-based Clustering of Distributed Data2024-02-02A Robust Quantile Huber Loss With Interpretable Parameter Adjustment In Distributional Reinforcement Learning2024-01-04Robust Non-parametric Knowledge-based Diffusion Least Mean Squares over Adaptive Networks2023-12-03Towards Robust Offline Reinforcement Learning under Diverse Data Corruption2023-10-19Dual Latent State Learning: Exploiting Regional Network Similarities for QoS Prediction2023-10-07High-dimensional robust regression under heavy-tailed data: Asymptotics and Universality2023-09-28A Huber Loss Minimization Approach to Byzantine Robust Federated Learning2023-08-24Machine learning in parameter estimation of nonlinear systems2023-08-23