TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/Self-Adjusting Smooth L1 Loss

Self-Adjusting Smooth L1 Loss

GeneralIntroduced 20001 papers
Source Paper

Description

Self-Adjusting Smooth L1 Loss is a loss function used in object detection that was introduced with RetinaMask. This is an improved version of Smooth L1. For Smooth L1 loss we have:

f(x)=0.5x2β if ∣x∣<βf(x) = 0.5 \frac{x^{2}}{\beta} \text{ if } |x| < \betaf(x)=0.5βx2​ if ∣x∣<β f(x)=∣x∣−0.5β otherwise f(x) = |x| -0.5\beta \text{ otherwise }f(x)=∣x∣−0.5β otherwise 

Here a point β\betaβ splits the positive axis range into two parts: L2L2L2 loss is used for targets in range [0,β][0, \beta][0,β], and L1L1L1 loss is used beyond β\betaβ to avoid over-penalizing utliers. The overall function is smooth (continuous, together with its derivative). However, the choice of control point (β\betaβ) is heuristic and is usually done by hyper parameter search.

Instead, with self-adjusting smooth L1 loss, inside the loss function the running mean and variance of the absolute loss are recorded. We use the running minibatch mean and variance with a momentum of 0.90.90.9 to update these two parameters.

Papers Using This Method

RetinaMask: Learning to predict masks improves state-of-the-art single-shot detection for free2019-01-10