Shihao Zhang, Linlin Yang, Michael Bi Mi, Xiaoxu Zheng, Angela Yao
In computer vision, it is often observed that formulating regression problems as a classification task often yields better performance. We investigate this curious phenomenon and provide a derivation to show that classification, with the cross-entropy loss, outperforms regression with a mean squared error loss in its ability to learn high-entropy feature representations. Based on the analysis, we propose an ordinal entropy loss to encourage higher-entropy feature spaces while maintaining ordinal relationships to improve the performance of regression tasks. Experiments on synthetic and real-world regression tasks demonstrate the importance and benefits of increasing entropy for regression.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Depth Estimation | NYU-Depth V2 | Delta < 1.25 | 0.932 | OrdinalEntropy |
| Depth Estimation | NYU-Depth V2 | RMSE | 0.321 | OrdinalEntropy |
| Depth Estimation | NYU-Depth V2 | absolute relative error | 0.089 | OrdinalEntropy |
| Depth Estimation | NYU-Depth V2 | log 10 | 0.039 | OrdinalEntropy |
| Crowds | ShanghaiTech B | MAE | 9.1 | OrdinalEntropy |
| Crowds | ShanghaiTech B | MSE | 14.5 | OrdinalEntropy |
| Crowds | ShanghaiTech A | MAE | 65.6 | OrdinalEntropy |
| Crowds | ShanghaiTech A | MSE | 105 | OrdinalEntropy |
| 3D | NYU-Depth V2 | Delta < 1.25 | 0.932 | OrdinalEntropy |
| 3D | NYU-Depth V2 | RMSE | 0.321 | OrdinalEntropy |
| 3D | NYU-Depth V2 | absolute relative error | 0.089 | OrdinalEntropy |
| 3D | NYU-Depth V2 | log 10 | 0.039 | OrdinalEntropy |