TY - GEN
T1 - Asymmetric Loss Functions for Learning with Noisy Labels
AU - Zhou, Xiong
AU - Liu, Xianming
AU - Jiang, Junjun
AU - Gao, Xin
AU - Ji, Xiangyang
N1 - KAUST Repository Item: Exported on 2023-06-20
Acknowledgements: This work was supported by National Key R&D Program of China under Grant 2018AAA0102801 and 2019YFE0109600, and National Natural Science Foundation of China under Grants 61922027 and 61932022.
PY - 2021/1/1
Y1 - 2021/1/1
N2 - Robust loss functions are essential for training deep neural networks with better generalization power in the presence of noisy labels. Symmetric loss functions are confirmed to be robust to label noise. However, the symmetric condition is overly restrictive. In this work, we propose a new class of loss functions, namely asymmetric loss functions, which are robust to learning with noisy labels for various types of noise. We investigate general theoretical properties of asymmetric loss functions, including classification calibration, excess risk bound, and noise tolerance. Meanwhile, we introduce the asymmetry ratio to measure the asymmetry of a loss function. The empirical results show that a higher ratio would provide better noise tolerance. Moreover, we modify several commonly-used loss functions and establish the necessary and sufficient conditions for them to be asymmetric. Experimental results on benchmark datasets demonstrate that asymmetric loss functions can outperform state-of-the-art methods.
AB - Robust loss functions are essential for training deep neural networks with better generalization power in the presence of noisy labels. Symmetric loss functions are confirmed to be robust to label noise. However, the symmetric condition is overly restrictive. In this work, we propose a new class of loss functions, namely asymmetric loss functions, which are robust to learning with noisy labels for various types of noise. We investigate general theoretical properties of asymmetric loss functions, including classification calibration, excess risk bound, and noise tolerance. Meanwhile, we introduce the asymmetry ratio to measure the asymmetry of a loss function. The empirical results show that a higher ratio would provide better noise tolerance. Moreover, we modify several commonly-used loss functions and establish the necessary and sufficient conditions for them to be asymmetric. Experimental results on benchmark datasets demonstrate that asymmetric loss functions can outperform state-of-the-art methods.
UR - http://hdl.handle.net/10754/692665
UR - https://proceedings.mlr.press/v139/zhou21f.html
UR - http://www.scopus.com/inward/record.url?scp=85161279939&partnerID=8YFLogxK
M3 - Conference contribution
SN - 9781713845065
SP - 12846
EP - 12856
BT - 38th International Conference on Machine Learning, ICML 2021
PB - ML Research Press
ER -