Abstract
Robust loss functions are essential for training deep neural networks with better generalization power in the presence of noisy labels. Symmetric loss functions are confirmed to be robust to label noise. However, the symmetric condition is overly restrictive. In this work, we propose a new class of loss functions, namely asymmetric loss functions, which are robust to learning with noisy labels for various types of noise. We investigate general theoretical properties of asymmetric loss functions, including classification calibration, excess risk bound, and noise tolerance. Meanwhile, we introduce the asymmetry ratio to measure the asymmetry of a loss function. The empirical results show that a higher ratio would provide better noise tolerance. Moreover, we modify several commonly-used loss functions and establish the necessary and sufficient conditions for them to be asymmetric. Experimental results on benchmark datasets demonstrate that asymmetric loss functions can outperform state-of-the-art methods.
Original language | English (US) |
---|---|
Title of host publication | 38th International Conference on Machine Learning, ICML 2021 |
Publisher | ML Research Press |
Pages | 12846-12856 |
Number of pages | 11 |
ISBN (Print) | 9781713845065 |
State | Published - Jan 1 2021 |
Bibliographical note
KAUST Repository Item: Exported on 2023-06-20Acknowledgements: This work was supported by National Key R&D Program of China under Grant 2018AAA0102801 and 2019YFE0109600, and National Natural Science Foundation of China under Grants 61922027 and 61932022.