Foster Adaptivity and Balance in Learning with Noisy Labels

1 Nanjing University of Science and Technology, 2 Horizon Robotics
European Conference on Computer Vision (ECCV), 2024
Poster Dataset Distillation (PoDD)

We propose a simple yet effective method, named SED, to learn with noisy labels in a Self-adaptivE and class-balanceD manner. SED divides the training set into clean and noisy subsets using dynamically updated thresholds, employs a mean-teacher model for label correction, and adaptively weights noisy samples based on correction confidence. The final objective integrates classification losses and consistency regularization to boost model performance.

Abstract

Label noise is ubiquitous in real-world scenarios, posing a practical challenge to supervised models due to its effect in hurting the generalization performance of deep neural networks. Existing methods primarily employ the sample selection paradigm and usually rely on dataset-dependent prior knowledge (e.g., a pre-defined threshold) to cope with label noise, inevitably degrading the adaptivity. Moreover, existing methods tend to neglect the class balance in selecting samples, leading to biased model performance. To this end, we propose a simple yet effective approach named SED to deal with label noise in a Self-adaptivE and class-balanceD manner. Specifically, we first design a novel sample selection strategy to empower self-adaptivity and class balance when identifying clean and noisy data. A mean-teacher model is then employed to correct labels of noisy samples. Subsequently, we propose a self-adaptive and class-balanced sample re-weighting mechanism to assign different weights to detected noisy samples. Finally, we additionally employ consistency regularization on selected clean samples to improve model generalization performance. Extensive experimental results on synthetic and real-world datasets demonstrate the effectiveness and superiority of our proposed method.

Self-Adaptive and Class-Balanced Sample Selection and Re-weighting

Self-Adaptive and Class-Balanced Sample Selection and Re-weighting

(a-b) Self-adaptive and class-balanced sample selection based on predicted probability w.r.t. given labels. The blue curve indicates the class-specific selection thresholds. (c-d) Self-adaptive and class-balanced sample re-weighting based on correction confidence. The orange curve represents the class-specific confidence threshold.

Average Test Accuracy on CIFAR100N and CIFAR80N Under Various Noise Conditions

Average Test Accuracy on CIFAR100N and CIFAR80N Under Various Noise Conditions

Experiments are conducted under symmetric and asymmetric label noise conditions. Results of existing methods are mainly drawn from previous works. † indicates methods re-implemented using their open-sourced code and default hyper-parameters.

Comparison of Test Accuracy on Real-World Noisy Datasets

Comparison of Test Accuracy on Real-World Noisy Datasets

The table shows the experimental results of existing methods and SED on Web-Aircraft, Web-Bird, and Web-Car datasets. Results of existing methods are mainly drawn from previous works. † indicates methods re-implemented using their open-sourced code and default hyper-parameters.

BibTeX

@article{sheng2024foster,
  title={Foster Adaptivity and Balance in Learning with Noisy Labels},
  author={Sheng, Mengmeng and Sun, Zeren and Chen, Tao and Pang, Shuchao and Wang, Yucheng and Yao, Yazhou},
  journal={European Conference on Computer Vision (ECCV)},
  year={2024}
}