AllMatch: Exploiting All Unlabeled Data for Semi-Supervised Learning

AllMatch: Exploiting All Unlabeled Data for Semi-Supervised Learning

Zhiyu Wu, Jinshi Cui

Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence
Main Track. Pages 5245-5253. https://doi.org/10.24963/ijcai.2024/580

Existing semi-supervised learning algorithms adopt pseudo-labeling and consistency regulation techniques to introduce supervision signals for unlabeled samples. To overcome the inherent limitation of threshold-based pseudo-labeling, prior studies have attempted to align the confidence threshold with the evolving learning status of the model, which is estimated through the predictions made on the unlabeled data. In this paper, we further reveal that classifier weights can reflect the differentiated learning status across categories and consequently propose a class-specific adaptive threshold mechanism. Additionally, considering that even the optimal threshold scheme cannot resolve the problem of discarding unlabeled samples, a binary classification consistency regulation approach is designed to distinguish candidate classes from negative options for all unlabeled samples. By combining the above strategies, we present a novel SSL algorithm named AllMatch, which achieves improved pseudo-label accuracy and a 100% utilization ratio for the unlabeled data. We extensively evaluate our approach on multiple benchmarks, encompassing both balanced and imbalanced settings. The results demonstrate that AllMatch consistently outperforms existing state-of-the-art methods.
Keywords:
Machine Learning: ML: Semi-supervised learning