Iterative ensemble pseudo-labeling for convolutional neural networks


Creative Commons License

Yildiz S., AMASYALI M. F.

Sigma Journal of Engineering and Natural Sciences, vol.42, no.3, pp.862-874, 2024 (ESCI) identifier

  • Publication Type: Article / Article
  • Volume: 42 Issue: 3
  • Publication Date: 2024
  • Doi Number: 10.14744/sigma.2024.00077
  • Journal Name: Sigma Journal of Engineering and Natural Sciences
  • Journal Indexes: Emerging Sources Citation Index (ESCI), Scopus, Academic Search Premier, Directory of Open Access Journals
  • Page Numbers: pp.862-874
  • Keywords: Ensemble Learning, Pseudo Labeling, Semi-Supervised Learning, STL-10
  • Yıldız Technical University Affiliated: Yes

Abstract

As is well known, the quantity of labeled samples determines the success of a convolutional neural network (CNN). However, creating the labeled dataset is a difficult and time-consuming process. In contrast, unlabeled data is cheap and easy to access. Semi-supervised methods incorporate unlabeled data into the training process, which allows the model to learn from unlabeled data as well. We propose a semi-supervised method based on the ensemble approach and the pseudo-labeling method. By balancing the unlabeled dataset with the labeled dataset during training, both the decision diversity between base-learner models and the individual success of base-learner models are high in our proposed training strategy. We show that using multiple CNN models can result in both higher success and a more robust model than training a single CNN model. For inference, we propose using both stacking and voting methodologies. We have shown that the most successful algorithm for the stacking approach is the Support Vector Machine (SVM). In experiments, we use the STL-10 dataset to evaluate models, and we increased accuracy by 15.9% over training using only labeled data. Since we propose a training method based on cross-entropy loss, it can be implemented combined with state-of-the-art algorithms.