Bibimbap: Pre-trained models ensemble for Domain Generalization

Jinho Kang, Taero Kim, Yewon Kim, Changdae Oh, Jiyoung Jung, Rakwoo Chang, Kyungwoo Song

Research output: Contribution to journalArticlepeer-review

Abstract

This paper addresses a machine learning problem often challenged by differences in the distributions of training and real-world data. We propose a framework that addresses the problem of underfitting in the ensembling method using pre-trained models and improves the performance and robustness of deep learning models through ensemble diversity. For the naive weight ensembling framework, we discovered that the ensembled models could not lie in the same loss basin under extreme domain shift conditions, suggesting that a loss barrier may exist. We used a fine-tuning step after the weighted ensemble to address the underfitting problem caused by the loss barrier and stabilize the batch normalization running parameters. We also inferred through qualitative analysis that the diversity of ensemble models affects domain generalization. We validate our method on a large-scale image dataset (ImageNet-1K) and chemical molecule data, which is suitable for testing with domain shift problems due to its data-splitting method.

Original languageEnglish
Article number110391
JournalPattern Recognition
Volume151
DOIs
StatePublished - Jul 2024

Keywords

  • Chemical dataset
  • Domain generalization
  • Ensemble learning
  • Molecular classification
  • Transfer learning
  • Weight averaging

Fingerprint

Dive into the research topics of 'Bibimbap: Pre-trained models ensemble for Domain Generalization'. Together they form a unique fingerprint.

Cite this