Crafting Distribution Shifts for Validation and Training
in Single Source Domain Generalization

WACV 2025 Oral

Nikos Efthymiadis*, Giorgos Tolias, Ondřej Chum

VRG, FEE, Czech Technical University in Prague

*Corresponding: efthynik@fel.cvut.cz

[arxiv] [code] [poster] [video] [bibtex]

TL;DR

In this paper, we propose a validation protocol for Single Source Domain Generalization that is capable to reflect the generalization abilities of each model, while being limited to the source domain validation set! Also proposes a SoTA method for Single Source Domain Generalization.

teaser

Motivation

Single-source domain generalization attempts to learn a model on a source domain and deploy it to unseen target domains. Limiting access only to source domain data imposes two key challenges – how to train a model that can generalize and how to verify that it does. The standard practice of validation on the training distribution does not accurately reflect the model’s generalization ability, while validation on the test distribution is a malpractice to avoid. In this work, we follow a fundamental direction in the generalization task, i.e., data augmentations, to synthesize new distributions, but in contrast to the standard practice, we apply them to the validation set instead, to estimate the method’s performance on multiple distribution shifts.

Approach

teaser

Since these augmentations are also valuable in the training phase, we propose a k-fold cross-validation scheme performed across augmentation types to get the best of both worlds. This way, the training set is augmented with challenging examples while, at the same time, the validation provides an unbiased estimate of performance on unseen distributions.

Besides the novel validation method for SSDG, we propose a family of classification methods parametrized by several train and test-time hyper-parameters. The values of these parameters are selected by the proposed validation method. We focus on enforcing shape bias, which is effectively demonstrated in prior work. We accomplish this by using a specialized image transformation technique, employing enhanced edge maps that eliminate textures while retaining crucial shape information. The transformation is performed both during training and testing.