Research

CDS: Cross-domain self-supervised pre-training

ICCV

Authors

Published on

10/17/2021

We present a two-stage pre-training approach that improves the generalization ability of standard single-domain pre-training. While standard pre-training on a single large dataset (such as ImageNet) can provide a good initial representation for transfer learning tasks, this approach may result in biased representations that impact the success of learning with new multi-domain data (e.g., different artistic styles) via methods like domain adaptation. We propose a novel pre-training approach called Cross-Domain Self-supervision (CDS), which directly employs unlabeled multi-domain data for downstream domain transfer tasks. Our approach uses self-supervision not only within a single domain but also across domains. In-domain instance discrimination is used to learn discriminative features on new data in a domain-adaptive manner, while cross-domain matching is used to learn domain-invariant features. We apply our method as a second pre-training step (after ImageNet pre-training), resulting in a significant target accuracy boost to diverse domain transfer tasks compared to standard one-stage pre-training.

Please cite our work using the BibTeX below.

@InProceedings{Kim_2021_ICCV,
    author    = {Kim, Donghyun and Saito, Kuniaki and Oh, Tae-Hyun and Plummer, Bryan A. and Sclaroff, Stan and Saenko, Kate},
    title     = {CDS: Cross-Domain Self-Supervised Pre-Training},
    booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},
    month     = {October},
    year      = {2021},
    pages     = {9123-9132}
}
Close Modal