CDS: Cross-domain self-supervised pre-training
Authors
Authors
- Donghyun Kim
- Kuniaki Saito
- Tae-Hyun Oh
- Bryan A. Plummer
- Stan Sclaroff
- Kate Saenko
Authors
- Donghyun Kim
- Kuniaki Saito
- Tae-Hyun Oh
- Bryan A. Plummer
- Stan Sclaroff
- Kate Saenko
Published on
10/17/2021
We present a two-stage pre-training approach that improves the generalization ability of standard single-domain pre-training. While standard pre-training on a single large dataset (such as ImageNet) can provide a good initial representation for transfer learning tasks, this approach may result in biased representations that impact the success of learning with new multi-domain data (e.g., different artistic styles) via methods like domain adaptation. We propose a novel pre-training approach called Cross-Domain Self-supervision (CDS), which directly employs unlabeled multi-domain data for downstream domain transfer tasks. Our approach uses self-supervision not only within a single domain but also across domains. In-domain instance discrimination is used to learn discriminative features on new data in a domain-adaptive manner, while cross-domain matching is used to learn domain-invariant features. We apply our method as a second pre-training step (after ImageNet pre-training), resulting in a significant target accuracy boost to diverse domain transfer tasks compared to standard one-stage pre-training.
Please cite our work using the BibTeX below.
@InProceedings{Kim_2021_ICCV,
author = {Kim, Donghyun and Saito, Kuniaki and Oh, Tae-Hyun and Plummer, Bryan A. and Sclaroff, Stan and Saenko, Kate},
title = {CDS: Cross-Domain Self-Supervised Pre-Training},
booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},
month = {October},
year = {2021},
pages = {9123-9132}
}