Professional Documents
Culture Documents
Harsh Chaurasia
October 2023
1 Introduction
TITLE: Universal Domain Adaptation Through Self-Supervision
AUTHORS: K. Saito, D. Kim, S. Sclaroff
Domain Adaptation (DA) is one of the most challenging and interesting problems in Computer Vision right now. Most of the current
unsupervised approaches assume that all the source categories are present in the target domain, which often proves to be impractical.
The authors in this paper have proposed Dance: Domain Adaptive Neighborhood Clustering via Entropy optimisation, a
novel method to enforce soft feature alignment paired with clustering in order to achieve universal domain adaptation.
Recent DA techniques make strong assumptions about category overlap between domains. For Ls and Lt being label sets from the
source and the target domain, the authors have tried to take care of all possible category shifts: OPEN-SET DA (Ls ⊂ Lt ),
CLOSED-SET DA (Ls = Lt ), PARTIAL DA (Lt ⊂ Ls ), and a mix of open-set and partial. Since the target domain is unlabelled in
a practical setting, we may not know which one of the above settings is at hand, leading to catastrophic misalignments. Moreover,
we often rely too much on source supervision, making it difficult to obtain discriminative features on the target. This leads to
failure in discriminating ”unknown” categories from the known categories. The earlier proposed self-supervision methods do not use
the cluster structure of the target domain, which has been taken care of in this paper. DANCE is proposed in this paper, which
harnesses the cluster structure of the target domain as well by using self-supervision. This preserves useful source features by using
distribution alignment by batch normalisation and entropy separation loss. DANCE is shown to achieve state-of-the-art results on
every source-only model.
Classical approaches for CDA measure distance between the source and target feature distributions and train a model to minimise
this distance. However, this method fails on ODA, PDA, and OPDA, defeating the purpose of UniDA. Additionally, self-supervised
methods have been used in the past, but directly using them is challenging as we require the number of clusters in the target domain,
which is not always readily available.
This is minimised to align each target sample to either a target neighbor or a prototype.
Some of the target samples still need to be aligned with known source categories. Unknown target samples are likely to have a
larger entropy of the source classifier’s output than known target samples as they do not share common features. For this, entropy
separation loss is used:
1 X
Les = Les (ρi )
|Bt |
i∈Bt