Timezone: »

Domain Adaptation with Invariant Representation Learning: What Transformations to Learn?
Petar Stojanov · Zijian Li · Mingming Gong · Ruichu Cai · Jaime Carbonell · Kun Zhang

Tue Dec 07 08:30 AM -- 10:00 AM (PST) @ None #None
Unsupervised domain adaptation, as a prevalent transfer learning setting, spans many real-world applications. With the increasing representational power and applicability of neural networks, state-of-the-art domain adaptation methods make use of deep architectures to map the input features $X$ to a latent representation $Z$ that has the same marginal distribution across domains. This has been shown to be insufficient for generating optimal representation for classification, and to find conditionally invariant representations, usually strong assumptions are needed. We provide reasoning why when the supports of the source and target data from overlap, any map of $X$ that is fixed across domains may not be suitable for domain adaptation via invariant features. Furthermore, we develop an efficient technique in which the optimal map from $X$ to $Z$ also takes domain-specific information as input, in addition to the features $X$. By using the property of minimal changes of causal mechanisms across domains, our model also takes into account the domain-specific information to ensure that the latent representation $Z$ does not discard valuable information about $Y$. We demonstrate the efficacy of our method via synthetic and real-world data experiments. The code is available at: \texttt{https://github.com/DMIRLAB-Group/DSAN}.

Author Information

Petar Stojanov (Carnegie Mellon University)
Zijian Li (Guangdong University of Technology)
Mingming Gong (University of Melbourne)
Ruichu Cai (Guangdong University of Technology)
Jaime Carbonell
Kun Zhang (CMU)

More from the Same Authors