Knowledge Adaptation: Teaching to Adapt

02/07/2017
by   Sebastian Ruder, et al.
0

Domain adaptation is crucial in many real-world applications where the distribution of the training data differs from the distribution of the test data. Previous Deep Learning-based approaches to domain adaptation need to be trained jointly on source and target domain data and are therefore unappealing in scenarios where models need to be adapted to a large number of domains or where a domain is evolving, e.g. spam detection where attackers continuously change their tactics. To fill this gap, we propose Knowledge Adaptation, an extension of Knowledge Distillation (Bucilua et al., 2006; Hinton et al., 2015) to the domain adaptation scenario. We show how a student model achieves state-of-the-art results on unsupervised domain adaptation from multiple sources on a standard sentiment analysis benchmark by taking into account the domain-specific expertise of multiple teachers and the similarities between their domains. When learning from a single teacher, using domain similarity to gauge trustworthiness is inadequate. To this end, we propose a simple metric that correlates well with the teacher's accuracy in the target domain. We demonstrate that incorporating high-confidence examples selected by this metric enables the student model to achieve state-of-the-art performance in the single-source scenario.

READ FULL TEXT

page 5

page 8

research
03/07/2022

Student Become Decathlon Master in Retinal Vessel Segmentation via Dual-teacher Multi-target Domain Adaptation

Unsupervised domain adaptation has been proposed recently to tackle the ...
research
06/16/2017

Self-ensembling for visual domain adaptation

This paper explores the use of self-ensembling for visual domain adaptat...
research
10/20/2020

Teacher-Student Consistency For Multi-Source Domain Adaptation

In Multi-Source Domain Adaptation (MSDA), models are trained on samples ...
research
10/22/2020

Knowledge Distillation for BERT Unsupervised Domain Adaptation

A pre-trained language model, BERT, has brought significant performance ...
research
07/19/2022

ML-BPM: Multi-teacher Learning with Bidirectional Photometric Mixing for Open Compound Domain Adaptation in Semantic Segmentation

Open compound domain adaptation (OCDA) considers the target domain as th...
research
03/17/2020

Teacher-Student Domain Adaptation for Biosensor Models

We present an approach to domain adaptation, addressing the case where d...
research
11/15/2022

Instance-aware Model Ensemble With Distillation For Unsupervised Domain Adaptation

The linear ensemble based strategy, i.e., averaging ensemble, has been p...

Please sign up or login with your details

Forgot password? Click here to reset