Dynamic Knowledge Distillation for Black-box Hypothesis Transfer Learning

07/24/2020
by   Yiqin Yu, et al.
19

In real world applications like healthcare, it is usually difficult to build a machine learning prediction model that works universally well across different institutions. At the same time, the available model is often proprietary, i.e., neither the model parameter nor the data set used for model training is accessible. In consequence, leveraging the knowledge hidden in the available model (aka. the hypothesis) and adapting it to a local data set becomes extremely challenging. Motivated by this situation, in this paper we aim to address such a specific case within the hypothesis transfer learning framework, in which 1) the source hypothesis is a black-box model and 2) the source domain data is unavailable.In particular, we introduce a novel algorithm called dynamic knowledge distillation for hypothesis transfer learning (dkdHTL). In this method, we use knowledge distillation with instance-wise weighting mechanism to adaptively transfer the "dark" knowledge from the source hypothesis to the target domain.The weighting coefficients of the distillation loss and the standard loss are determined by the consistency between the predicted probability of the source hypothesis and the target ground-truth label.Empirical results on both transfer learning benchmark datasets and a healthcare dataset demonstrate the effectiveness of our method.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/03/2018

A Hybrid Instance-based Transfer Learning Method

In recent years, supervised machine learning models have demonstrated tr...
research
06/07/2021

Zero-Shot Knowledge Distillation from a Decision-Based Black-Box Model

Knowledge distillation (KD) is a successful approach for deep neural net...
research
05/13/2023

Black-box Source-free Domain Adaptation via Two-stage Knowledge Distillation

Source-free domain adaptation aims to adapt deep neural networks using o...
research
05/03/2021

Black-Box Dissector: Towards Erasing-based Hard-Label Model Stealing Attack

Model stealing attack aims to create a substitute model that steals the ...
research
05/28/2022

Divide to Adapt: Mitigating Confirmation Bias for Domain Adaptation of Black-Box Predictors

Domain Adaptation of Black-box Predictors (DABP) aims to learn a model o...
research
10/23/2022

Respecting Transfer Gap in Knowledge Distillation

Knowledge distillation (KD) is essentially a process of transferring a t...
research
08/15/2019

Multitask and Transfer Learning for Autotuning Exascale Applications

Multitask learning and transfer learning have proven to be useful in the...

Please sign up or login with your details

Forgot password? Click here to reset