Learning the Wrong Lessons: Inserting Trojans During Knowledge Distillation

03/09/2023
by   Leonard Tang, et al.
0

In recent years, knowledge distillation has become a cornerstone of efficiently deployed machine learning, with labs and industries using knowledge distillation to train models that are inexpensive and resource-optimized. Trojan attacks have contemporaneously gained significant prominence, revealing fundamental vulnerabilities in deep learning models. Given the widespread use of knowledge distillation, in this work we seek to exploit the unlabelled data knowledge distillation process to embed Trojans in a student model without introducing conspicuous behavior in the teacher. We ultimately devise a Trojan attack that effectively reduces student accuracy, does not alter teacher performance, and is efficiently constructible in practice.

READ FULL TEXT
research
05/31/2022

What Knowledge Gets Distilled in Knowledge Distillation?

Knowledge distillation aims to transfer useful information from a teache...
research
07/17/2020

Knowledge Distillation in Deep Learning and its Applications

Deep learning based models are relatively large, and it is hard to deplo...
research
11/30/2020

A Selective Survey on Versatile Knowledge Distillation Paradigm for Neural Network Models

This paper aims to provide a selective survey about knowledge distillati...
research
01/27/2022

Dynamic Rectification Knowledge Distillation

Knowledge Distillation is a technique which aims to utilize dark knowled...
research
12/01/2018

On Compressing U-net Using Knowledge Distillation

We study the use of knowledge distillation to compress the U-net archite...
research
11/23/2020

Generative Adversarial Simulator

Knowledge distillation between machine learning models has opened many n...
research
03/14/2023

Teacher-Student Knowledge Distillation for Radar Perception on Embedded Accelerators

Many radar signal processing methodologies are being developed for criti...

Please sign up or login with your details

Forgot password? Click here to reset