Distilling Universal and Joint Knowledge for Cross-Domain Model Compression on Time Series Data

07/07/2023
by   Qing Xu, et al.
0

For many real-world time series tasks, the computational complexity of prevalent deep leaning models often hinders the deployment on resource-limited environments (e.g., smartphones). Moreover, due to the inevitable domain shift between model training (source) and deploying (target) stages, compressing those deep models under cross-domain scenarios becomes more challenging. Although some of existing works have already explored cross-domain knowledge distillation for model compression, they are either biased to source data or heavily tangled between source and target data. To this end, we design a novel end-to-end framework called Universal and joint knowledge distillation (UNI-KD) for cross-domain model compression. In particular, we propose to transfer both the universal feature-level knowledge across source and target domains and the joint logit-level knowledge shared by both domains from the teacher to the student model via an adversarial learning scheme. More specifically, a feature-domain discriminator is employed to align teacher's and student's representations for universal knowledge transfer. A data-domain discriminator is utilized to prioritize the domain-shared samples for joint knowledge transfer. Extensive experimental results on four time series datasets demonstrate the superiority of our proposed method over state-of-the-art (SOTA) benchmarks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/29/2021

Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer

Recent applications pose requirements of both cross-domain knowledge tra...
research
12/02/2020

Meta-KD: A Meta Knowledge Distillation Framework for Language Model Compression across Domains

Pre-trained language models have been applied to various NLP tasks with ...
research
06/07/2020

ADMP: An Adversarial Double Masks Based Pruning Framework For Unsupervised Cross-Domain Compression

Despite the recent progress of network pruning, directly applying it to ...
research
07/08/2019

Multivariate-Information Adversarial Ensemble for Scalable Joint Distribution Matching

A broad range of cross-m-domain generation researches boil down to match...
research
01/20/2021

Learning to Augment for Data-Scarce Domain BERT Knowledge Distillation

Despite pre-trained language models such as BERT have achieved appealing...
research
08/19/2023

Prototypical Cross-domain Knowledge Transfer for Cervical Dysplasia Visual Inspection

Early detection of dysplasia of the cervix is critical for cervical canc...
research
04/05/2022

A Two-student Learning Framework for Mixed Supervised Target Sound Detection

Target sound detection (TSD) aims to detect the target sound from mixtur...

Please sign up or login with your details

Forgot password? Click here to reset