Pool of Experts: Realtime Querying Specialized Knowledge in Massive Neural Networks

07/03/2021
by   Hakbin Kim, et al.
0

In spite of the great success of deep learning technologies, training and delivery of a practically serviceable model is still a highly time-consuming process. Furthermore, a resulting model is usually too generic and heavyweight, and hence essentially goes through another expensive model compression phase to fit in a resource-limited device like embedded systems. Inspired by the fact that a machine learning task specifically requested by mobile users is often much simpler than it is supported by a massive generic model, this paper proposes a framework, called Pool of Experts (PoE), that instantly builds a lightweight and task-specific model without any training process. For a realtime model querying service, PoE first extracts a pool of primitive components, called experts, from a well-trained and sufficiently generic network by exploiting a novel conditional knowledge distillation method, and then performs our train-free knowledge consolidation to quickly combine necessary experts into a lightweight network for a target task. Thanks to this train-free property, in our thorough empirical study, PoE can build a fairly accurate yet compact model in a realtime manner, whereas it takes a few minutes per query for the other training methods to achieve a similar level of the accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/19/2020

Learning from a Lightweight Teacher for Efficient Knowledge Distillation

Knowledge Distillation (KD) is an effective framework for compressing de...
research
08/05/2019

Knowledge Isomorphism between Neural Networks

This paper aims to analyze knowledge isomorphism between pre-trained dee...
research
10/20/2020

Asynchronous Edge Learning using Cloned Knowledge Distillation

With the increasing demand for more and more data, the federated learnin...
research
11/13/2018

Private Model Compression via Knowledge Distillation

The soaring demand for intelligent mobile applications calls for deployi...
research
05/03/2023

A Systematic Study of Knowledge Distillation for Natural Language Generation with Pseudo-Target Training

Modern Natural Language Generation (NLG) models come with massive comput...
research
03/31/2022

Training strategy for a lightweight countermeasure model for automatic speaker verification

The countermeasure (CM) model is developed to protect Automatic Speaker ...
research
11/03/2020

Towards a Universal Gating Network for Mixtures of Experts

The combination and aggregation of knowledge from multiple neural networ...

Please sign up or login with your details

Forgot password? Click here to reset