Early Exiting with Ensemble Internal Classifiers

05/28/2021
by   Tianxiang Sun, et al.
0

As a simple technique to accelerate inference of large-scale pre-trained models, early exiting has gained much attention in the NLP community. It allows samples to exit early at internal classifiers without passing through the entire model. Most existing work usually trains the internal classifiers independently and employs an exiting strategy to decide whether or not to exit based on the confidence of the current internal classifier. However, none of these works takes full advantage of the fact that the internal classifiers are trained to solve the same task therefore can be used to construct an ensemble. In this paper, we show that a novel objective function for the training of the ensemble internal classifiers can be naturally induced from the perspective of ensemble learning and information theory. The proposed training objective consists of two terms: one for accuracy and the other for the diversity of the internal classifiers. In contrast, the objective used in prior work is exactly the accuracy term of our training objective therefore only optimizes the accuracy but not diversity. Further, we propose a simple voting-based strategy that considers predictions of all the past internal classifiers to infer the correct label and decide whether to exit. Experimental results on various NLP tasks show that our proposed objective function and voting-based strategy can achieve better accuracy-speed trade-offs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/07/2020

BERT Loses Patience: Fast and Robust Inference with Early Exit

In this paper, we propose Patience-based Early Exit, a straightforward y...
research
10/15/2020

SpaML: a Bimodal Ensemble Learning Spam Detector based on NLP Techniques

In this paper, we put forward a new tool, called SpaML, for spam detecti...
research
07/09/2021

Specialists Outperform Generalists in Ensemble Classification

Consider an ensemble of k individual classifiers whose accuracies are kn...
research
06/09/2021

Zero Time Waste: Recycling Predictions in Early Exit Neural Networks

The problem of reducing processing time of large deep learning models is...
research
06/09/2022

On Margins and Generalisation for Voting Classifiers

We study the generalisation properties of majority voting on finite ense...
research
01/28/2023

Anticipate, Ensemble and Prune: Improving Convolutional Neural Networks via Aggregated Early Exits

Today, artificial neural networks are the state of the art for solving a...
research
02/07/2021

Sparsely ensembled convolutional neural network classifiers via reinforcement learning

We consider convolutional neural network (CNN) ensemble learning with th...

Please sign up or login with your details

Forgot password? Click here to reset