Deep Active Learning by Model Interpretability

07/23/2020
by   Qiang Liu, et al.
0

Recent successes of Deep Neural Networks (DNNs) in a variety of research tasks, however, heavily rely on the large amounts of labeled samples. This may require considerable annotation cost in real-world applications. Fortunately, active learning is a promising methodology to train high-performing model with minimal annotation cost. In the deep learning context, the critical question of active learning is how to precisely identify the informativeness of samples for DNN. In this paper, inspired by piece-wise linear interpretability in DNN, we first introduce the linear separable regions of samples to the problem of active learning, and propose a novel Deep Active learning approach by Model Interpretability (DAMI). To keep the maximal representativeness of the entire unlabeled data, DAMI tries to select and label samples on different linear separable regions introduced by the piece-wise linear interpretability in DNN. We focus on two scenarios: 1) Multi-Layer Perception (MLP) for modeling tabular data; 2) language models for modeling textual data. On tabular data, we use the local piece-wise interpretation in DNN as the representation of each sample, and directly run K-mediods clustering to select and label the central sample in each cluster. On textual data, we propose a novel aggregator to find the most informative word in each sentence, and use its local piece-wise interpretation as the representation of the sentence. To be noted, this whole process of DAMI does not require any hyper-parameters to tune manually. To verify the effectiveness of our approach, extensive experiments have been conducted on both tabular datasets and textual datasets. The experimental results demonstrate that DAMI constantly outperforms several state-of-the-art compared methods.

READ FULL TEXT

page 6

page 7

research
08/15/2021

Deep Active Learning for Text Classification with Diverse Interpretations

Recently, Deep Neural Networks (DNNs) have made remarkable progress for ...
research
08/22/2020

DNN2LR: Interpretation-inspired Feature Crossing for Real-world Tabular Data

For sake of reliability, it is necessary for models in real-world applic...
research
12/20/2022

Temporal Output Discrepancy for Loss Estimation-based Active Learning

While deep learning succeeds in a wide range of tasks, it highly depends...
research
04/10/2020

State-Relabeling Adversarial Active Learning

Active learning is to design label-efficient algorithms by sampling the ...
research
06/13/2022

Efficient Human-in-the-loop System for Guiding DNNs Attention

Attention guidance is an approach to addressing dataset bias in deep lea...
research
07/27/2020

Active Learning for Video Description With Cluster-Regularized Ensemble Ranking

Automatic video captioning aims to train models to generate text descrip...
research
08/07/2021

Reducing Annotating Load: Active Learning with Synthetic Images in Surgical Instrument Segmentation

Accurate instrument segmentation in endoscopic vision of robot-assisted ...

Please sign up or login with your details

Forgot password? Click here to reset