Meta-learning Pathologies from Radiology Reports using Variance Aware Prototypical Networks

10/22/2022
by   Arijit Sehanobish, et al.
0

Large pretrained Transformer-based language models like BERT and GPT have changed the landscape of Natural Language Processing (NLP). However, fine tuning such models still requires a large number of training examples for each target task, thus annotating multiple datasets and training these models on various downstream tasks becomes time consuming and expensive. In this work, we propose a simple extension of the Prototypical Networks for few-shot text classification. Our main idea is to replace the class prototypes by Gaussians and introduce a regularization term that encourages the examples to be clustered near the appropriate class centroids. Experimental results show that our method outperforms various strong baselines on 13 public and 4 internal datasets. Furthermore, we use the class distributions as a tool for detecting potential out-of-distribution (OOD) data points during deployment.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/27/2022

Grad2Task: Improved Few-shot Text Classification Using Gradients for Task Representation

Large pretrained language models (LMs) like BERT have improved performan...
research
11/10/2019

Learning to Few-Shot Learn Across Diverse Natural Language Classification Tasks

Self-supervised pre-training of transformer models has shown enormous su...
research
07/13/2023

Unsupervised Calibration through Prior Adaptation for Text Classification using Large Language Models

A wide variety of natural language tasks are currently being addressed w...
research
08/29/2023

TransPrompt v2: A Transferable Prompting Framework for Cross-task Text Classification

Text classification is one of the most imperative tasks in natural langu...
research
05/24/2021

PTR: Prompt Tuning with Rules for Text Classification

Fine-tuned pre-trained language models (PLMs) have achieved awesome perf...
research
10/08/2020

Discriminatively-Tuned Generative Classifiers for Robust Natural Language Inference

While discriminative neural network classifiers are generally preferred,...
research
07/27/2023

Metric-Based In-context Learning: A Case Study in Text Simplification

In-context learning (ICL) for large language models has proven to be a p...

Please sign up or login with your details

Forgot password? Click here to reset