Accelerating Thematic Investment with Prompt Tuned Pretrained Language Models

Prompt Tuning is emerging as a scalable and cost-effective method to fine-tune Pretrained Language Models (PLMs). This study benchmarks the performance and computational efficiency of Prompt Tuning and baseline methods on a multi-label text classification task. This is applied to the use case of classifying companies into an investment firm's proprietary industry taxonomy, supporting their thematic investment strategy. Text-to-text classification with PLMs is frequently reported to outperform classification with a classification head, but has several limitations when applied to a multi-label classification problem where each label consists of multiple tokens: (a) Generated labels may not match any label in the industry taxonomy; (b) During fine-tuning, multiple labels must be provided in an arbitrary order; (c) The model provides a binary decision for each label, rather than an appropriate confidence score. Limitation (a) is addressed by applying constrained decoding using Trie Search, which slightly improves classification performance. All limitations (a), (b), and (c) are addressed by replacing the PLM's language head with a classification head. This improves performance significantly, while also reducing computational costs during inference. The results indicate the continuing need to adapt state-of-the-art methods to domain-specific tasks, even in the era of PLMs with strong generalization abilities.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/18/2022

HFT-ONLSTM: Hierarchical and Fine-Tuning Multi-label Text Classification

Many important classification problems in the real-world consist of a la...
research
05/03/2023

Using Language Models on Low-end Hardware

This paper evaluates the viability of using fixed language models for tr...
research
04/28/2022

HPT: Hierarchy-aware Prompt Tuning for Hierarchical Text Classification

Hierarchical text classification (HTC) is a challenging subtask of multi...
research
10/01/2021

Fast Multi-Resolution Transformer Fine-tuning for Extreme Multi-label Text Classification

Extreme multi-label text classification (XMC) seeks to find relevant lab...
research
08/26/2021

Fine-tuning Pretrained Language Models with Label Attention for Explainable Biomedical Text Classification

The massive growth of digital biomedical data is making biomedical text ...
research
10/24/2020

Large Scale Legal Text Classification Using Transformer Models

Large multi-label text classification is a challenging Natural Language ...
research
10/30/2022

Parameter-Efficient Tuning Makes a Good Classification Head

In recent years, pretrained models revolutionized the paradigm of natura...

Please sign up or login with your details

Forgot password? Click here to reset