Hierarchical Transformers for Long Document Classification

10/23/2019
by   Raghavendra Pappagari, et al.
0

BERT, which stands for Bidirectional Encoder Representations from Transformers, is a recently introduced language representation model based upon the transfer learning paradigm. We extend its fine-tuning procedure to address one of its major limitations - applicability to inputs longer than a few hundred words, such as transcripts of human call conversations. Our method is conceptually simple. We segment the input into smaller chunks and feed each of them into the base model. Then, we propagate each output through a single recurrent layer, or another transformer, followed by a softmax activation. We obtain the final classification decision after the last segment has been consumed. We show that both BERT extensions are quick to fine-tune and converge after as little as 1 epoch of training on a small, domain-specific data set. We successfully apply them in three different tasks involving customer call satisfaction prediction and topic classification, and obtain a significant improvement over the baseline models in two of them.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/18/2022

Hierarchical Neural Network Approaches for Long Document Classification

Text classification algorithms investigate the intricate relationships b...
research
11/20/2020

Fine-Tuning BERT for Sentiment Analysis of Vietnamese Reviews

Sentiment analysis is an important task in the field ofNature Language P...
research
02/28/2019

BERT for Joint Intent Classification and Slot Filling

Intent classification and slot filling are two essential tasks for natur...
research
08/19/2021

How Hateful are Movies? A Study and Prediction on Movie Subtitles

In this research, we investigate techniques to detect hate speech in mov...
research
09/13/2022

CNN-Trans-Enc: A CNN-Enhanced Transformer-Encoder On Top Of Static BERT representations for Document Classification

BERT achieves remarkable results in text classification tasks, it is yet...
research
03/25/2022

A Comparative Evaluation Of Transformer Models For De-Identification Of Clinical Text Data

Objective: To comparatively evaluate several transformer model architect...
research
02/17/2020

A Financial Service Chatbot based on Deep Bidirectional Transformers

We develop a chatbot using Deep Bidirectional Transformer models (BERT) ...

Please sign up or login with your details

Forgot password? Click here to reset