(Psycho-)Linguistic Features Meet Transformer Models for Improved Explainable and Controllable Text Simplification

12/19/2022
by   Yu Qiao, et al.
0

State-of-the-art text simplification (TS) systems adopt end-to-end neural network models to directly generate the simplified version of the input text, and usually function as a blackbox. Moreover, TS is usually treated as an all-purpose generic task under the assumption of homogeneity, where the same simplification is suitable for all. In recent years, however, there has been increasing recognition of the need to adapt the simplification techniques to the specific needs of different target groups. In this work, we aim to advance current research on explainable and controllable TS in two ways: First, building on recently proposed work to increase the transparency of TS systems, we use a large set of (psycho-)linguistic features in combination with pre-trained language models to improve explainable complexity prediction. Second, based on the results of this preliminary task, we extend a state-of-the-art Seq2Seq TS model, ACCESS, to enable explicit control of ten attributes. The results of experiments show (1) that our approach improves the performance of state-of-the-art models for predicting explainable complexity and (2) that explicitly conditioning the Seq2Seq model on ten attributes leads to a significant improvement in performance in both within-domain and out-of-domain settings.

READ FULL TEXT

page 19

page 20

page 21

page 22

research
09/21/2021

TrOCR: Transformer-based Optical Character Recognition with Pre-trained Models

Text recognition is a long-standing research problem for document digita...
research
07/31/2020

An Empirical Study on Explainable Prediction of Text Complexity: Preliminaries for Text Simplification

Text simplification is concerned with reducing the language complexity a...
research
10/07/2019

Controllable Sentence Simplification

Text simplification aims at making a text easier to read and understand ...
research
05/07/2023

Stanford MLab at SemEval-2023 Task 10: Exploring GloVe- and Transformer-Based Methods for the Explainable Detection of Online Sexism

In this paper, we discuss the methods we applied at SemEval-2023 Task 10...
research
09/08/2022

Towards explainable evaluation of language models on the semantic similarity of visual concepts

Recent breakthroughs in NLP research, such as the advent of Transformer ...
research
07/05/2023

Multilingual Controllable Transformer-Based Lexical Simplification

Text is by far the most ubiquitous source of knowledge and information a...
research
10/21/2020

Controllable Text Simplification with Explicit Paraphrasing

Text Simplification improves the readability of sentences through severa...

Please sign up or login with your details

Forgot password? Click here to reset