FR: Folded Rationalization with a Unified Encoder

09/17/2022
by   Wei Liu, et al.
3

Conventional works generally employ a two-phase model in which a generator selects the most important pieces, followed by a predictor that makes predictions based on the selected pieces. However, such a two-phase model may incur the degeneration problem where the predictor overfits to the noise generated by a not yet well-trained generator and in turn, leads the generator to converge to a sub-optimal model that tends to select senseless pieces. To tackle this challenge, we propose Folded Rationalization (FR) that folds the two phases of the rationale model into one from the perspective of text semantic extraction. The key idea of FR is to employ a unified encoder between the generator and predictor, based on which FR can facilitate a better predictor by access to valuable information blocked by the generator in the traditional two-phase model and thus bring a better generator. Empirically, we show that FR improves the F1 score by up to 10.3 state-of-the-art methods.

READ FULL TEXT

page 2

page 3

page 5

research
05/08/2023

MGR: Multi-generator based Rationalization

Rationalization is to employ a generator and a predictor to construct a ...
research
05/23/2023

Decoupled Rationalization with Asymmetric Learning Rates: A Flexible Lipshitz Restraint

A self-explaining rationalization model is generally constructed by a co...
research
12/16/2020

Learning from the Best: Rationalizing Prediction by Adversarial Information Calibration

Explaining the predictions of AI models is paramount in safety-critical ...
research
01/15/2023

Rationalizing Predictions by Adversarial Information Calibration

Explaining the predictions of AI models is paramount in safety-critical ...
research
05/27/2023

Unsupervised Selective Rationalization with Noise Injection

A major issue with using deep learning models in sensitive applications ...
research
10/26/2021

Understanding Interlocking Dynamics of Cooperative Rationalization

Selective rationalization explains the prediction of complex neural netw...
research
06/17/2021

On Anytime Learning at Macroscale

Classical machine learning frameworks assume access to a possibly large ...

Please sign up or login with your details

Forgot password? Click here to reset