Methods for Estimating and Improving Robustness of Language Models

06/16/2022
by   Michal Štefánik, et al.
0

Despite their outstanding performance, large language models (LLMs) suffer notorious flaws related to their preference for simple, surface-level textual relations over full semantic complexity of the problem. This proposal investigates a common denominator of this problem in their weak ability to generalise outside of the training domain. We survey diverse research directions providing estimations of model generalisation ability and find that incorporating some of these measures in the training objectives leads to enhanced distributional robustness of neural models. Based on these findings, we present future research directions towards enhancing the robustness of LLMs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/25/2022

Shortcut Learning of Large Language Models in Natural Language Understanding: A Survey

Large language models (LLMs) have achieved state-of-the-art performance ...
research
07/12/2023

A Comprehensive Overview of Large Language Models

Large Language Models (LLMs) have shown excellent generalization capabil...
research
06/09/2019

A Survey on Neural Network Language Models

As the core component of Natural Language Processing (NLP) system, Langu...
research
03/08/2022

Adapt𝒪r: Objective-Centric Adaptation Framework for Language Models

Progress in natural language processing research is catalyzed by the pos...
research
05/28/2021

What Is Considered Complete for Visual Recognition?

This is an opinion paper. We hope to deliver a key message that current ...
research
05/21/2023

A PhD Student's Perspective on Research in NLP in the Era of Very Large Language Models

Recent progress in large language models has enabled the deployment of m...

Please sign up or login with your details

Forgot password? Click here to reset