MRCLens: an MRC Dataset Bias Detection Toolkit

07/18/2022
by   Yifan Zhong, et al.
23

Many recent neural models have shown remarkable empirical results in Machine Reading Comprehension, but evidence suggests sometimes the models take advantage of dataset biases to predict and fail to generalize on out-of-sample data. While many other approaches have been proposed to address this issue from the computation perspective such as new architectures or training procedures, we believe a method that allows researchers to discover biases, and adjust the data or the models in an earlier stage will be beneficial. Thus, we introduce MRCLens, a toolkit that detects whether biases exist before users train the full model. For the convenience of introducing the toolkit, we also provide a categorization of common biases in MRC.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/28/2019

Sogou Machine Reading Comprehension Toolkit

Machine reading comprehension have been intensively studied in recent ye...
research
10/04/2022

Text Characterization Toolkit

In NLP, models are usually evaluated by reporting single-number performa...
research
02/11/2020

ReClor: A Reading Comprehension Dataset Requiring Logical Reasoning

Recent powerful pre-trained language models have achieved remarkable per...
research
05/24/2021

Using Adversarial Attacks to Reveal the Statistical Bias in Machine Reading Comprehension Models

Pre-trained language models have achieved human-level performance on man...
research
12/21/2018

NeuroX: A Toolkit for Analyzing Individual Neurons in Neural Networks

We present a toolkit to facilitate the interpretation and understanding ...
research
10/22/2021

Challenges in Procedural Multimodal Machine Comprehension:A Novel Way To Benchmark

We focus on Multimodal Machine Reading Comprehension (M3C) where a model...

Please sign up or login with your details

Forgot password? Click here to reset