ScreenQA: Large-Scale Question-Answer Pairs over Mobile App Screenshots

09/16/2022
by   Yu-Chung Hsiao, et al.
13

We present a new task and dataset, ScreenQA, for screen content understanding via question answering. The existing screen datasets are focused either on structure and component-level understanding, or on a much higher-level composite task such as navigation and task completion. We attempt to bridge the gap between these two by annotating 80,000+ question-answer pairs over the RICO dataset in hope to benchmark the screen reading comprehension capacity.

READ FULL TEXT

page 3

page 8

page 13

research
05/03/2023

NorQuAD: Norwegian Question Answering Dataset

In this paper we present NorQuAD: the first Norwegian question answering...
research
09/17/2021

CodeQA: A Question Answering Dataset for Source Code Comprehension

We propose CodeQA, a free-form question answering dataset for the purpos...
research
01/27/2021

VisualMRC: Machine Reading Comprehension on Document Images

Recent studies on machine reading comprehension have focused on text-lev...
research
02/04/2022

Interactive Mobile App Navigation with Uncertain or Under-specified Natural Language Commands

We introduce Mobile app Tasks with Iterative Feedback (MoTIF), a new dat...
research
12/06/2022

Dataset vs Reality: Understanding Model Performance from the Perspective of Information Need

Deep learning technologies have brought us many models that outperform h...
research
09/24/2018

Stochastic Answer Networks for SQuAD 2.0

This paper presents an extension of the Stochastic Answer Network (SAN),...
research
03/04/2016

Text Understanding with the Attention Sum Reader Network

Several large cloze-style context-question-answer datasets have been int...

Please sign up or login with your details

Forgot password? Click here to reset