ScreenQA: Large-Scale Question-Answer Pairs over Mobile App Screenshots

09/16/2022
by   Yu-Chung Hsiao, et al.
13

We present a new task and dataset, ScreenQA, for screen content understanding via question answering. The existing screen datasets are focused either on structure and component-level understanding, or on a much higher-level composite task such as navigation and task completion. We attempt to bridge the gap between these two by annotating 80,000+ question-answer pairs over the RICO dataset in hope to benchmark the screen reading comprehension capacity.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset