Papers
arxiv:2209.08199

ScreenQA: Large-Scale Question-Answer Pairs over Mobile App Screenshots

Published on Sep 16, 2022
Authors:
,
,

Abstract

We present a new task and dataset, ScreenQA, for screen content understanding via question answering. The existing screen datasets are focused either on structure and component-level understanding, or on a much higher-level composite task such as navigation and task completion. We attempt to bridge the gap between these two by annotating 86K question-answer pairs over the RICO dataset in hope to benchmark the screen reading comprehension capacity.

Community

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2209.08199 in a model README.md to link it from this page.

Datasets citing this paper 1

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2209.08199 in a Space README.md to link it from this page.

Collections including this paper 1