QLEVR: A Diagnostic Dataset for Quantificational Language and Elementary Visual Reasoning
Research output: Chapter in Book/Report/Conference proceeding › Article in proceedings › Research › peer-review
Standard
QLEVR : A Diagnostic Dataset for Quantificational Language and Elementary Visual Reasoning. / Li, Zechen; Sogaard, Anders.
Findings of the Association for Computational Linguistics: NAACL 2022 - Findings. Association for Computational Linguistics (ACL), 2022. p. 980-996.Research output: Chapter in Book/Report/Conference proceeding › Article in proceedings › Research › peer-review
Harvard
APA
Vancouver
Author
Bibtex
}
RIS
TY - GEN
T1 - QLEVR
T2 - 2022 Findings of the Association for Computational Linguistics: NAACL 2022
AU - Li, Zechen
AU - Sogaard, Anders
N1 - Publisher Copyright: © Findings of the Association for Computational Linguistics: NAACL 2022 - Findings.
PY - 2022
Y1 - 2022
N2 - Synthetic datasets have successfully been used to probe visual question-answering datasets for their reasoning abilities. CLEVR (Johnson et al., 2017), for example, tests a range of visual reasoning abilities. The questions in CLEVR focus on comparisons of shapes, colors, and sizes, numerical reasoning, and existence claims. This paper introduces a minimally biased, diagnostic visual questionanswering dataset, QLEVR, that goes beyond existential and numerical quantification and focus on more complex quantifiers and their combinations, e.g., asking whether there are more than two red balls that are smaller than at least three blue balls in an image. We describe how the dataset was created and present a first evaluation of state-of-the-art visual question-answering models, showing that QLEVR presents a formidable challenge to our current models. Code and Dataset are available at https://github.com/ zechenli03/QLEVR.
AB - Synthetic datasets have successfully been used to probe visual question-answering datasets for their reasoning abilities. CLEVR (Johnson et al., 2017), for example, tests a range of visual reasoning abilities. The questions in CLEVR focus on comparisons of shapes, colors, and sizes, numerical reasoning, and existence claims. This paper introduces a minimally biased, diagnostic visual questionanswering dataset, QLEVR, that goes beyond existential and numerical quantification and focus on more complex quantifiers and their combinations, e.g., asking whether there are more than two red balls that are smaller than at least three blue balls in an image. We describe how the dataset was created and present a first evaluation of state-of-the-art visual question-answering models, showing that QLEVR presents a formidable challenge to our current models. Code and Dataset are available at https://github.com/ zechenli03/QLEVR.
UR - http://www.scopus.com/inward/record.url?scp=85137332365&partnerID=8YFLogxK
U2 - 10.18653/v1/2022.findings-naacl.73
DO - 10.18653/v1/2022.findings-naacl.73
M3 - Article in proceedings
AN - SCOPUS:85137332365
SP - 980
EP - 996
BT - Findings of the Association for Computational Linguistics
PB - Association for Computational Linguistics (ACL)
Y2 - 10 July 2022 through 15 July 2022
ER -
ID: 341493689