IconQA

Icon Question Answering

Dataset Information
Modalities
Images, Texts
Languages
English
Introduced
2021
License
Homepage

Overview

Current visual question answering (VQA) tasks mainly consider answering human-annotated questions for natural images in the daily-life context. Icon question answering (IconQA) is a benchmark which aims to highlight the importance of abstract diagram understanding and comprehensive cognitive reasoning in real-world diagram word problems. For this benchmark, a large-scale IconQA dataset is built that consists of three sub-tasks: multi-image-choice, multi-text-choice, and filling-in-the-blank. Compared to existing VQA benchmarks, IconQA requires not only perception skills like object recognition and text understanding, but also diverse cognitive reasoning skills, such as geometric reasoning, commonsense reasoning, and arithmetic reasoning.

Description from: IconQA

Variants: IconQA

Associated Benchmarks

This dataset is used in 1 benchmark:

  • Visual Question Answering (VQA) -

Recent Benchmark Submissions

Task Model Paper Date
Visual Question Answering (VQA) Patch-TRM IconQA: A New Benchmark for … 2021-10-25
Visual Question Answering (VQA) ViLT IconQA: A New Benchmark for … 2021-10-25
Visual Question Answering (VQA) ViT IconQA: A New Benchmark for … 2021-10-25
Visual Question Answering (VQA) UNITER IconQA: A New Benchmark for … 2021-10-25
Visual Question Answering (VQA) DFAF IconQA: A New Benchmark for … 2021-10-25
Visual Question Answering (VQA) MCAN IconQA: A New Benchmark for … 2021-10-25
Visual Question Answering (VQA) ViLBERT IconQA: A New Benchmark for … 2021-10-25
Visual Question Answering (VQA) BAN IconQA: A New Benchmark for … 2021-10-25
Visual Question Answering (VQA) Top-Down IconQA: A New Benchmark for … 2021-10-25
Visual Question Answering (VQA) Random IconQA: A New Benchmark for … 2021-10-25
Visual Question Answering (VQA) Q-Only IconQA: A New Benchmark for … 2021-10-25
Visual Question Answering (VQA) I-Only IconQA: A New Benchmark for … 2021-10-25

Research Papers

Recent papers with results on this dataset: