Icon Question Answering
Current visual question answering (VQA) tasks mainly consider answering human-annotated questions for natural images in the daily-life context. Icon question answering (IconQA) is a benchmark which aims to highlight the importance of abstract diagram understanding and comprehensive cognitive reasoning in real-world diagram word problems. For this benchmark, a large-scale IconQA dataset is built that consists of three sub-tasks: multi-image-choice, multi-text-choice, and filling-in-the-blank. Compared to existing VQA benchmarks, IconQA requires not only perception skills like object recognition and text understanding, but also diverse cognitive reasoning skills, such as geometric reasoning, commonsense reasoning, and arithmetic reasoning.
Description from: IconQA
Variants: IconQA
This dataset is used in 1 benchmark:
Task | Model | Paper | Date |
---|---|---|---|
Visual Question Answering (VQA) | Patch-TRM | IconQA: A New Benchmark for … | 2021-10-25 |
Visual Question Answering (VQA) | ViLT | IconQA: A New Benchmark for … | 2021-10-25 |
Visual Question Answering (VQA) | ViT | IconQA: A New Benchmark for … | 2021-10-25 |
Visual Question Answering (VQA) | UNITER | IconQA: A New Benchmark for … | 2021-10-25 |
Visual Question Answering (VQA) | DFAF | IconQA: A New Benchmark for … | 2021-10-25 |
Visual Question Answering (VQA) | MCAN | IconQA: A New Benchmark for … | 2021-10-25 |
Visual Question Answering (VQA) | ViLBERT | IconQA: A New Benchmark for … | 2021-10-25 |
Visual Question Answering (VQA) | BAN | IconQA: A New Benchmark for … | 2021-10-25 |
Visual Question Answering (VQA) | Top-Down | IconQA: A New Benchmark for … | 2021-10-25 |
Visual Question Answering (VQA) | Random | IconQA: A New Benchmark for … | 2021-10-25 |
Visual Question Answering (VQA) | Q-Only | IconQA: A New Benchmark for … | 2021-10-25 |
Visual Question Answering (VQA) | I-Only | IconQA: A New Benchmark for … | 2021-10-25 |
Recent papers with results on this dataset: