ML Research Wiki / Benchmarks / Question Answering / SIQA

SIQA

Question Answering Benchmark

Performance Over Time

📊 Showing 24 results | 📏 Metric: Accuracy

Top Performing Models

Rank Model Paper Accuracy Date Code
1 Unicorn 11B (fine-tuned) UNICORN on RAINBOW: A Universal Commonsense Reasoning Model on a New Multitask Benchmark 83.20 2021-03-24 📦 allenai/rainbow
2 LLaMA-2 13B + MixLoRA MixLoRA: Enhancing Large Language Models Fine-Tuning with LoRA-based Mixture of Experts 82.50 2024-04-22 📦 TUDB-Labs/MixLoRA 📦 mikecovlee/mLoRA
3 CompassMTL 567M with Tailor Task Compass: Scaling Multi-task Pre-training with Task Prefix 82.20 2022-10-12 📦 cooelf/compassmtl
4 CompassMTL 567M Task Compass: Scaling Multi-task Pre-training with Task Prefix 81.70 2022-10-12 📦 cooelf/compassmtl
5 LLaMA-3 8B+MoSLoRA (fine-tuned) Mixture-of-Subspaces in Low-Rank Adaptation 81.00 2024-06-16 📦 wutaiqiang/moslora
6 DeBERTa-Large 304M Two is Better than Many? Binary Classification as an Effective Approach to Multi-Choice Question Answering 80.20 2022-10-29 📦 declare-lab/team
7 DeBERTa-Large 304M (classification-based) Two is Better than Many? Binary Classification as an Effective Approach to Multi-Choice Question Answering 79.90 2022-10-29 📦 declare-lab/team
8 UnifiedQA 3B UnifiedQA: Crossing Format Boundaries With a Single QA System 79.80 2020-05-02 📦 allenai/unifiedqa 📦 facebookresearch/metaicl
9 ExDeBERTa 567M Task Compass: Scaling Multi-task Pre-training with Task Prefix 79.60 2022-10-12 📦 cooelf/compassmtl
10 LLaMA-3 8B + MixLoRA MixLoRA: Enhancing Large Language Models Fine-Tuning with LoRA-based Mixture of Experts 78.80 2024-04-22 📦 TUDB-Labs/MixLoRA 📦 mikecovlee/mLoRA

All Papers (24)