PaLM 540B (finetuned)
|
PaLM: Scaling Language Modeling with Pathways
|
90.10
|
2022-04-05
|
|
ST-MoE-32B 269B (fine-tuned)
|
ST-MoE: Designing Stable and Transferable Sparse …
|
89.60
|
2022-02-17
|
|
Turing NLR v5 XXL 5.4B (fine-tuned)
|
Toward Efficient Language Model Pretraining and D…
|
88.40
|
2022-12-04
|
|
DeBERTa-1.5B
|
DeBERTa: Decoding-enhanced BERT with Disentangled…
|
88.20
|
2020-06-05
|
|
Vega v2 6B (fine-tuned)
|
Toward Efficient Language Model Pretraining and D…
|
88.20
|
2022-12-04
|
|
PaLM 2-L (one-shot)
|
PaLM 2 Technical Report
|
88.20
|
2023-05-17
|
|
T5-XXL 11B (fine-tuned)
|
Exploring the Limits of Transfer Learning with a …
|
88.10
|
2019-10-23
|
|
ST-MoE-L 4.1B (fine-tuned)
|
ST-MoE: Designing Stable and Transferable Sparse …
|
86.00
|
2022-02-17
|
|
PaLM 2-M (one-shot)
|
PaLM 2 Technical Report
|
84.10
|
2023-05-17
|
|
PaLM 2-S (one-shot)
|
PaLM 2 Technical Report
|
84.00
|
2023-05-17
|
|
FLAN 137B (prompt-tuned)
|
Finetuned Language Models Are Zero-Shot Learners
|
83.40
|
2021-09-03
|
|
FLAN 137B (zero-shot)
|
Finetuned Language Models Are Zero-Shot Learners
|
77.50
|
2021-09-03
|
|
GPT-3 175B (Few-Shot)
|
Language Models are Few-Shot Learners
|
75.40
|
2020-05-28
|
|
FLAN 137B (1-shot)
|
Finetuned Language Models Are Zero-Shot Learners
|
72.10
|
2021-09-03
|
|
KELM (finetuning BERT-large based single model)
|
KELM: Knowledge Enhanced Pre-Trained Language Rep…
|
70.80
|
2021-09-09
|
|
BERT-large(single model)
|
BERT: Pre-training of Deep Bidirectional Transfor…
|
70.00
|
2018-10-11
|
|
Neo-6B (QA + WS)
|
Ask Me Anything: A simple strategy for prompting …
|
63.80
|
2022-10-05
|
|
T5-11B
|
Exploring the Limits of Transfer Learning with a …
|
63.30
|
2019-10-23
|
|
Bloomberg GPT 50B (1-shot)
|
BloombergGPT: A Large Language Model for Finance
|
62.30
|
2023-03-30
|
|
N-Grammer 343M
|
N-Grammer: Augmenting Transformers with latent n-…
|
62.00
|
2022-07-13
|
|
Neo-6B (few-shot)
|
Ask Me Anything: A simple strategy for prompting …
|
60.80
|
2022-10-05
|
|
Hybrid H3 355M (3-shot, logit scoring)
|
Hungry Hungry Hippos: Towards Language Modeling w…
|
59.70
|
2022-12-28
|
|
AlexaTM 20B
|
AlexaTM 20B: Few-Shot Learning Using a Large-Scal…
|
59.60
|
2022-08-02
|
|
Hybrid H3 355M (0-shot, logit scoring)
|
Hungry Hungry Hippos: Towards Language Modeling w…
|
59.50
|
2022-12-28
|
|
Neo-6B (QA)
|
Ask Me Anything: A simple strategy for prompting …
|
58.80
|
2022-10-05
|
|
Hybrid H3 125M (0-shot, logit scoring)
|
Hungry Hungry Hippos: Towards Language Modeling w…
|
51.40
|
2022-12-28
|
|
Hybrid H3 125M (3-shot, logit scoring)
|
Hungry Hungry Hippos: Towards Language Modeling w…
|
48.90
|
2022-12-28
|
|
BLOOM 176B (1-shot)
|
BloombergGPT: A Large Language Model for Finance
|
26.70
|
2023-03-30
|
|
GPT-NeoX 20B (1-shot)
|
BloombergGPT: A Large Language Model for Finance
|
22.90
|
2023-03-30
|
|
OPT 66B (1-shot)
|
BloombergGPT: A Large Language Model for Finance
|
18.80
|
2023-03-30
|
|