SMARTRoBERTa
|
SMART: Robust and Efficient Fine-Tuning for Pre-t…
|
92.80
|
2019-11-08
|
|
DeBERTa (large)
|
DeBERTa: Decoding-enhanced BERT with Disentangled…
|
92.50
|
2020-06-05
|
|
SMART-BERT
|
SMART: Robust and Efficient Fine-Tuning for Pre-t…
|
90.00
|
2019-11-08
|
|
MT-DNN-SMART
|
SMART: Robust and Efficient Fine-Tuning for Pre-t…
|
0.93
|
2019-11-08
|
|
StructBERTRoBERTa ensemble
|
StructBERT: Incorporating Language Structures int…
|
0.93
|
2019-08-13
|
|
Mnet-Sim
|
MNet-Sim: A Multi-layered Semantic Similarity Net…
|
0.93
|
2021-11-09
|
|
T5-11B
|
Exploring the Limits of Transfer Learning with a …
|
0.93
|
2019-10-23
|
|
ALBERT
|
ALBERT: A Lite BERT for Self-supervised Learning …
|
0.93
|
2019-09-26
|
|
XLNet (single model)
|
XLNet: Generalized Autoregressive Pretraining for…
|
0.93
|
2019-06-19
|
|
RoBERTa
|
RoBERTa: A Robustly Optimized BERT Pretraining Ap…
|
0.92
|
2019-07-26
|
|
RoBERTa-large 355M (MLP quantized vector-wise, fine-tuned)
|
LLM.int8(): 8-bit Matrix Multiplication for Trans…
|
0.92
|
2022-08-15
|
|
PSQ (Chen et al., 2020)
|
A Statistical Framework for Low-bitwidth Training…
|
0.92
|
2020-10-27
|
|
RoBERTa-large 355M + Entailment as Few-shot Learner
|
Entailment as Few-Shot Learner
|
0.92
|
2021-04-29
|
|
ERNIE 2.0 Large
|
ERNIE 2.0: A Continual Pre-training Framework for…
|
0.91
|
2019-07-29
|
|
Q-BERT (Shen et al., 2020)
|
Q-BERT: Hessian Based Ultra Low Precision Quantiz…
|
0.91
|
2019-09-12
|
|
Q8BERT (Zafrir et al., 2019)
|
Q8BERT: Quantized 8Bit BERT
|
0.91
|
2019-10-14
|
|
DistilBERT 66M
|
DistilBERT, a distilled version of BERT: smaller,…
|
0.91
|
2019-10-02
|
|
T5-3B
|
Exploring the Limits of Transfer Learning with a …
|
0.91
|
2019-10-23
|
|
MLM+ del-word
|
CLEAR: Contrastive Learning for Sentence Represen…
|
0.91
|
2020-12-31
|
|
RealFormer
|
RealFormer: Transformer Likes Residual Attention
|
0.90
|
2020-12-21
|
|
T5-Large
|
Exploring the Limits of Transfer Learning with a …
|
0.90
|
2019-10-23
|
|
SpanBERT
|
SpanBERT: Improving Pre-training by Representing …
|
0.90
|
2019-07-24
|
|
AnglE-LLaMA-13B
|
AnglE-optimized Text Embeddings
|
0.90
|
2023-09-22
|
|
T5-Base
|
Exploring the Limits of Transfer Learning with a …
|
0.89
|
2019-10-23
|
|
ASA + RoBERTa
|
Adversarial Self-Attention for Language Understan…
|
0.89
|
2022-06-25
|
|
PromptEOL+CSE+LLaMA-30B
|
Scaling Sentence Embeddings with Large Language M…
|
0.89
|
2023-07-31
|
|
AnglE-LLaMA-7B
|
AnglE-optimized Text Embeddings
|
0.89
|
2023-09-22
|
|
AnglE-LLaMA-7B-v2
|
AnglE-optimized Text Embeddings
|
0.89
|
2023-09-22
|
|
T5-Large 770M
|
Exploring the Limits of Transfer Learning with a …
|
0.89
|
2019-10-23
|
|
PromptEOL+CSE+OPT-13B
|
Scaling Sentence Embeddings with Large Language M…
|
0.89
|
2023-07-31
|
|
PromptEOL+CSE+OPT-2.7B
|
Scaling Sentence Embeddings with Large Language M…
|
0.88
|
2023-07-31
|
|
PromCSE-RoBERTa-large (0.355B)
|
Improved Universal Sentence Embeddings with Promp…
|
0.88
|
2022-03-14
|
|
BigBird
|
Big Bird: Transformers for Longer Sequences
|
0.88
|
2020-07-28
|
|
ERNIE 2.0 Base
|
ERNIE 2.0: A Continual Pre-training Framework for…
|
0.88
|
2019-07-29
|
|
Charformer-Tall
|
Charformer: Fast Character Transformers via Gradi…
|
0.87
|
2021-06-23
|
|
SimCSE-RoBERTalarge
|
SimCSE: Simple Contrastive Learning of Sentence E…
|
0.87
|
2021-04-18
|
|
Trans-Encoder-RoBERTa-large-cross (unsup.)
|
Trans-Encoder: Unsupervised sentence-pair modelli…
|
0.87
|
2021-09-27
|
|
Trans-Encoder-RoBERTa-large-bi (unsup.)
|
Trans-Encoder: Unsupervised sentence-pair modelli…
|
0.87
|
2021-09-27
|
|
BERT-LARGE
|
BERT: Pre-training of Deep Bidirectional Transfor…
|
0.87
|
2018-10-11
|
|
ASA + BERT-base
|
Adversarial Self-Attention for Language Understan…
|
0.87
|
2022-06-25
|
|
Trans-Encoder-BERT-large-bi (unsup.)
|
Trans-Encoder: Unsupervised sentence-pair modelli…
|
0.86
|
2021-09-27
|
|
SRoBERTa-NLI-STSb-large
|
Sentence-BERT: Sentence Embeddings using Siamese …
|
0.86
|
2019-08-27
|
|
T5-Small
|
Exploring the Limits of Transfer Learning with a …
|
0.86
|
2019-10-23
|
|
SBERT-STSb-base
|
Sentence-BERT: Sentence Embeddings using Siamese …
|
0.85
|
2019-08-27
|
|
Trans-Encoder-RoBERTa-base-cross (unsup.)
|
Trans-Encoder: Unsupervised sentence-pair modelli…
|
0.85
|
2021-09-27
|
|
SBERT-STSb-large
|
Sentence-BERT: Sentence Embeddings using Siamese …
|
0.84
|
2019-08-27
|
|
FNet-Large
|
FNet: Mixing Tokens with Fourier Transforms
|
0.84
|
2021-05-09
|
|
Trans-Encoder-BERT-base-bi (unsup.)
|
Trans-Encoder: Unsupervised sentence-pair modelli…
|
0.84
|
2021-09-27
|
|
ERNIE
|
ERNIE: Enhanced Language Representation with Info…
|
0.83
|
2019-05-17
|
|
24hBERT
|
How to Train BERT with an Academic Budget
|
0.82
|
2021-04-15
|
|
TinyBERT-4 14.5M
|
TinyBERT: Distilling BERT for Natural Language Un…
|
0.80
|
2019-09-23
|
|
SBERT-NLI-large
|
Sentence-BERT: Sentence Embeddings using Siamese …
|
0.79
|
2019-08-27
|
|
Mirror-RoBERTa-base (unsup.)
|
Fast, Effective, and Self-Supervised: Transformin…
|
0.79
|
2021-04-16
|
|
USE_T
|
Universal Sentence Encoder
|
0.78
|
2018-03-29
|
|
Dino (STSb/̄🦕)
|
Generating Datasets with Pretrained Language Mode…
|
0.78
|
2021-04-15
|
|
SRoBERTa-NLI-base
|
Sentence-BERT: Sentence Embeddings using Siamese …
|
0.78
|
2019-08-27
|
|
SBERT-NLI-base
|
Sentence-BERT: Sentence Embeddings using Siamese …
|
0.77
|
2019-08-27
|
|
Dino (STS/̄🦕)
|
Generating Datasets with Pretrained Language Mode…
|
0.77
|
2021-04-15
|
|
Mirror-BERT-base (unsup.)
|
Fast, Effective, and Self-Supervised: Transformin…
|
0.76
|
2021-04-16
|
|
BERTlarge-flow (target)
|
On the Sentence Embeddings from Pre-trained Langu…
|
0.72
|
2020-11-02
|
|
IS-BERT-NLI
|
An Unsupervised Sentence Embedding Method by Mutu…
|
0.69
|
2020-09-25
|
|
Rematch
|
Rematch: Robust and Efficient Matching of Local K…
|
0.67
|
2024-04-02
|
|