MPNet-multilingual
|
MTEB: Massive Text Embedding Benchmark
|
31.57
|
2022-10-13
|
|
ST5-Base
|
MTEB: Massive Text Embedding Benchmark
|
31.39
|
2022-10-13
|
|
SimCSE-BERT-unsup
|
MTEB: Massive Text Embedding Benchmark
|
31.15
|
2022-10-13
|
|
MiniLM-L6
|
MTEB: Massive Text Embedding Benchmark
|
30.81
|
2022-10-13
|
|
MiniLM-L12-multilingual
|
MTEB: Massive Text Embedding Benchmark
|
30.67
|
2022-10-13
|
|
GTR-XXL
|
MTEB: Massive Text Embedding Benchmark
|
30.64
|
2022-10-13
|
|
Komninos
|
MTEB: Massive Text Embedding Benchmark
|
30.49
|
2022-10-13
|
|
Contriever
|
MTEB: Massive Text Embedding Benchmark
|
30.36
|
2022-10-13
|
|
SGPT-125M-nli
|
MTEB: Massive Text Embedding Benchmark
|
30.26
|
2022-10-13
|
|
GTR-XL
|
MTEB: Massive Text Embedding Benchmark
|
30.21
|
2022-10-13
|
|
ST5-XXL
|
MTEB: Massive Text Embedding Benchmark
|
30.08
|
2022-10-13
|
|
ST5-XL
|
MTEB: Massive Text Embedding Benchmark
|
29.91
|
2022-10-13
|
|
BERT
|
MTEB: Massive Text Embedding Benchmark
|
29.82
|
2022-10-13
|
|
GTR-Base
|
MTEB: Massive Text Embedding Benchmark
|
29.67
|
2022-10-13
|
|
ST5-Large
|
MTEB: Massive Text Embedding Benchmark
|
29.64
|
2022-10-13
|
|
coCondenser-msmarco
|
MTEB: Massive Text Embedding Benchmark
|
29.50
|
2022-10-13
|
|
Glove
|
MTEB: Massive Text Embedding Benchmark
|
28.87
|
2022-10-13
|
|
MiniLM-L12
|
MTEB: Massive Text Embedding Benchmark
|
27.90
|
2022-10-13
|
|
SPECTER
|
MTEB: Massive Text Embedding Benchmark
|
27.66
|
2022-10-13
|
|
MPNet
|
MTEB: Massive Text Embedding Benchmark
|
27.49
|
2022-10-13
|
|
Ada Similarity
|
MTEB: Massive Text Embedding Benchmark
|
26.94
|
2022-10-13
|
|
LASER2
|
MTEB: Massive Text Embedding Benchmark
|
26.80
|
2022-10-13
|
|
SGPT-1.3B-msmarco
|
MTEB: Massive Text Embedding Benchmark
|
25.44
|
2022-10-13
|
|
SGPT-BLOOM-7.1B-msmarco
|
MTEB: Massive Text Embedding Benchmark
|
24.99
|
2022-10-13
|
|
SGPT-5.8B-msmarco
|
MTEB: Massive Text Embedding Benchmark
|
24.75
|
2022-10-13
|
|
SimCSE-BERT-sup
|
MTEB: Massive Text Embedding Benchmark
|
23.31
|
2022-10-13
|
|