Upload folder using huggingface_hub
Browse files
README.md
CHANGED
@@ -4,21 +4,15 @@ language:
|
|
4 |
- en
|
5 |
library_name: model2vec
|
6 |
license: mit
|
7 |
-
tags:
|
8 |
-
- embeddings
|
9 |
-
- static-embeddings
|
10 |
-
- mteb
|
11 |
model-index:
|
12 |
- name: M2V_base_glove_subword
|
13 |
results:
|
14 |
-
-
|
15 |
-
type: Classification
|
16 |
-
dataset:
|
17 |
-
name: MTEB AmazonCounterfactualClassification (en-ext)
|
18 |
-
type: mteb/amazon_counterfactual
|
19 |
config: en-ext
|
20 |
-
|
21 |
revision: e8379541af4e31359cca9fbcf4b00f2671dba205
|
|
|
|
|
22 |
metrics:
|
23 |
- type: accuracy
|
24 |
value: 66.4167916041979
|
@@ -32,14 +26,14 @@ model-index:
|
|
32 |
value: 72.84623161234782
|
33 |
- type: main_score
|
34 |
value: 66.4167916041979
|
35 |
-
|
36 |
type: Classification
|
37 |
-
|
38 |
-
name: MTEB AmazonCounterfactualClassification (en)
|
39 |
-
type: mteb/amazon_counterfactual
|
40 |
config: en
|
41 |
-
|
42 |
revision: e8379541af4e31359cca9fbcf4b00f2671dba205
|
|
|
|
|
43 |
metrics:
|
44 |
- type: accuracy
|
45 |
value: 68.044776119403
|
@@ -53,14 +47,14 @@ model-index:
|
|
53 |
value: 71.2243167389672
|
54 |
- type: main_score
|
55 |
value: 68.044776119403
|
56 |
-
|
57 |
type: Classification
|
58 |
-
|
59 |
-
name: MTEB AmazonPolarityClassification (default)
|
60 |
-
type: mteb/amazon_polarity
|
61 |
config: default
|
62 |
-
|
63 |
revision: e2d317d38cd51312af73b3d32a06d1a08b442046
|
|
|
|
|
64 |
metrics:
|
65 |
- type: accuracy
|
66 |
value: 67.21602499999999
|
@@ -74,14 +68,14 @@ model-index:
|
|
74 |
value: 66.68107362746888
|
75 |
- type: main_score
|
76 |
value: 67.21602499999999
|
77 |
-
|
78 |
type: Classification
|
79 |
-
|
80 |
-
name: MTEB AmazonReviewsClassification (en)
|
81 |
-
type: mteb/amazon_reviews_multi
|
82 |
config: en
|
83 |
-
|
84 |
revision: 1399c76144fd37290681b995c656ef9b2e06e26d
|
|
|
|
|
85 |
metrics:
|
86 |
- type: accuracy
|
87 |
value: 32.384
|
@@ -91,14 +85,14 @@ model-index:
|
|
91 |
value: 32.05276706247388
|
92 |
- type: main_score
|
93 |
value: 32.384
|
94 |
-
|
95 |
-
type:
|
96 |
-
|
97 |
-
name: MTEB ArguAna (default)
|
98 |
-
type: mteb/arguana
|
99 |
config: default
|
100 |
-
|
101 |
revision: c22ab2a51041ffd869aaddef7af8d8215647e41a
|
|
|
|
|
102 |
metrics:
|
103 |
- type: main_score
|
104 |
value: 29.599999999999998
|
@@ -382,14 +376,14 @@ model-index:
|
|
382 |
value: 28.733999999999998
|
383 |
- type: recall_at_5
|
384 |
value: 36.131
|
385 |
-
|
386 |
-
type:
|
387 |
-
|
388 |
-
name: MTEB ArxivClusteringP2P (default)
|
389 |
-
type: mteb/arxiv-clustering-p2p
|
390 |
config: default
|
391 |
-
|
392 |
revision: a122ad7f3f0291bf49cc6f4d32aa80929df69d5d
|
|
|
|
|
393 |
metrics:
|
394 |
- type: main_score
|
395 |
value: 35.46255145204994
|
@@ -397,14 +391,14 @@ model-index:
|
|
397 |
value: 35.46255145204994
|
398 |
- type: v_measure_std
|
399 |
value: 14.146815377034603
|
400 |
-
|
401 |
type: Clustering
|
402 |
-
|
403 |
-
name: MTEB ArxivClusteringS2S (default)
|
404 |
-
type: mteb/arxiv-clustering-s2s
|
405 |
config: default
|
406 |
-
|
407 |
revision: f910caf1a6075f7329cdf8c1a6135696f37dbd53
|
|
|
|
|
408 |
metrics:
|
409 |
- type: main_score
|
410 |
value: 26.34189987196252
|
@@ -412,14 +406,14 @@ model-index:
|
|
412 |
value: 26.34189987196252
|
413 |
- type: v_measure_std
|
414 |
value: 14.798697652139317
|
415 |
-
|
416 |
-
type:
|
417 |
-
|
418 |
-
name: MTEB AskUbuntuDupQuestions (default)
|
419 |
-
type: mteb/askubuntudupquestions-reranking
|
420 |
config: default
|
421 |
-
|
422 |
revision: 2000358ca161889fa9c082cb41daa8dcfb161a54
|
|
|
|
|
423 |
metrics:
|
424 |
- type: main_score
|
425 |
value: 52.85912447389551
|
@@ -439,14 +433,14 @@ model-index:
|
|
439 |
value: 19.44964881732576
|
440 |
- type: nAUC_mrr_std
|
441 |
value: 9.313450884539453
|
442 |
-
|
443 |
-
type:
|
444 |
-
|
445 |
-
name: MTEB BIOSSES (default)
|
446 |
-
type: mteb/biosses-sts
|
447 |
config: default
|
448 |
-
|
449 |
revision: d3fb88f8f02e40887cd149695127462bbcf29b4a
|
|
|
|
|
450 |
metrics:
|
451 |
- type: cosine_pearson
|
452 |
value: 73.38282679412139
|
@@ -466,14 +460,14 @@ model-index:
|
|
466 |
value: 73.38282679412139
|
467 |
- type: spearman
|
468 |
value: 75.59389113278942
|
469 |
-
|
470 |
-
type:
|
471 |
-
|
472 |
-
name: MTEB Banking77Classification (default)
|
473 |
-
type: mteb/banking77
|
474 |
config: default
|
475 |
-
|
476 |
revision: 0fd18e25b25c072e09e0d92ab615fda904d66300
|
|
|
|
|
477 |
metrics:
|
478 |
- type: accuracy
|
479 |
value: 71.38636363636363
|
@@ -483,14 +477,14 @@ model-index:
|
|
483 |
value: 71.55994805461263
|
484 |
- type: main_score
|
485 |
value: 71.38636363636363
|
486 |
-
|
487 |
-
type:
|
488 |
-
|
489 |
-
name: MTEB BiorxivClusteringP2P (default)
|
490 |
-
type: mteb/biorxiv-clustering-p2p
|
491 |
config: default
|
492 |
-
|
493 |
revision: 65b79d1d13f80053f67aca9498d9402c2d9f1f40
|
|
|
|
|
494 |
metrics:
|
495 |
- type: main_score
|
496 |
value: 31.47309865069476
|
@@ -498,14 +492,14 @@ model-index:
|
|
498 |
value: 31.47309865069476
|
499 |
- type: v_measure_std
|
500 |
value: 0.6360736715097297
|
501 |
-
|
502 |
type: Clustering
|
503 |
-
|
504 |
-
name: MTEB BiorxivClusteringS2S (default)
|
505 |
-
type: mteb/biorxiv-clustering-s2s
|
506 |
config: default
|
507 |
-
|
508 |
revision: 258694dd0231531bc1fd9de6ceb52a0853c6d908
|
|
|
|
|
509 |
metrics:
|
510 |
- type: main_score
|
511 |
value: 22.58199120148109
|
@@ -513,14 +507,14 @@ model-index:
|
|
513 |
value: 22.58199120148109
|
514 |
- type: v_measure_std
|
515 |
value: 1.1055877138914942
|
516 |
-
|
517 |
-
type:
|
518 |
-
|
519 |
-
name: MTEB CQADupstackAndroidRetrieval (default)
|
520 |
-
type: mteb/cqadupstack-android
|
521 |
config: default
|
522 |
-
|
523 |
revision: f46a197baaae43b4f621051089b82a364682dfeb
|
|
|
|
|
524 |
metrics:
|
525 |
- type: main_score
|
526 |
value: 28.518
|
@@ -804,14 +798,14 @@ model-index:
|
|
804 |
value: 26.183
|
805 |
- type: recall_at_5
|
806 |
value: 30.653000000000002
|
807 |
-
|
808 |
type: Retrieval
|
809 |
-
|
810 |
-
name: MTEB CQADupstackEnglishRetrieval (default)
|
811 |
-
type: mteb/cqadupstack-english
|
812 |
config: default
|
813 |
-
|
814 |
revision: ad9991cb51e31e31e430383c75ffb2885547b5f0
|
|
|
|
|
815 |
metrics:
|
816 |
- type: main_score
|
817 |
value: 24.698999999999998
|
@@ -1095,14 +1089,14 @@ model-index:
|
|
1095 |
value: 23.247999999999998
|
1096 |
- type: recall_at_5
|
1097 |
value: 25.611
|
1098 |
-
|
1099 |
type: Retrieval
|
1100 |
-
|
1101 |
-
name: MTEB CQADupstackGamingRetrieval (default)
|
1102 |
-
type: mteb/cqadupstack-gaming
|
1103 |
config: default
|
1104 |
-
|
1105 |
revision: 4885aa143210c98657558c04aaf3dc47cfb54340
|
|
|
|
|
1106 |
metrics:
|
1107 |
- type: main_score
|
1108 |
value: 34.837
|
@@ -1386,14 +1380,14 @@ model-index:
|
|
1386 |
value: 33.351
|
1387 |
- type: recall_at_5
|
1388 |
value: 39.345
|
1389 |
-
|
1390 |
type: Retrieval
|
1391 |
-
|
1392 |
-
name: MTEB CQADupstackGisRetrieval (default)
|
1393 |
-
type: mteb/cqadupstack-gis
|
1394 |
config: default
|
1395 |
-
|
1396 |
revision: 5003b3064772da1887988e05400cf3806fe491f2
|
|
|
|
|
1397 |
metrics:
|
1398 |
- type: main_score
|
1399 |
value: 15.889000000000001
|
@@ -1677,14 +1671,14 @@ model-index:
|
|
1677 |
value: 15.383
|
1678 |
- type: recall_at_5
|
1679 |
value: 18.174
|
1680 |
-
|
1681 |
type: Retrieval
|
1682 |
-
|
1683 |
-
name: MTEB CQADupstackMathematicaRetrieval (default)
|
1684 |
-
type: mteb/cqadupstack-mathematica
|
1685 |
config: default
|
1686 |
-
|
1687 |
revision: 90fceea13679c63fe563ded68f3b6f06e50061de
|
|
|
|
|
1688 |
metrics:
|
1689 |
- type: main_score
|
1690 |
value: 11.0
|
@@ -1968,14 +1962,14 @@ model-index:
|
|
1968 |
value: 9.735000000000001
|
1969 |
- type: recall_at_5
|
1970 |
value: 12.595999999999998
|
1971 |
-
|
1972 |
type: Retrieval
|
1973 |
-
|
1974 |
-
name: MTEB CQADupstackPhysicsRetrieval (default)
|
1975 |
-
type: mteb/cqadupstack-physics
|
1976 |
config: default
|
1977 |
-
|
1978 |
revision: 79531abbd1fb92d06c6d6315a0cbbbf5bb247ea4
|
|
|
|
|
1979 |
metrics:
|
1980 |
- type: main_score
|
1981 |
value: 22.671
|
@@ -2259,14 +2253,14 @@ model-index:
|
|
2259 |
value: 21.583
|
2260 |
- type: recall_at_5
|
2261 |
value: 24.937
|
2262 |
-
|
2263 |
type: Retrieval
|
2264 |
-
|
2265 |
-
name: MTEB CQADupstackProgrammersRetrieval (default)
|
2266 |
-
type: mteb/cqadupstack-programmers
|
2267 |
config: default
|
2268 |
-
|
2269 |
revision: 6184bc1440d2dbc7612be22b50686b8826d22b32
|
|
|
|
|
2270 |
metrics:
|
2271 |
- type: main_score
|
2272 |
value: 16.306
|
@@ -2550,27 +2544,27 @@ model-index:
|
|
2550 |
value: 14.935
|
2551 |
- type: recall_at_5
|
2552 |
value: 17.862000000000002
|
2553 |
-
|
2554 |
type: Retrieval
|
2555 |
-
|
2556 |
-
name: MTEB CQADupstackRetrieval (default)
|
2557 |
-
type: CQADupstackRetrieval_is_a_combined_dataset
|
2558 |
config: default
|
2559 |
-
|
2560 |
revision: CQADupstackRetrieval_is_a_combined_dataset
|
|
|
|
|
2561 |
metrics:
|
2562 |
- type: main_score
|
2563 |
value: 19.211500000000004
|
2564 |
- type: ndcg_at_10
|
2565 |
value: 19.211500000000004
|
2566 |
-
|
2567 |
type: Retrieval
|
2568 |
-
|
2569 |
-
name: MTEB CQADupstackStatsRetrieval (default)
|
2570 |
-
type: mteb/cqadupstack-stats
|
2571 |
config: default
|
2572 |
-
|
2573 |
revision: 65ac3a16b8e91f9cee4c9828cc7c335575432a2a
|
|
|
|
|
2574 |
metrics:
|
2575 |
- type: main_score
|
2576 |
value: 13.274
|
@@ -2854,14 +2848,14 @@ model-index:
|
|
2854 |
value: 12.113999999999999
|
2855 |
- type: recall_at_5
|
2856 |
value: 14.371
|
2857 |
-
|
2858 |
type: Retrieval
|
2859 |
-
|
2860 |
-
name: MTEB CQADupstackTexRetrieval (default)
|
2861 |
-
type: mteb/cqadupstack-tex
|
2862 |
config: default
|
2863 |
-
|
2864 |
revision: 46989137a86843e03a6195de44b09deda022eec7
|
|
|
|
|
2865 |
metrics:
|
2866 |
- type: main_score
|
2867 |
value: 10.994
|
@@ -3145,14 +3139,14 @@ model-index:
|
|
3145 |
value: 10.381
|
3146 |
- type: recall_at_5
|
3147 |
value: 12.354999999999999
|
3148 |
-
|
3149 |
type: Retrieval
|
3150 |
-
|
3151 |
-
name: MTEB CQADupstackUnixRetrieval (default)
|
3152 |
-
type: mteb/cqadupstack-unix
|
3153 |
config: default
|
3154 |
-
|
3155 |
revision: 6c6430d3a6d36f8d2a829195bc5dc94d7e063e53
|
|
|
|
|
3156 |
metrics:
|
3157 |
- type: main_score
|
3158 |
value: 17.415
|
@@ -3436,14 +3430,14 @@ model-index:
|
|
3436 |
value: 16.458000000000002
|
3437 |
- type: recall_at_5
|
3438 |
value: 18.087
|
3439 |
-
|
3440 |
type: Retrieval
|
3441 |
-
|
3442 |
-
name: MTEB CQADupstackWebmastersRetrieval (default)
|
3443 |
-
type: mteb/cqadupstack-webmasters
|
3444 |
config: default
|
3445 |
-
|
3446 |
revision: 160c094312a0e1facb97e55eeddb698c0abe3571
|
|
|
|
|
3447 |
metrics:
|
3448 |
- type: main_score
|
3449 |
value: 21.398
|
@@ -3727,14 +3721,14 @@ model-index:
|
|
3727 |
value: 20.148
|
3728 |
- type: recall_at_5
|
3729 |
value: 23.296
|
3730 |
-
|
3731 |
type: Retrieval
|
3732 |
-
|
3733 |
-
name: MTEB CQADupstackWordpressRetrieval (default)
|
3734 |
-
type: mteb/cqadupstack-wordpress
|
3735 |
config: default
|
3736 |
-
|
3737 |
revision: 4ffe81d471b1924886b33c7567bfb200e9eec5c4
|
|
|
|
|
3738 |
metrics:
|
3739 |
- type: main_score
|
3740 |
value: 13.536999999999999
|
@@ -4018,14 +4012,14 @@ model-index:
|
|
4018 |
value: 12.187000000000001
|
4019 |
- type: recall_at_5
|
4020 |
value: 14.868
|
4021 |
-
|
4022 |
type: Retrieval
|
4023 |
-
|
4024 |
-
name: MTEB ClimateFEVER (default)
|
4025 |
-
type: mteb/climate-fever
|
4026 |
config: default
|
4027 |
-
|
4028 |
revision: 47f2ac6acb640fc46020b02a5b59fdda04d39380
|
|
|
|
|
4029 |
metrics:
|
4030 |
- type: main_score
|
4031 |
value: 14.015
|
@@ -4309,14 +4303,14 @@ model-index:
|
|
4309 |
value: 9.713
|
4310 |
- type: recall_at_5
|
4311 |
value: 12.415
|
4312 |
-
|
4313 |
type: Retrieval
|
4314 |
-
|
4315 |
-
name: MTEB DBPedia (default)
|
4316 |
-
type: mteb/dbpedia
|
4317 |
config: default
|
4318 |
-
|
4319 |
revision: c0f706b76e590d620bd6618b3ca8efdd34e2d659
|
|
|
|
|
4320 |
metrics:
|
4321 |
- type: main_score
|
4322 |
value: 19.899
|
@@ -4600,14 +4594,14 @@ model-index:
|
|
4600 |
value: 6.425
|
4601 |
- type: recall_at_5
|
4602 |
value: 8.818
|
4603 |
-
|
4604 |
-
type:
|
4605 |
-
|
4606 |
-
name: MTEB EmotionClassification (default)
|
4607 |
-
type: mteb/emotion
|
4608 |
config: default
|
4609 |
-
|
4610 |
revision: 4f58c6b202a23cf9a4da393831edf4f9183cad37
|
|
|
|
|
4611 |
metrics:
|
4612 |
- type: accuracy
|
4613 |
value: 38.339999999999996
|
@@ -4617,14 +4611,14 @@ model-index:
|
|
4617 |
value: 40.51989104726522
|
4618 |
- type: main_score
|
4619 |
value: 38.339999999999996
|
4620 |
-
|
4621 |
-
type:
|
4622 |
-
|
4623 |
-
name: MTEB FEVER (default)
|
4624 |
-
type: mteb/fever
|
4625 |
config: default
|
4626 |
-
|
4627 |
revision: bea83ef9e8fb933d90a2f1d5515737465d613e12
|
|
|
|
|
4628 |
metrics:
|
4629 |
- type: main_score
|
4630 |
value: 25.006
|
@@ -4908,14 +4902,14 @@ model-index:
|
|
4908 |
value: 24.082
|
4909 |
- type: recall_at_5
|
4910 |
value: 29.563
|
4911 |
-
|
4912 |
type: Retrieval
|
4913 |
-
|
4914 |
-
name: MTEB FiQA2018 (default)
|
4915 |
-
type: mteb/fiqa
|
4916 |
config: default
|
4917 |
-
|
4918 |
revision: 27a168819829fe9bcd655c2df245fb19452e8e06
|
|
|
|
|
4919 |
metrics:
|
4920 |
- type: main_score
|
4921 |
value: 11.167
|
@@ -5199,14 +5193,14 @@ model-index:
|
|
5199 |
value: 8.863999999999999
|
5200 |
- type: recall_at_5
|
5201 |
value: 11.354000000000001
|
5202 |
-
|
5203 |
type: Retrieval
|
5204 |
-
|
5205 |
-
name: MTEB HotpotQA (default)
|
5206 |
-
type: mteb/hotpotqa
|
5207 |
config: default
|
5208 |
-
|
5209 |
revision: ab518f4d6fcca38d87c25209f94beba119d02014
|
|
|
|
|
5210 |
metrics:
|
5211 |
- type: main_score
|
5212 |
value: 30.837999999999997
|
@@ -5490,14 +5484,14 @@ model-index:
|
|
5490 |
value: 25.779999999999998
|
5491 |
- type: recall_at_5
|
5492 |
value: 29.250999999999998
|
5493 |
-
|
5494 |
-
type:
|
5495 |
-
|
5496 |
-
name: MTEB ImdbClassification (default)
|
5497 |
-
type: mteb/imdb
|
5498 |
config: default
|
5499 |
-
|
5500 |
revision: 3d86128a09e091d6018b6d26cad27f2739fc2db7
|
|
|
|
|
5501 |
metrics:
|
5502 |
- type: accuracy
|
5503 |
value: 66.5616
|
@@ -5511,14 +5505,14 @@ model-index:
|
|
5511 |
value: 66.15361405073978
|
5512 |
- type: main_score
|
5513 |
value: 66.5616
|
5514 |
-
|
5515 |
-
type:
|
5516 |
-
|
5517 |
-
name: MTEB MSMARCO (default)
|
5518 |
-
type: mteb/msmarco
|
5519 |
config: default
|
5520 |
-
|
5521 |
revision: c5a29a104738b98a9e76336939199e264163d4a0
|
|
|
|
|
5522 |
metrics:
|
5523 |
- type: main_score
|
5524 |
value: 28.034
|
@@ -5802,14 +5796,14 @@ model-index:
|
|
5802 |
value: 2.138
|
5803 |
- type: recall_at_5
|
5804 |
value: 3.3770000000000002
|
5805 |
-
|
5806 |
-
type:
|
5807 |
-
|
5808 |
-
name: MTEB MTOPDomainClassification (en)
|
5809 |
-
type: mteb/mtop_domain
|
5810 |
config: en
|
5811 |
-
|
5812 |
revision: d80d48c1eb48d3562165c59d59d0034df9fff0bf
|
|
|
|
|
5813 |
metrics:
|
5814 |
- type: accuracy
|
5815 |
value: 84.81988144094848
|
@@ -5819,14 +5813,14 @@ model-index:
|
|
5819 |
value: 84.95181538630469
|
5820 |
- type: main_score
|
5821 |
value: 84.81988144094848
|
5822 |
-
|
5823 |
type: Classification
|
5824 |
-
|
5825 |
-
name: MTEB MTOPIntentClassification (en)
|
5826 |
-
type: mteb/mtop_intent
|
5827 |
config: en
|
5828 |
-
|
5829 |
revision: ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba
|
|
|
|
|
5830 |
metrics:
|
5831 |
- type: accuracy
|
5832 |
value: 62.41222070223438
|
@@ -5836,14 +5830,14 @@ model-index:
|
|
5836 |
value: 66.23266420473301
|
5837 |
- type: main_score
|
5838 |
value: 62.41222070223438
|
5839 |
-
|
5840 |
type: Classification
|
5841 |
-
|
5842 |
-
name: MTEB MassiveIntentClassification (en)
|
5843 |
-
type: mteb/amazon_massive_intent
|
5844 |
config: en
|
5845 |
-
|
5846 |
revision: 4672e20407010da34463acc759c162ca9734bca6
|
|
|
|
|
5847 |
metrics:
|
5848 |
- type: accuracy
|
5849 |
value: 62.50168123739073
|
@@ -5853,14 +5847,14 @@ model-index:
|
|
5853 |
value: 62.787680759907204
|
5854 |
- type: main_score
|
5855 |
value: 62.50168123739073
|
5856 |
-
|
5857 |
type: Classification
|
5858 |
-
|
5859 |
-
name: MTEB MassiveScenarioClassification (en)
|
5860 |
-
type: mteb/amazon_massive_scenario
|
5861 |
config: en
|
5862 |
-
|
5863 |
revision: fad2c6e8459f9e1c45d9315f4953d921437d70f8
|
|
|
|
|
5864 |
metrics:
|
5865 |
- type: accuracy
|
5866 |
value: 66.09280430396772
|
@@ -5870,14 +5864,14 @@ model-index:
|
|
5870 |
value: 66.15203456480924
|
5871 |
- type: main_score
|
5872 |
value: 66.09280430396772
|
5873 |
-
|
5874 |
-
type:
|
5875 |
-
|
|
|
5876 |
name: MTEB MedrxivClusteringP2P (default)
|
|
|
|
|
5877 |
type: mteb/medrxiv-clustering-p2p
|
5878 |
-
config: default
|
5879 |
-
split: test
|
5880 |
-
revision: e7a26af6f3ae46b30dde8737f02c07b1505bcc73
|
5881 |
metrics:
|
5882 |
- type: main_score
|
5883 |
value: 26.932942933622616
|
@@ -5885,14 +5879,14 @@ model-index:
|
|
5885 |
value: 26.932942933622616
|
5886 |
- type: v_measure_std
|
5887 |
value: 1.593124055965666
|
5888 |
-
|
5889 |
type: Clustering
|
5890 |
-
|
5891 |
-
name: MTEB MedrxivClusteringS2S (default)
|
5892 |
-
type: mteb/medrxiv-clustering-s2s
|
5893 |
config: default
|
5894 |
-
|
5895 |
revision: 35191c8c0dca72d8ff3efcd72aa802307d469663
|
|
|
|
|
5896 |
metrics:
|
5897 |
- type: main_score
|
5898 |
value: 22.9594415386389
|
@@ -5900,14 +5894,14 @@ model-index:
|
|
5900 |
value: 22.9594415386389
|
5901 |
- type: v_measure_std
|
5902 |
value: 1.2719806552652395
|
5903 |
-
|
5904 |
-
type:
|
5905 |
-
|
5906 |
-
name: MTEB MindSmallReranking (default)
|
5907 |
-
type: mteb/mind_small
|
5908 |
config: default
|
5909 |
-
|
5910 |
revision: 59042f120c80e8afa9cdbb224f67076cec0fc9a7
|
|
|
|
|
5911 |
metrics:
|
5912 |
- type: main_score
|
5913 |
value: 28.527234738258063
|
@@ -5927,14 +5921,14 @@ model-index:
|
|
5927 |
value: -26.591327847291947
|
5928 |
- type: nAUC_mrr_std
|
5929 |
value: -11.52072949105865
|
5930 |
-
|
5931 |
-
type:
|
5932 |
-
|
5933 |
-
name: MTEB NFCorpus (default)
|
5934 |
-
type: mteb/nfcorpus
|
5935 |
config: default
|
5936 |
-
|
5937 |
revision: ec0fa4fe99da2ff19ca1214b7966684033a58814
|
|
|
|
|
5938 |
metrics:
|
5939 |
- type: main_score
|
5940 |
value: 23.318
|
@@ -6218,14 +6212,14 @@ model-index:
|
|
6218 |
value: 6.773999999999999
|
6219 |
- type: recall_at_5
|
6220 |
value: 8.713
|
6221 |
-
|
6222 |
type: Retrieval
|
6223 |
-
|
6224 |
-
name: MTEB NQ (default)
|
6225 |
-
type: mteb/nq
|
6226 |
config: default
|
6227 |
-
|
6228 |
revision: b774495ed302d8c44a3a7ea25c90dbce03968f31
|
|
|
|
|
6229 |
metrics:
|
6230 |
- type: main_score
|
6231 |
value: 17.682000000000002
|
@@ -6509,14 +6503,14 @@ model-index:
|
|
6509 |
value: 16.338
|
6510 |
- type: recall_at_5
|
6511 |
value: 21.217
|
6512 |
-
|
6513 |
type: Retrieval
|
6514 |
-
|
6515 |
-
name: MTEB QuoraRetrieval (default)
|
6516 |
-
type: mteb/quora
|
6517 |
config: default
|
6518 |
-
|
6519 |
revision: e4e08e0b7dbe3c8700f0daef558ff32256715259
|
|
|
|
|
6520 |
metrics:
|
6521 |
- type: main_score
|
6522 |
value: 74.323
|
@@ -6800,14 +6794,14 @@ model-index:
|
|
6800 |
value: 73.137
|
6801 |
- type: recall_at_5
|
6802 |
value: 78.655
|
6803 |
-
|
6804 |
-
type:
|
6805 |
-
|
6806 |
-
name: MTEB RedditClustering (default)
|
6807 |
-
type: mteb/reddit-clustering
|
6808 |
config: default
|
6809 |
-
|
6810 |
revision: 24640382cdbf8abc73003fb0fa6d111a705499eb
|
|
|
|
|
6811 |
metrics:
|
6812 |
- type: main_score
|
6813 |
value: 28.89014544508522
|
@@ -6815,14 +6809,14 @@ model-index:
|
|
6815 |
value: 28.89014544508522
|
6816 |
- type: v_measure_std
|
6817 |
value: 4.477854992673074
|
6818 |
-
|
6819 |
type: Clustering
|
6820 |
-
|
6821 |
-
name: MTEB RedditClusteringP2P (default)
|
6822 |
-
type: mteb/reddit-clustering-p2p
|
6823 |
config: default
|
6824 |
-
|
6825 |
revision: 385e3cb46b4cfa89021f56c4380204149d0efe33
|
|
|
|
|
6826 |
metrics:
|
6827 |
- type: main_score
|
6828 |
value: 41.588064041506414
|
@@ -6830,14 +6824,14 @@ model-index:
|
|
6830 |
value: 41.588064041506414
|
6831 |
- type: v_measure_std
|
6832 |
value: 12.234957713539355
|
6833 |
-
|
6834 |
-
type:
|
6835 |
-
|
6836 |
-
name: MTEB SCIDOCS (default)
|
6837 |
-
type: mteb/scidocs
|
6838 |
config: default
|
6839 |
-
|
6840 |
revision: f8c2fcf00f625baaa80f62ec5bd9e1fff3b8ae88
|
|
|
|
|
6841 |
metrics:
|
6842 |
- type: main_score
|
6843 |
value: 9.923
|
@@ -7121,14 +7115,14 @@ model-index:
|
|
7121 |
value: 5.295
|
7122 |
- type: recall_at_5
|
7123 |
value: 7.37
|
7124 |
-
|
7125 |
-
type:
|
7126 |
-
|
7127 |
-
name: MTEB SICK-R (default)
|
7128 |
-
type: mteb/sickr-sts
|
7129 |
config: default
|
7130 |
-
|
7131 |
revision: 20a6d6f312dd54037fe07a32d58e5e168867909d
|
|
|
|
|
7132 |
metrics:
|
7133 |
- type: cosine_pearson
|
7134 |
value: 75.0950047498747
|
@@ -7148,14 +7142,14 @@ model-index:
|
|
7148 |
value: 75.0950047498747
|
7149 |
- type: spearman
|
7150 |
value: 66.17240782538595
|
7151 |
-
|
7152 |
type: STS
|
7153 |
-
|
7154 |
-
name: MTEB STS12 (default)
|
7155 |
-
type: mteb/sts12-sts
|
7156 |
config: default
|
7157 |
-
|
7158 |
revision: a0d554a64d88156834ff5ae9920b964011b16384
|
|
|
|
|
7159 |
metrics:
|
7160 |
- type: cosine_pearson
|
7161 |
value: 70.27191745166907
|
@@ -7175,14 +7169,14 @@ model-index:
|
|
7175 |
value: 70.27191745166907
|
7176 |
- type: spearman
|
7177 |
value: 61.89139464648924
|
7178 |
-
|
7179 |
type: STS
|
7180 |
-
|
7181 |
-
name: MTEB STS13 (default)
|
7182 |
-
type: mteb/sts13-sts
|
7183 |
config: default
|
7184 |
-
|
7185 |
revision: 7e90230a92c190f1bf69ae9002b8cea547a64cca
|
|
|
|
|
7186 |
metrics:
|
7187 |
- type: cosine_pearson
|
7188 |
value: 70.19582039979868
|
@@ -7202,14 +7196,14 @@ model-index:
|
|
7202 |
value: 70.19582039979868
|
7203 |
- type: spearman
|
7204 |
value: 71.66792475528088
|
7205 |
-
|
7206 |
type: STS
|
7207 |
-
|
7208 |
-
name: MTEB STS14 (default)
|
7209 |
-
type: mteb/sts14-sts
|
7210 |
config: default
|
7211 |
-
|
7212 |
revision: 6031580fec1f6af667f0bd2da0a551cf4f0b2375
|
|
|
|
|
7213 |
metrics:
|
7214 |
- type: cosine_pearson
|
7215 |
value: 69.52140108419252
|
@@ -7229,14 +7223,14 @@ model-index:
|
|
7229 |
value: 69.52140108419252
|
7230 |
- type: spearman
|
7231 |
value: 67.82634222687376
|
7232 |
-
|
7233 |
type: STS
|
7234 |
-
|
7235 |
-
name: MTEB STS15 (default)
|
7236 |
-
type: mteb/sts15-sts
|
7237 |
config: default
|
7238 |
-
|
7239 |
revision: ae752c7c21bf194d8b67fd573edf7ae58183cbe3
|
|
|
|
|
7240 |
metrics:
|
7241 |
- type: cosine_pearson
|
7242 |
value: 73.66221619412464
|
@@ -7256,14 +7250,14 @@ model-index:
|
|
7256 |
value: 73.66221619412464
|
7257 |
- type: spearman
|
7258 |
value: 75.48765072240437
|
7259 |
-
|
7260 |
type: STS
|
7261 |
-
|
7262 |
-
name: MTEB STS16 (default)
|
7263 |
-
type: mteb/sts16-sts
|
7264 |
config: default
|
7265 |
-
|
7266 |
revision: 4d8694f8f0e0100860b497b999b3dbed754a0513
|
|
|
|
|
7267 |
metrics:
|
7268 |
- type: cosine_pearson
|
7269 |
value: 62.328630460915925
|
@@ -7283,14 +7277,14 @@ model-index:
|
|
7283 |
value: 62.328630460915925
|
7284 |
- type: spearman
|
7285 |
value: 66.48155706668948
|
7286 |
-
|
7287 |
type: STS
|
7288 |
-
|
7289 |
-
name: MTEB STS17 (fr-en)
|
7290 |
-
type: mteb/sts17-crosslingual-sts
|
7291 |
config: fr-en
|
7292 |
-
|
7293 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
|
|
|
|
7294 |
metrics:
|
7295 |
- type: cosine_pearson
|
7296 |
value: 21.344883409729785
|
@@ -7310,14 +7304,14 @@ model-index:
|
|
7310 |
value: 21.344883409729785
|
7311 |
- type: spearman
|
7312 |
value: 19.492480027372526
|
7313 |
-
|
7314 |
type: STS
|
7315 |
-
|
7316 |
-
name: MTEB STS17 (es-en)
|
7317 |
-
type: mteb/sts17-crosslingual-sts
|
7318 |
config: es-en
|
7319 |
-
|
7320 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
|
|
|
|
7321 |
metrics:
|
7322 |
- type: cosine_pearson
|
7323 |
value: 14.966581838953037
|
@@ -7337,14 +7331,14 @@ model-index:
|
|
7337 |
value: 14.966581838953037
|
7338 |
- type: spearman
|
7339 |
value: 13.24509138766898
|
7340 |
-
|
7341 |
type: STS
|
7342 |
-
|
7343 |
-
name: MTEB STS17 (nl-en)
|
7344 |
-
type: mteb/sts17-crosslingual-sts
|
7345 |
config: nl-en
|
7346 |
-
|
7347 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
|
|
|
|
7348 |
metrics:
|
7349 |
- type: cosine_pearson
|
7350 |
value: 18.309414985775234
|
@@ -7364,14 +7358,14 @@ model-index:
|
|
7364 |
value: 18.309414985775234
|
7365 |
- type: spearman
|
7366 |
value: 14.341489363671842
|
7367 |
-
|
7368 |
type: STS
|
7369 |
-
|
7370 |
-
name: MTEB STS17 (en-de)
|
7371 |
-
type: mteb/sts17-crosslingual-sts
|
7372 |
config: en-de
|
7373 |
-
|
7374 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
|
|
|
|
7375 |
metrics:
|
7376 |
- type: cosine_pearson
|
7377 |
value: 21.301586456013037
|
@@ -7391,14 +7385,14 @@ model-index:
|
|
7391 |
value: 21.301586456013037
|
7392 |
- type: spearman
|
7393 |
value: 22.571419522164376
|
7394 |
-
|
7395 |
type: STS
|
7396 |
-
|
7397 |
-
name: MTEB STS17 (it-en)
|
7398 |
-
type: mteb/sts17-crosslingual-sts
|
7399 |
config: it-en
|
7400 |
-
|
7401 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
|
|
|
|
7402 |
metrics:
|
7403 |
- type: cosine_pearson
|
7404 |
value: 16.140292893693204
|
@@ -7418,14 +7412,14 @@ model-index:
|
|
7418 |
value: 16.140292893693204
|
7419 |
- type: spearman
|
7420 |
value: 10.216376215477217
|
7421 |
-
|
7422 |
type: STS
|
7423 |
-
|
7424 |
-
name: MTEB STS17 (en-en)
|
7425 |
-
type: mteb/sts17-crosslingual-sts
|
7426 |
config: en-en
|
7427 |
-
|
7428 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
|
|
|
|
7429 |
metrics:
|
7430 |
- type: cosine_pearson
|
7431 |
value: 78.42242639560023
|
@@ -7445,14 +7439,14 @@ model-index:
|
|
7445 |
value: 78.42242639560023
|
7446 |
- type: spearman
|
7447 |
value: 80.2472005970173
|
7448 |
-
|
7449 |
type: STS
|
7450 |
-
|
7451 |
-
name: MTEB STS17 (en-ar)
|
7452 |
-
type: mteb/sts17-crosslingual-sts
|
7453 |
config: en-ar
|
7454 |
-
|
7455 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
|
|
|
|
7456 |
metrics:
|
7457 |
- type: cosine_pearson
|
7458 |
value: -5.762967943082491
|
@@ -7472,14 +7466,14 @@ model-index:
|
|
7472 |
value: -5.762967943082491
|
7473 |
- type: spearman
|
7474 |
value: -6.184248227377756
|
7475 |
-
|
7476 |
type: STS
|
7477 |
-
|
7478 |
-
name: MTEB STS17 (en-tr)
|
7479 |
-
type: mteb/sts17-crosslingual-sts
|
7480 |
config: en-tr
|
7481 |
-
|
7482 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
|
|
|
|
7483 |
metrics:
|
7484 |
- type: cosine_pearson
|
7485 |
value: -8.666319610669559
|
@@ -7499,14 +7493,14 @@ model-index:
|
|
7499 |
value: -8.666319610669559
|
7500 |
- type: spearman
|
7501 |
value: -10.0877070299522
|
7502 |
-
|
7503 |
type: STS
|
7504 |
-
|
7505 |
-
name: MTEB STS22 (es-en)
|
7506 |
-
type: mteb/sts22-crosslingual-sts
|
7507 |
config: es-en
|
7508 |
-
|
7509 |
revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
|
|
|
|
|
7510 |
metrics:
|
7511 |
- type: cosine_pearson
|
7512 |
value: 16.880423266497427
|
@@ -7526,14 +7520,14 @@ model-index:
|
|
7526 |
value: 16.880423266497427
|
7527 |
- type: spearman
|
7528 |
value: 18.497107178067477
|
7529 |
-
|
7530 |
type: STS
|
7531 |
-
|
7532 |
-
name: MTEB STS22 (en)
|
7533 |
-
type: mteb/sts22-crosslingual-sts
|
7534 |
config: en
|
7535 |
-
|
7536 |
revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
|
|
|
|
|
7537 |
metrics:
|
7538 |
- type: cosine_pearson
|
7539 |
value: 44.98861387948161
|
@@ -7553,14 +7547,14 @@ model-index:
|
|
7553 |
value: 44.98861387948161
|
7554 |
- type: spearman
|
7555 |
value: 59.04270974068145
|
7556 |
-
|
7557 |
type: STS
|
7558 |
-
|
7559 |
-
name: MTEB STS22 (de-en)
|
7560 |
-
type: mteb/sts22-crosslingual-sts
|
7561 |
config: de-en
|
7562 |
-
|
7563 |
revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
|
|
|
|
|
7564 |
metrics:
|
7565 |
- type: cosine_pearson
|
7566 |
value: 26.505168004689462
|
@@ -7580,14 +7574,14 @@ model-index:
|
|
7580 |
value: 26.505168004689462
|
7581 |
- type: spearman
|
7582 |
value: 28.591720613248732
|
7583 |
-
|
7584 |
type: STS
|
7585 |
-
|
7586 |
-
name: MTEB STS22 (zh-en)
|
7587 |
-
type: mteb/sts22-crosslingual-sts
|
7588 |
config: zh-en
|
7589 |
-
|
7590 |
revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
|
|
|
|
|
7591 |
metrics:
|
7592 |
- type: cosine_pearson
|
7593 |
value: 11.552622364692777
|
@@ -7607,14 +7601,14 @@ model-index:
|
|
7607 |
value: 11.552622364692777
|
7608 |
- type: spearman
|
7609 |
value: 10.973019756392695
|
7610 |
-
|
7611 |
type: STS
|
7612 |
-
|
7613 |
-
name: MTEB STS22 (pl-en)
|
7614 |
-
type: mteb/sts22-crosslingual-sts
|
7615 |
config: pl-en
|
7616 |
-
|
7617 |
revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
|
|
|
|
|
7618 |
metrics:
|
7619 |
- type: cosine_pearson
|
7620 |
value: 10.466988163502029
|
@@ -7634,14 +7628,14 @@ model-index:
|
|
7634 |
value: 10.466988163502029
|
7635 |
- type: spearman
|
7636 |
value: -0.21879166839686814
|
7637 |
-
|
7638 |
type: STS
|
7639 |
-
|
7640 |
-
name: MTEB STSBenchmark (default)
|
7641 |
-
type: mteb/stsbenchmark-sts
|
7642 |
config: default
|
7643 |
-
|
7644 |
revision: b0fddb56ed78048fa8b90373c8a3cfc37b684831
|
|
|
|
|
7645 |
metrics:
|
7646 |
- type: cosine_pearson
|
7647 |
value: 66.80057012864974
|
@@ -7661,14 +7655,14 @@ model-index:
|
|
7661 |
value: 66.80057012864974
|
7662 |
- type: spearman
|
7663 |
value: 66.52235871936412
|
7664 |
-
|
7665 |
-
type:
|
7666 |
-
|
7667 |
-
name: MTEB SciDocsRR (default)
|
7668 |
-
type: mteb/scidocs-reranking
|
7669 |
config: default
|
7670 |
-
|
7671 |
revision: d3c5e1fc0b855ab6097bf1cda04dd73947d7caab
|
|
|
|
|
7672 |
metrics:
|
7673 |
- type: main_score
|
7674 |
value: 71.11074203128574
|
@@ -7688,14 +7682,14 @@ model-index:
|
|
7688 |
value: 72.52180244204617
|
7689 |
- type: nAUC_mrr_std
|
7690 |
value: 64.6185657337566
|
7691 |
-
|
7692 |
-
type:
|
7693 |
-
|
7694 |
-
name: MTEB SciFact (default)
|
7695 |
-
type: mteb/scifact
|
7696 |
config: default
|
7697 |
-
|
7698 |
revision: 0228b52cf27578f30900b9e5271d331663a030d7
|
|
|
|
|
7699 |
metrics:
|
7700 |
- type: main_score
|
7701 |
value: 50.663000000000004
|
@@ -7979,14 +7973,14 @@ model-index:
|
|
7979 |
value: 52.983000000000004
|
7980 |
- type: recall_at_5
|
7981 |
value: 58.428000000000004
|
7982 |
-
|
7983 |
-
type:
|
7984 |
-
|
7985 |
-
name: MTEB SprintDuplicateQuestions (default)
|
7986 |
-
type: mteb/sprintduplicatequestions-pairclassification
|
7987 |
config: default
|
7988 |
-
|
7989 |
revision: d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46
|
|
|
|
|
7990 |
metrics:
|
7991 |
- type: cosine_accuracy
|
7992 |
value: 99.6029702970297
|
@@ -8070,14 +8064,14 @@ model-index:
|
|
8070 |
value: 78.80597014925374
|
8071 |
- type: similarity_recall
|
8072 |
value: 79.2
|
8073 |
-
|
8074 |
-
type:
|
8075 |
-
|
8076 |
-
name: MTEB StackExchangeClustering (default)
|
8077 |
-
type: mteb/stackexchange-clustering
|
8078 |
config: default
|
8079 |
-
|
8080 |
revision: 6cbc1f7b2bc0622f2e39d2c77fa502909748c259
|
|
|
|
|
8081 |
metrics:
|
8082 |
- type: main_score
|
8083 |
value: 40.01875953666112
|
@@ -8085,14 +8079,14 @@ model-index:
|
|
8085 |
value: 40.01875953666112
|
8086 |
- type: v_measure_std
|
8087 |
value: 4.519991014119391
|
8088 |
-
|
8089 |
type: Clustering
|
8090 |
-
|
8091 |
-
name: MTEB StackExchangeClusteringP2P (default)
|
8092 |
-
type: mteb/stackexchange-clustering-p2p
|
8093 |
config: default
|
8094 |
-
|
8095 |
revision: 815ca46b2622cec33ccafc3735d572c266efdb44
|
|
|
|
|
8096 |
metrics:
|
8097 |
- type: main_score
|
8098 |
value: 28.81354037080584
|
@@ -8100,14 +8094,14 @@ model-index:
|
|
8100 |
value: 28.81354037080584
|
8101 |
- type: v_measure_std
|
8102 |
value: 1.4144350664362755
|
8103 |
-
|
8104 |
-
type:
|
8105 |
-
|
8106 |
-
name: MTEB StackOverflowDupQuestions (default)
|
8107 |
-
type: mteb/stackoverflowdupquestions-reranking
|
8108 |
config: default
|
8109 |
-
|
8110 |
revision: e185fbe320c72810689fc5848eb6114e1ef5ec69
|
|
|
|
|
8111 |
metrics:
|
8112 |
- type: main_score
|
8113 |
value: 44.09716409649705
|
@@ -8127,14 +8121,14 @@ model-index:
|
|
8127 |
value: 17.207604918830953
|
8128 |
- type: nAUC_mrr_std
|
8129 |
value: 6.664790258906265
|
8130 |
-
|
8131 |
-
type:
|
8132 |
-
|
8133 |
-
name: MTEB SummEval (default)
|
8134 |
-
type: mteb/summeval
|
8135 |
config: default
|
8136 |
-
|
8137 |
revision: cda12ad7615edc362dbf25a00fdd61d3b1eaf93c
|
|
|
|
|
8138 |
metrics:
|
8139 |
- type: cosine_pearson
|
8140 |
value: 29.294245469087553
|
@@ -8150,14 +8144,14 @@ model-index:
|
|
8150 |
value: 29.294245469087553
|
8151 |
- type: spearman
|
8152 |
value: 30.080488918284974
|
8153 |
-
|
8154 |
-
type:
|
8155 |
-
|
8156 |
-
name: MTEB TRECCOVID (default)
|
8157 |
-
type: mteb/trec-covid
|
8158 |
config: default
|
8159 |
-
|
8160 |
revision: bb9466bac8153a0349341eb1b22e06409e78ef4e
|
|
|
|
|
8161 |
metrics:
|
8162 |
- type: main_score
|
8163 |
value: 39.983999999999995
|
@@ -8441,14 +8435,14 @@ model-index:
|
|
8441 |
value: 0.302
|
8442 |
- type: recall_at_5
|
8443 |
value: 0.482
|
8444 |
-
|
8445 |
type: Retrieval
|
8446 |
-
|
8447 |
-
name: MTEB Touche2020 (default)
|
8448 |
-
type: mteb/touche2020
|
8449 |
config: default
|
8450 |
-
|
8451 |
revision: a34f9a33db75fa0cbb21bb5cfc3dae8dc8bec93f
|
|
|
|
|
8452 |
metrics:
|
8453 |
- type: main_score
|
8454 |
value: 13.753000000000002
|
@@ -8732,14 +8726,14 @@ model-index:
|
|
8732 |
value: 3.024
|
8733 |
- type: recall_at_5
|
8734 |
value: 4.776
|
8735 |
-
|
8736 |
-
type:
|
8737 |
-
|
8738 |
-
name: MTEB ToxicConversationsClassification (default)
|
8739 |
-
type: mteb/toxic_conversations_50k
|
8740 |
config: default
|
8741 |
-
|
8742 |
revision: edfaf9da55d3dd50d43143d90c1ac476895ae6de
|
|
|
|
|
8743 |
metrics:
|
8744 |
- type: accuracy
|
8745 |
value: 65.6884765625
|
@@ -8753,14 +8747,14 @@ model-index:
|
|
8753 |
value: 73.34420433686675
|
8754 |
- type: main_score
|
8755 |
value: 65.6884765625
|
8756 |
-
|
8757 |
type: Classification
|
8758 |
-
|
8759 |
-
name: MTEB TweetSentimentExtractionClassification (default)
|
8760 |
-
type: mteb/tweet_sentiment_extraction
|
8761 |
config: default
|
8762 |
-
|
8763 |
revision: d604517c81ca91fe16a244d1248fc021f9ecee7a
|
|
|
|
|
8764 |
metrics:
|
8765 |
- type: accuracy
|
8766 |
value: 49.83305036785513
|
@@ -8770,14 +8764,14 @@ model-index:
|
|
8770 |
value: 49.32130156716104
|
8771 |
- type: main_score
|
8772 |
value: 49.83305036785513
|
8773 |
-
|
8774 |
-
type:
|
8775 |
-
|
8776 |
-
name: MTEB TwentyNewsgroupsClustering (default)
|
8777 |
-
type: mteb/twentynewsgroups-clustering
|
8778 |
config: default
|
8779 |
-
|
8780 |
revision: 6125ec4e24fa026cec8a478383ee943acfbd5449
|
|
|
|
|
8781 |
metrics:
|
8782 |
- type: main_score
|
8783 |
value: 25.27920179659098
|
@@ -8785,14 +8779,14 @@ model-index:
|
|
8785 |
value: 25.27920179659098
|
8786 |
- type: v_measure_std
|
8787 |
value: 2.092324622279832
|
8788 |
-
|
8789 |
-
type:
|
8790 |
-
|
8791 |
-
name: MTEB TwitterSemEval2015 (default)
|
8792 |
-
type: mteb/twittersemeval2015-pairclassification
|
8793 |
config: default
|
8794 |
-
|
8795 |
revision: 70970daeab8776df92f5ea462b6173c0b46fd2d1
|
|
|
|
|
8796 |
metrics:
|
8797 |
- type: cosine_accuracy
|
8798 |
value: 82.19586338439531
|
@@ -8876,14 +8870,14 @@ model-index:
|
|
8876 |
value: 54.3767840152236
|
8877 |
- type: similarity_recall
|
8878 |
value: 60.31662269129288
|
8879 |
-
|
8880 |
type: PairClassification
|
8881 |
-
|
8882 |
-
name: MTEB TwitterURLCorpus (default)
|
8883 |
-
type: mteb/twitterurlcorpus-pairclassification
|
8884 |
config: default
|
8885 |
-
|
8886 |
revision: 8b6510b0b1fa4e4c4f879467980e9be563ec1cdf
|
|
|
|
|
8887 |
metrics:
|
8888 |
- type: cosine_accuracy
|
8889 |
value: 85.86758256684907
|
@@ -8967,6 +8961,13 @@ model-index:
|
|
8967 |
value: 69.39391707784078
|
8968 |
- type: similarity_recall
|
8969 |
value: 72.55158607945796
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8970 |
---
|
8971 |
|
8972 |
# M2V_base_glove_subword Model Card
|
@@ -8987,7 +8988,7 @@ Load this model using the `from_pretrained` method:
|
|
8987 |
from model2vec import StaticModel
|
8988 |
|
8989 |
# Load a pretrained Model2Vec model
|
8990 |
-
model = StaticModel.from_pretrained("M2V_base_glove_subword")
|
8991 |
|
8992 |
# Compute text embeddings
|
8993 |
embeddings = model.encode(["Example sentence"])
|
|
|
4 |
- en
|
5 |
library_name: model2vec
|
6 |
license: mit
|
|
|
|
|
|
|
|
|
7 |
model-index:
|
8 |
- name: M2V_base_glove_subword
|
9 |
results:
|
10 |
+
- dataset:
|
|
|
|
|
|
|
|
|
11 |
config: en-ext
|
12 |
+
name: MTEB AmazonCounterfactualClassification (en-ext)
|
13 |
revision: e8379541af4e31359cca9fbcf4b00f2671dba205
|
14 |
+
split: test
|
15 |
+
type: mteb/amazon_counterfactual
|
16 |
metrics:
|
17 |
- type: accuracy
|
18 |
value: 66.4167916041979
|
|
|
26 |
value: 72.84623161234782
|
27 |
- type: main_score
|
28 |
value: 66.4167916041979
|
29 |
+
task:
|
30 |
type: Classification
|
31 |
+
- dataset:
|
|
|
|
|
32 |
config: en
|
33 |
+
name: MTEB AmazonCounterfactualClassification (en)
|
34 |
revision: e8379541af4e31359cca9fbcf4b00f2671dba205
|
35 |
+
split: test
|
36 |
+
type: mteb/amazon_counterfactual
|
37 |
metrics:
|
38 |
- type: accuracy
|
39 |
value: 68.044776119403
|
|
|
47 |
value: 71.2243167389672
|
48 |
- type: main_score
|
49 |
value: 68.044776119403
|
50 |
+
task:
|
51 |
type: Classification
|
52 |
+
- dataset:
|
|
|
|
|
53 |
config: default
|
54 |
+
name: MTEB AmazonPolarityClassification (default)
|
55 |
revision: e2d317d38cd51312af73b3d32a06d1a08b442046
|
56 |
+
split: test
|
57 |
+
type: mteb/amazon_polarity
|
58 |
metrics:
|
59 |
- type: accuracy
|
60 |
value: 67.21602499999999
|
|
|
68 |
value: 66.68107362746888
|
69 |
- type: main_score
|
70 |
value: 67.21602499999999
|
71 |
+
task:
|
72 |
type: Classification
|
73 |
+
- dataset:
|
|
|
|
|
74 |
config: en
|
75 |
+
name: MTEB AmazonReviewsClassification (en)
|
76 |
revision: 1399c76144fd37290681b995c656ef9b2e06e26d
|
77 |
+
split: test
|
78 |
+
type: mteb/amazon_reviews_multi
|
79 |
metrics:
|
80 |
- type: accuracy
|
81 |
value: 32.384
|
|
|
85 |
value: 32.05276706247388
|
86 |
- type: main_score
|
87 |
value: 32.384
|
88 |
+
task:
|
89 |
+
type: Classification
|
90 |
+
- dataset:
|
|
|
|
|
91 |
config: default
|
92 |
+
name: MTEB ArguAna (default)
|
93 |
revision: c22ab2a51041ffd869aaddef7af8d8215647e41a
|
94 |
+
split: test
|
95 |
+
type: mteb/arguana
|
96 |
metrics:
|
97 |
- type: main_score
|
98 |
value: 29.599999999999998
|
|
|
376 |
value: 28.733999999999998
|
377 |
- type: recall_at_5
|
378 |
value: 36.131
|
379 |
+
task:
|
380 |
+
type: Retrieval
|
381 |
+
- dataset:
|
|
|
|
|
382 |
config: default
|
383 |
+
name: MTEB ArxivClusteringP2P (default)
|
384 |
revision: a122ad7f3f0291bf49cc6f4d32aa80929df69d5d
|
385 |
+
split: test
|
386 |
+
type: mteb/arxiv-clustering-p2p
|
387 |
metrics:
|
388 |
- type: main_score
|
389 |
value: 35.46255145204994
|
|
|
391 |
value: 35.46255145204994
|
392 |
- type: v_measure_std
|
393 |
value: 14.146815377034603
|
394 |
+
task:
|
395 |
type: Clustering
|
396 |
+
- dataset:
|
|
|
|
|
397 |
config: default
|
398 |
+
name: MTEB ArxivClusteringS2S (default)
|
399 |
revision: f910caf1a6075f7329cdf8c1a6135696f37dbd53
|
400 |
+
split: test
|
401 |
+
type: mteb/arxiv-clustering-s2s
|
402 |
metrics:
|
403 |
- type: main_score
|
404 |
value: 26.34189987196252
|
|
|
406 |
value: 26.34189987196252
|
407 |
- type: v_measure_std
|
408 |
value: 14.798697652139317
|
409 |
+
task:
|
410 |
+
type: Clustering
|
411 |
+
- dataset:
|
|
|
|
|
412 |
config: default
|
413 |
+
name: MTEB AskUbuntuDupQuestions (default)
|
414 |
revision: 2000358ca161889fa9c082cb41daa8dcfb161a54
|
415 |
+
split: test
|
416 |
+
type: mteb/askubuntudupquestions-reranking
|
417 |
metrics:
|
418 |
- type: main_score
|
419 |
value: 52.85912447389551
|
|
|
433 |
value: 19.44964881732576
|
434 |
- type: nAUC_mrr_std
|
435 |
value: 9.313450884539453
|
436 |
+
task:
|
437 |
+
type: Reranking
|
438 |
+
- dataset:
|
|
|
|
|
439 |
config: default
|
440 |
+
name: MTEB BIOSSES (default)
|
441 |
revision: d3fb88f8f02e40887cd149695127462bbcf29b4a
|
442 |
+
split: test
|
443 |
+
type: mteb/biosses-sts
|
444 |
metrics:
|
445 |
- type: cosine_pearson
|
446 |
value: 73.38282679412139
|
|
|
460 |
value: 73.38282679412139
|
461 |
- type: spearman
|
462 |
value: 75.59389113278942
|
463 |
+
task:
|
464 |
+
type: STS
|
465 |
+
- dataset:
|
|
|
|
|
466 |
config: default
|
467 |
+
name: MTEB Banking77Classification (default)
|
468 |
revision: 0fd18e25b25c072e09e0d92ab615fda904d66300
|
469 |
+
split: test
|
470 |
+
type: mteb/banking77
|
471 |
metrics:
|
472 |
- type: accuracy
|
473 |
value: 71.38636363636363
|
|
|
477 |
value: 71.55994805461263
|
478 |
- type: main_score
|
479 |
value: 71.38636363636363
|
480 |
+
task:
|
481 |
+
type: Classification
|
482 |
+
- dataset:
|
|
|
|
|
483 |
config: default
|
484 |
+
name: MTEB BiorxivClusteringP2P (default)
|
485 |
revision: 65b79d1d13f80053f67aca9498d9402c2d9f1f40
|
486 |
+
split: test
|
487 |
+
type: mteb/biorxiv-clustering-p2p
|
488 |
metrics:
|
489 |
- type: main_score
|
490 |
value: 31.47309865069476
|
|
|
492 |
value: 31.47309865069476
|
493 |
- type: v_measure_std
|
494 |
value: 0.6360736715097297
|
495 |
+
task:
|
496 |
type: Clustering
|
497 |
+
- dataset:
|
|
|
|
|
498 |
config: default
|
499 |
+
name: MTEB BiorxivClusteringS2S (default)
|
500 |
revision: 258694dd0231531bc1fd9de6ceb52a0853c6d908
|
501 |
+
split: test
|
502 |
+
type: mteb/biorxiv-clustering-s2s
|
503 |
metrics:
|
504 |
- type: main_score
|
505 |
value: 22.58199120148109
|
|
|
507 |
value: 22.58199120148109
|
508 |
- type: v_measure_std
|
509 |
value: 1.1055877138914942
|
510 |
+
task:
|
511 |
+
type: Clustering
|
512 |
+
- dataset:
|
|
|
|
|
513 |
config: default
|
514 |
+
name: MTEB CQADupstackAndroidRetrieval (default)
|
515 |
revision: f46a197baaae43b4f621051089b82a364682dfeb
|
516 |
+
split: test
|
517 |
+
type: mteb/cqadupstack-android
|
518 |
metrics:
|
519 |
- type: main_score
|
520 |
value: 28.518
|
|
|
798 |
value: 26.183
|
799 |
- type: recall_at_5
|
800 |
value: 30.653000000000002
|
801 |
+
task:
|
802 |
type: Retrieval
|
803 |
+
- dataset:
|
|
|
|
|
804 |
config: default
|
805 |
+
name: MTEB CQADupstackEnglishRetrieval (default)
|
806 |
revision: ad9991cb51e31e31e430383c75ffb2885547b5f0
|
807 |
+
split: test
|
808 |
+
type: mteb/cqadupstack-english
|
809 |
metrics:
|
810 |
- type: main_score
|
811 |
value: 24.698999999999998
|
|
|
1089 |
value: 23.247999999999998
|
1090 |
- type: recall_at_5
|
1091 |
value: 25.611
|
1092 |
+
task:
|
1093 |
type: Retrieval
|
1094 |
+
- dataset:
|
|
|
|
|
1095 |
config: default
|
1096 |
+
name: MTEB CQADupstackGamingRetrieval (default)
|
1097 |
revision: 4885aa143210c98657558c04aaf3dc47cfb54340
|
1098 |
+
split: test
|
1099 |
+
type: mteb/cqadupstack-gaming
|
1100 |
metrics:
|
1101 |
- type: main_score
|
1102 |
value: 34.837
|
|
|
1380 |
value: 33.351
|
1381 |
- type: recall_at_5
|
1382 |
value: 39.345
|
1383 |
+
task:
|
1384 |
type: Retrieval
|
1385 |
+
- dataset:
|
|
|
|
|
1386 |
config: default
|
1387 |
+
name: MTEB CQADupstackGisRetrieval (default)
|
1388 |
revision: 5003b3064772da1887988e05400cf3806fe491f2
|
1389 |
+
split: test
|
1390 |
+
type: mteb/cqadupstack-gis
|
1391 |
metrics:
|
1392 |
- type: main_score
|
1393 |
value: 15.889000000000001
|
|
|
1671 |
value: 15.383
|
1672 |
- type: recall_at_5
|
1673 |
value: 18.174
|
1674 |
+
task:
|
1675 |
type: Retrieval
|
1676 |
+
- dataset:
|
|
|
|
|
1677 |
config: default
|
1678 |
+
name: MTEB CQADupstackMathematicaRetrieval (default)
|
1679 |
revision: 90fceea13679c63fe563ded68f3b6f06e50061de
|
1680 |
+
split: test
|
1681 |
+
type: mteb/cqadupstack-mathematica
|
1682 |
metrics:
|
1683 |
- type: main_score
|
1684 |
value: 11.0
|
|
|
1962 |
value: 9.735000000000001
|
1963 |
- type: recall_at_5
|
1964 |
value: 12.595999999999998
|
1965 |
+
task:
|
1966 |
type: Retrieval
|
1967 |
+
- dataset:
|
|
|
|
|
1968 |
config: default
|
1969 |
+
name: MTEB CQADupstackPhysicsRetrieval (default)
|
1970 |
revision: 79531abbd1fb92d06c6d6315a0cbbbf5bb247ea4
|
1971 |
+
split: test
|
1972 |
+
type: mteb/cqadupstack-physics
|
1973 |
metrics:
|
1974 |
- type: main_score
|
1975 |
value: 22.671
|
|
|
2253 |
value: 21.583
|
2254 |
- type: recall_at_5
|
2255 |
value: 24.937
|
2256 |
+
task:
|
2257 |
type: Retrieval
|
2258 |
+
- dataset:
|
|
|
|
|
2259 |
config: default
|
2260 |
+
name: MTEB CQADupstackProgrammersRetrieval (default)
|
2261 |
revision: 6184bc1440d2dbc7612be22b50686b8826d22b32
|
2262 |
+
split: test
|
2263 |
+
type: mteb/cqadupstack-programmers
|
2264 |
metrics:
|
2265 |
- type: main_score
|
2266 |
value: 16.306
|
|
|
2544 |
value: 14.935
|
2545 |
- type: recall_at_5
|
2546 |
value: 17.862000000000002
|
2547 |
+
task:
|
2548 |
type: Retrieval
|
2549 |
+
- dataset:
|
|
|
|
|
2550 |
config: default
|
2551 |
+
name: MTEB CQADupstackRetrieval (default)
|
2552 |
revision: CQADupstackRetrieval_is_a_combined_dataset
|
2553 |
+
split: test
|
2554 |
+
type: CQADupstackRetrieval_is_a_combined_dataset
|
2555 |
metrics:
|
2556 |
- type: main_score
|
2557 |
value: 19.211500000000004
|
2558 |
- type: ndcg_at_10
|
2559 |
value: 19.211500000000004
|
2560 |
+
task:
|
2561 |
type: Retrieval
|
2562 |
+
- dataset:
|
|
|
|
|
2563 |
config: default
|
2564 |
+
name: MTEB CQADupstackStatsRetrieval (default)
|
2565 |
revision: 65ac3a16b8e91f9cee4c9828cc7c335575432a2a
|
2566 |
+
split: test
|
2567 |
+
type: mteb/cqadupstack-stats
|
2568 |
metrics:
|
2569 |
- type: main_score
|
2570 |
value: 13.274
|
|
|
2848 |
value: 12.113999999999999
|
2849 |
- type: recall_at_5
|
2850 |
value: 14.371
|
2851 |
+
task:
|
2852 |
type: Retrieval
|
2853 |
+
- dataset:
|
|
|
|
|
2854 |
config: default
|
2855 |
+
name: MTEB CQADupstackTexRetrieval (default)
|
2856 |
revision: 46989137a86843e03a6195de44b09deda022eec7
|
2857 |
+
split: test
|
2858 |
+
type: mteb/cqadupstack-tex
|
2859 |
metrics:
|
2860 |
- type: main_score
|
2861 |
value: 10.994
|
|
|
3139 |
value: 10.381
|
3140 |
- type: recall_at_5
|
3141 |
value: 12.354999999999999
|
3142 |
+
task:
|
3143 |
type: Retrieval
|
3144 |
+
- dataset:
|
|
|
|
|
3145 |
config: default
|
3146 |
+
name: MTEB CQADupstackUnixRetrieval (default)
|
3147 |
revision: 6c6430d3a6d36f8d2a829195bc5dc94d7e063e53
|
3148 |
+
split: test
|
3149 |
+
type: mteb/cqadupstack-unix
|
3150 |
metrics:
|
3151 |
- type: main_score
|
3152 |
value: 17.415
|
|
|
3430 |
value: 16.458000000000002
|
3431 |
- type: recall_at_5
|
3432 |
value: 18.087
|
3433 |
+
task:
|
3434 |
type: Retrieval
|
3435 |
+
- dataset:
|
|
|
|
|
3436 |
config: default
|
3437 |
+
name: MTEB CQADupstackWebmastersRetrieval (default)
|
3438 |
revision: 160c094312a0e1facb97e55eeddb698c0abe3571
|
3439 |
+
split: test
|
3440 |
+
type: mteb/cqadupstack-webmasters
|
3441 |
metrics:
|
3442 |
- type: main_score
|
3443 |
value: 21.398
|
|
|
3721 |
value: 20.148
|
3722 |
- type: recall_at_5
|
3723 |
value: 23.296
|
3724 |
+
task:
|
3725 |
type: Retrieval
|
3726 |
+
- dataset:
|
|
|
|
|
3727 |
config: default
|
3728 |
+
name: MTEB CQADupstackWordpressRetrieval (default)
|
3729 |
revision: 4ffe81d471b1924886b33c7567bfb200e9eec5c4
|
3730 |
+
split: test
|
3731 |
+
type: mteb/cqadupstack-wordpress
|
3732 |
metrics:
|
3733 |
- type: main_score
|
3734 |
value: 13.536999999999999
|
|
|
4012 |
value: 12.187000000000001
|
4013 |
- type: recall_at_5
|
4014 |
value: 14.868
|
4015 |
+
task:
|
4016 |
type: Retrieval
|
4017 |
+
- dataset:
|
|
|
|
|
4018 |
config: default
|
4019 |
+
name: MTEB ClimateFEVER (default)
|
4020 |
revision: 47f2ac6acb640fc46020b02a5b59fdda04d39380
|
4021 |
+
split: test
|
4022 |
+
type: mteb/climate-fever
|
4023 |
metrics:
|
4024 |
- type: main_score
|
4025 |
value: 14.015
|
|
|
4303 |
value: 9.713
|
4304 |
- type: recall_at_5
|
4305 |
value: 12.415
|
4306 |
+
task:
|
4307 |
type: Retrieval
|
4308 |
+
- dataset:
|
|
|
|
|
4309 |
config: default
|
4310 |
+
name: MTEB DBPedia (default)
|
4311 |
revision: c0f706b76e590d620bd6618b3ca8efdd34e2d659
|
4312 |
+
split: test
|
4313 |
+
type: mteb/dbpedia
|
4314 |
metrics:
|
4315 |
- type: main_score
|
4316 |
value: 19.899
|
|
|
4594 |
value: 6.425
|
4595 |
- type: recall_at_5
|
4596 |
value: 8.818
|
4597 |
+
task:
|
4598 |
+
type: Retrieval
|
4599 |
+
- dataset:
|
|
|
|
|
4600 |
config: default
|
4601 |
+
name: MTEB EmotionClassification (default)
|
4602 |
revision: 4f58c6b202a23cf9a4da393831edf4f9183cad37
|
4603 |
+
split: test
|
4604 |
+
type: mteb/emotion
|
4605 |
metrics:
|
4606 |
- type: accuracy
|
4607 |
value: 38.339999999999996
|
|
|
4611 |
value: 40.51989104726522
|
4612 |
- type: main_score
|
4613 |
value: 38.339999999999996
|
4614 |
+
task:
|
4615 |
+
type: Classification
|
4616 |
+
- dataset:
|
|
|
|
|
4617 |
config: default
|
4618 |
+
name: MTEB FEVER (default)
|
4619 |
revision: bea83ef9e8fb933d90a2f1d5515737465d613e12
|
4620 |
+
split: test
|
4621 |
+
type: mteb/fever
|
4622 |
metrics:
|
4623 |
- type: main_score
|
4624 |
value: 25.006
|
|
|
4902 |
value: 24.082
|
4903 |
- type: recall_at_5
|
4904 |
value: 29.563
|
4905 |
+
task:
|
4906 |
type: Retrieval
|
4907 |
+
- dataset:
|
|
|
|
|
4908 |
config: default
|
4909 |
+
name: MTEB FiQA2018 (default)
|
4910 |
revision: 27a168819829fe9bcd655c2df245fb19452e8e06
|
4911 |
+
split: test
|
4912 |
+
type: mteb/fiqa
|
4913 |
metrics:
|
4914 |
- type: main_score
|
4915 |
value: 11.167
|
|
|
5193 |
value: 8.863999999999999
|
5194 |
- type: recall_at_5
|
5195 |
value: 11.354000000000001
|
5196 |
+
task:
|
5197 |
type: Retrieval
|
5198 |
+
- dataset:
|
|
|
|
|
5199 |
config: default
|
5200 |
+
name: MTEB HotpotQA (default)
|
5201 |
revision: ab518f4d6fcca38d87c25209f94beba119d02014
|
5202 |
+
split: test
|
5203 |
+
type: mteb/hotpotqa
|
5204 |
metrics:
|
5205 |
- type: main_score
|
5206 |
value: 30.837999999999997
|
|
|
5484 |
value: 25.779999999999998
|
5485 |
- type: recall_at_5
|
5486 |
value: 29.250999999999998
|
5487 |
+
task:
|
5488 |
+
type: Retrieval
|
5489 |
+
- dataset:
|
|
|
|
|
5490 |
config: default
|
5491 |
+
name: MTEB ImdbClassification (default)
|
5492 |
revision: 3d86128a09e091d6018b6d26cad27f2739fc2db7
|
5493 |
+
split: test
|
5494 |
+
type: mteb/imdb
|
5495 |
metrics:
|
5496 |
- type: accuracy
|
5497 |
value: 66.5616
|
|
|
5505 |
value: 66.15361405073978
|
5506 |
- type: main_score
|
5507 |
value: 66.5616
|
5508 |
+
task:
|
5509 |
+
type: Classification
|
5510 |
+
- dataset:
|
|
|
|
|
5511 |
config: default
|
5512 |
+
name: MTEB MSMARCO (default)
|
5513 |
revision: c5a29a104738b98a9e76336939199e264163d4a0
|
5514 |
+
split: test
|
5515 |
+
type: mteb/msmarco
|
5516 |
metrics:
|
5517 |
- type: main_score
|
5518 |
value: 28.034
|
|
|
5796 |
value: 2.138
|
5797 |
- type: recall_at_5
|
5798 |
value: 3.3770000000000002
|
5799 |
+
task:
|
5800 |
+
type: Retrieval
|
5801 |
+
- dataset:
|
|
|
|
|
5802 |
config: en
|
5803 |
+
name: MTEB MTOPDomainClassification (en)
|
5804 |
revision: d80d48c1eb48d3562165c59d59d0034df9fff0bf
|
5805 |
+
split: test
|
5806 |
+
type: mteb/mtop_domain
|
5807 |
metrics:
|
5808 |
- type: accuracy
|
5809 |
value: 84.81988144094848
|
|
|
5813 |
value: 84.95181538630469
|
5814 |
- type: main_score
|
5815 |
value: 84.81988144094848
|
5816 |
+
task:
|
5817 |
type: Classification
|
5818 |
+
- dataset:
|
|
|
|
|
5819 |
config: en
|
5820 |
+
name: MTEB MTOPIntentClassification (en)
|
5821 |
revision: ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba
|
5822 |
+
split: test
|
5823 |
+
type: mteb/mtop_intent
|
5824 |
metrics:
|
5825 |
- type: accuracy
|
5826 |
value: 62.41222070223438
|
|
|
5830 |
value: 66.23266420473301
|
5831 |
- type: main_score
|
5832 |
value: 62.41222070223438
|
5833 |
+
task:
|
5834 |
type: Classification
|
5835 |
+
- dataset:
|
|
|
|
|
5836 |
config: en
|
5837 |
+
name: MTEB MassiveIntentClassification (en)
|
5838 |
revision: 4672e20407010da34463acc759c162ca9734bca6
|
5839 |
+
split: test
|
5840 |
+
type: mteb/amazon_massive_intent
|
5841 |
metrics:
|
5842 |
- type: accuracy
|
5843 |
value: 62.50168123739073
|
|
|
5847 |
value: 62.787680759907204
|
5848 |
- type: main_score
|
5849 |
value: 62.50168123739073
|
5850 |
+
task:
|
5851 |
type: Classification
|
5852 |
+
- dataset:
|
|
|
|
|
5853 |
config: en
|
5854 |
+
name: MTEB MassiveScenarioClassification (en)
|
5855 |
revision: fad2c6e8459f9e1c45d9315f4953d921437d70f8
|
5856 |
+
split: test
|
5857 |
+
type: mteb/amazon_massive_scenario
|
5858 |
metrics:
|
5859 |
- type: accuracy
|
5860 |
value: 66.09280430396772
|
|
|
5864 |
value: 66.15203456480924
|
5865 |
- type: main_score
|
5866 |
value: 66.09280430396772
|
5867 |
+
task:
|
5868 |
+
type: Classification
|
5869 |
+
- dataset:
|
5870 |
+
config: default
|
5871 |
name: MTEB MedrxivClusteringP2P (default)
|
5872 |
+
revision: e7a26af6f3ae46b30dde8737f02c07b1505bcc73
|
5873 |
+
split: test
|
5874 |
type: mteb/medrxiv-clustering-p2p
|
|
|
|
|
|
|
5875 |
metrics:
|
5876 |
- type: main_score
|
5877 |
value: 26.932942933622616
|
|
|
5879 |
value: 26.932942933622616
|
5880 |
- type: v_measure_std
|
5881 |
value: 1.593124055965666
|
5882 |
+
task:
|
5883 |
type: Clustering
|
5884 |
+
- dataset:
|
|
|
|
|
5885 |
config: default
|
5886 |
+
name: MTEB MedrxivClusteringS2S (default)
|
5887 |
revision: 35191c8c0dca72d8ff3efcd72aa802307d469663
|
5888 |
+
split: test
|
5889 |
+
type: mteb/medrxiv-clustering-s2s
|
5890 |
metrics:
|
5891 |
- type: main_score
|
5892 |
value: 22.9594415386389
|
|
|
5894 |
value: 22.9594415386389
|
5895 |
- type: v_measure_std
|
5896 |
value: 1.2719806552652395
|
5897 |
+
task:
|
5898 |
+
type: Clustering
|
5899 |
+
- dataset:
|
|
|
|
|
5900 |
config: default
|
5901 |
+
name: MTEB MindSmallReranking (default)
|
5902 |
revision: 59042f120c80e8afa9cdbb224f67076cec0fc9a7
|
5903 |
+
split: test
|
5904 |
+
type: mteb/mind_small
|
5905 |
metrics:
|
5906 |
- type: main_score
|
5907 |
value: 28.527234738258063
|
|
|
5921 |
value: -26.591327847291947
|
5922 |
- type: nAUC_mrr_std
|
5923 |
value: -11.52072949105865
|
5924 |
+
task:
|
5925 |
+
type: Reranking
|
5926 |
+
- dataset:
|
|
|
|
|
5927 |
config: default
|
5928 |
+
name: MTEB NFCorpus (default)
|
5929 |
revision: ec0fa4fe99da2ff19ca1214b7966684033a58814
|
5930 |
+
split: test
|
5931 |
+
type: mteb/nfcorpus
|
5932 |
metrics:
|
5933 |
- type: main_score
|
5934 |
value: 23.318
|
|
|
6212 |
value: 6.773999999999999
|
6213 |
- type: recall_at_5
|
6214 |
value: 8.713
|
6215 |
+
task:
|
6216 |
type: Retrieval
|
6217 |
+
- dataset:
|
|
|
|
|
6218 |
config: default
|
6219 |
+
name: MTEB NQ (default)
|
6220 |
revision: b774495ed302d8c44a3a7ea25c90dbce03968f31
|
6221 |
+
split: test
|
6222 |
+
type: mteb/nq
|
6223 |
metrics:
|
6224 |
- type: main_score
|
6225 |
value: 17.682000000000002
|
|
|
6503 |
value: 16.338
|
6504 |
- type: recall_at_5
|
6505 |
value: 21.217
|
6506 |
+
task:
|
6507 |
type: Retrieval
|
6508 |
+
- dataset:
|
|
|
|
|
6509 |
config: default
|
6510 |
+
name: MTEB QuoraRetrieval (default)
|
6511 |
revision: e4e08e0b7dbe3c8700f0daef558ff32256715259
|
6512 |
+
split: test
|
6513 |
+
type: mteb/quora
|
6514 |
metrics:
|
6515 |
- type: main_score
|
6516 |
value: 74.323
|
|
|
6794 |
value: 73.137
|
6795 |
- type: recall_at_5
|
6796 |
value: 78.655
|
6797 |
+
task:
|
6798 |
+
type: Retrieval
|
6799 |
+
- dataset:
|
|
|
|
|
6800 |
config: default
|
6801 |
+
name: MTEB RedditClustering (default)
|
6802 |
revision: 24640382cdbf8abc73003fb0fa6d111a705499eb
|
6803 |
+
split: test
|
6804 |
+
type: mteb/reddit-clustering
|
6805 |
metrics:
|
6806 |
- type: main_score
|
6807 |
value: 28.89014544508522
|
|
|
6809 |
value: 28.89014544508522
|
6810 |
- type: v_measure_std
|
6811 |
value: 4.477854992673074
|
6812 |
+
task:
|
6813 |
type: Clustering
|
6814 |
+
- dataset:
|
|
|
|
|
6815 |
config: default
|
6816 |
+
name: MTEB RedditClusteringP2P (default)
|
6817 |
revision: 385e3cb46b4cfa89021f56c4380204149d0efe33
|
6818 |
+
split: test
|
6819 |
+
type: mteb/reddit-clustering-p2p
|
6820 |
metrics:
|
6821 |
- type: main_score
|
6822 |
value: 41.588064041506414
|
|
|
6824 |
value: 41.588064041506414
|
6825 |
- type: v_measure_std
|
6826 |
value: 12.234957713539355
|
6827 |
+
task:
|
6828 |
+
type: Clustering
|
6829 |
+
- dataset:
|
|
|
|
|
6830 |
config: default
|
6831 |
+
name: MTEB SCIDOCS (default)
|
6832 |
revision: f8c2fcf00f625baaa80f62ec5bd9e1fff3b8ae88
|
6833 |
+
split: test
|
6834 |
+
type: mteb/scidocs
|
6835 |
metrics:
|
6836 |
- type: main_score
|
6837 |
value: 9.923
|
|
|
7115 |
value: 5.295
|
7116 |
- type: recall_at_5
|
7117 |
value: 7.37
|
7118 |
+
task:
|
7119 |
+
type: Retrieval
|
7120 |
+
- dataset:
|
|
|
|
|
7121 |
config: default
|
7122 |
+
name: MTEB SICK-R (default)
|
7123 |
revision: 20a6d6f312dd54037fe07a32d58e5e168867909d
|
7124 |
+
split: test
|
7125 |
+
type: mteb/sickr-sts
|
7126 |
metrics:
|
7127 |
- type: cosine_pearson
|
7128 |
value: 75.0950047498747
|
|
|
7142 |
value: 75.0950047498747
|
7143 |
- type: spearman
|
7144 |
value: 66.17240782538595
|
7145 |
+
task:
|
7146 |
type: STS
|
7147 |
+
- dataset:
|
|
|
|
|
7148 |
config: default
|
7149 |
+
name: MTEB STS12 (default)
|
7150 |
revision: a0d554a64d88156834ff5ae9920b964011b16384
|
7151 |
+
split: test
|
7152 |
+
type: mteb/sts12-sts
|
7153 |
metrics:
|
7154 |
- type: cosine_pearson
|
7155 |
value: 70.27191745166907
|
|
|
7169 |
value: 70.27191745166907
|
7170 |
- type: spearman
|
7171 |
value: 61.89139464648924
|
7172 |
+
task:
|
7173 |
type: STS
|
7174 |
+
- dataset:
|
|
|
|
|
7175 |
config: default
|
7176 |
+
name: MTEB STS13 (default)
|
7177 |
revision: 7e90230a92c190f1bf69ae9002b8cea547a64cca
|
7178 |
+
split: test
|
7179 |
+
type: mteb/sts13-sts
|
7180 |
metrics:
|
7181 |
- type: cosine_pearson
|
7182 |
value: 70.19582039979868
|
|
|
7196 |
value: 70.19582039979868
|
7197 |
- type: spearman
|
7198 |
value: 71.66792475528088
|
7199 |
+
task:
|
7200 |
type: STS
|
7201 |
+
- dataset:
|
|
|
|
|
7202 |
config: default
|
7203 |
+
name: MTEB STS14 (default)
|
7204 |
revision: 6031580fec1f6af667f0bd2da0a551cf4f0b2375
|
7205 |
+
split: test
|
7206 |
+
type: mteb/sts14-sts
|
7207 |
metrics:
|
7208 |
- type: cosine_pearson
|
7209 |
value: 69.52140108419252
|
|
|
7223 |
value: 69.52140108419252
|
7224 |
- type: spearman
|
7225 |
value: 67.82634222687376
|
7226 |
+
task:
|
7227 |
type: STS
|
7228 |
+
- dataset:
|
|
|
|
|
7229 |
config: default
|
7230 |
+
name: MTEB STS15 (default)
|
7231 |
revision: ae752c7c21bf194d8b67fd573edf7ae58183cbe3
|
7232 |
+
split: test
|
7233 |
+
type: mteb/sts15-sts
|
7234 |
metrics:
|
7235 |
- type: cosine_pearson
|
7236 |
value: 73.66221619412464
|
|
|
7250 |
value: 73.66221619412464
|
7251 |
- type: spearman
|
7252 |
value: 75.48765072240437
|
7253 |
+
task:
|
7254 |
type: STS
|
7255 |
+
- dataset:
|
|
|
|
|
7256 |
config: default
|
7257 |
+
name: MTEB STS16 (default)
|
7258 |
revision: 4d8694f8f0e0100860b497b999b3dbed754a0513
|
7259 |
+
split: test
|
7260 |
+
type: mteb/sts16-sts
|
7261 |
metrics:
|
7262 |
- type: cosine_pearson
|
7263 |
value: 62.328630460915925
|
|
|
7277 |
value: 62.328630460915925
|
7278 |
- type: spearman
|
7279 |
value: 66.48155706668948
|
7280 |
+
task:
|
7281 |
type: STS
|
7282 |
+
- dataset:
|
|
|
|
|
7283 |
config: fr-en
|
7284 |
+
name: MTEB STS17 (fr-en)
|
7285 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
7286 |
+
split: test
|
7287 |
+
type: mteb/sts17-crosslingual-sts
|
7288 |
metrics:
|
7289 |
- type: cosine_pearson
|
7290 |
value: 21.344883409729785
|
|
|
7304 |
value: 21.344883409729785
|
7305 |
- type: spearman
|
7306 |
value: 19.492480027372526
|
7307 |
+
task:
|
7308 |
type: STS
|
7309 |
+
- dataset:
|
|
|
|
|
7310 |
config: es-en
|
7311 |
+
name: MTEB STS17 (es-en)
|
7312 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
7313 |
+
split: test
|
7314 |
+
type: mteb/sts17-crosslingual-sts
|
7315 |
metrics:
|
7316 |
- type: cosine_pearson
|
7317 |
value: 14.966581838953037
|
|
|
7331 |
value: 14.966581838953037
|
7332 |
- type: spearman
|
7333 |
value: 13.24509138766898
|
7334 |
+
task:
|
7335 |
type: STS
|
7336 |
+
- dataset:
|
|
|
|
|
7337 |
config: nl-en
|
7338 |
+
name: MTEB STS17 (nl-en)
|
7339 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
7340 |
+
split: test
|
7341 |
+
type: mteb/sts17-crosslingual-sts
|
7342 |
metrics:
|
7343 |
- type: cosine_pearson
|
7344 |
value: 18.309414985775234
|
|
|
7358 |
value: 18.309414985775234
|
7359 |
- type: spearman
|
7360 |
value: 14.341489363671842
|
7361 |
+
task:
|
7362 |
type: STS
|
7363 |
+
- dataset:
|
|
|
|
|
7364 |
config: en-de
|
7365 |
+
name: MTEB STS17 (en-de)
|
7366 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
7367 |
+
split: test
|
7368 |
+
type: mteb/sts17-crosslingual-sts
|
7369 |
metrics:
|
7370 |
- type: cosine_pearson
|
7371 |
value: 21.301586456013037
|
|
|
7385 |
value: 21.301586456013037
|
7386 |
- type: spearman
|
7387 |
value: 22.571419522164376
|
7388 |
+
task:
|
7389 |
type: STS
|
7390 |
+
- dataset:
|
|
|
|
|
7391 |
config: it-en
|
7392 |
+
name: MTEB STS17 (it-en)
|
7393 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
7394 |
+
split: test
|
7395 |
+
type: mteb/sts17-crosslingual-sts
|
7396 |
metrics:
|
7397 |
- type: cosine_pearson
|
7398 |
value: 16.140292893693204
|
|
|
7412 |
value: 16.140292893693204
|
7413 |
- type: spearman
|
7414 |
value: 10.216376215477217
|
7415 |
+
task:
|
7416 |
type: STS
|
7417 |
+
- dataset:
|
|
|
|
|
7418 |
config: en-en
|
7419 |
+
name: MTEB STS17 (en-en)
|
7420 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
7421 |
+
split: test
|
7422 |
+
type: mteb/sts17-crosslingual-sts
|
7423 |
metrics:
|
7424 |
- type: cosine_pearson
|
7425 |
value: 78.42242639560023
|
|
|
7439 |
value: 78.42242639560023
|
7440 |
- type: spearman
|
7441 |
value: 80.2472005970173
|
7442 |
+
task:
|
7443 |
type: STS
|
7444 |
+
- dataset:
|
|
|
|
|
7445 |
config: en-ar
|
7446 |
+
name: MTEB STS17 (en-ar)
|
7447 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
7448 |
+
split: test
|
7449 |
+
type: mteb/sts17-crosslingual-sts
|
7450 |
metrics:
|
7451 |
- type: cosine_pearson
|
7452 |
value: -5.762967943082491
|
|
|
7466 |
value: -5.762967943082491
|
7467 |
- type: spearman
|
7468 |
value: -6.184248227377756
|
7469 |
+
task:
|
7470 |
type: STS
|
7471 |
+
- dataset:
|
|
|
|
|
7472 |
config: en-tr
|
7473 |
+
name: MTEB STS17 (en-tr)
|
7474 |
revision: faeb762787bd10488a50c8b5be4a3b82e411949c
|
7475 |
+
split: test
|
7476 |
+
type: mteb/sts17-crosslingual-sts
|
7477 |
metrics:
|
7478 |
- type: cosine_pearson
|
7479 |
value: -8.666319610669559
|
|
|
7493 |
value: -8.666319610669559
|
7494 |
- type: spearman
|
7495 |
value: -10.0877070299522
|
7496 |
+
task:
|
7497 |
type: STS
|
7498 |
+
- dataset:
|
|
|
|
|
7499 |
config: es-en
|
7500 |
+
name: MTEB STS22 (es-en)
|
7501 |
revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
|
7502 |
+
split: test
|
7503 |
+
type: mteb/sts22-crosslingual-sts
|
7504 |
metrics:
|
7505 |
- type: cosine_pearson
|
7506 |
value: 16.880423266497427
|
|
|
7520 |
value: 16.880423266497427
|
7521 |
- type: spearman
|
7522 |
value: 18.497107178067477
|
7523 |
+
task:
|
7524 |
type: STS
|
7525 |
+
- dataset:
|
|
|
|
|
7526 |
config: en
|
7527 |
+
name: MTEB STS22 (en)
|
7528 |
revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
|
7529 |
+
split: test
|
7530 |
+
type: mteb/sts22-crosslingual-sts
|
7531 |
metrics:
|
7532 |
- type: cosine_pearson
|
7533 |
value: 44.98861387948161
|
|
|
7547 |
value: 44.98861387948161
|
7548 |
- type: spearman
|
7549 |
value: 59.04270974068145
|
7550 |
+
task:
|
7551 |
type: STS
|
7552 |
+
- dataset:
|
|
|
|
|
7553 |
config: de-en
|
7554 |
+
name: MTEB STS22 (de-en)
|
7555 |
revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
|
7556 |
+
split: test
|
7557 |
+
type: mteb/sts22-crosslingual-sts
|
7558 |
metrics:
|
7559 |
- type: cosine_pearson
|
7560 |
value: 26.505168004689462
|
|
|
7574 |
value: 26.505168004689462
|
7575 |
- type: spearman
|
7576 |
value: 28.591720613248732
|
7577 |
+
task:
|
7578 |
type: STS
|
7579 |
+
- dataset:
|
|
|
|
|
7580 |
config: zh-en
|
7581 |
+
name: MTEB STS22 (zh-en)
|
7582 |
revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
|
7583 |
+
split: test
|
7584 |
+
type: mteb/sts22-crosslingual-sts
|
7585 |
metrics:
|
7586 |
- type: cosine_pearson
|
7587 |
value: 11.552622364692777
|
|
|
7601 |
value: 11.552622364692777
|
7602 |
- type: spearman
|
7603 |
value: 10.973019756392695
|
7604 |
+
task:
|
7605 |
type: STS
|
7606 |
+
- dataset:
|
|
|
|
|
7607 |
config: pl-en
|
7608 |
+
name: MTEB STS22 (pl-en)
|
7609 |
revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
|
7610 |
+
split: test
|
7611 |
+
type: mteb/sts22-crosslingual-sts
|
7612 |
metrics:
|
7613 |
- type: cosine_pearson
|
7614 |
value: 10.466988163502029
|
|
|
7628 |
value: 10.466988163502029
|
7629 |
- type: spearman
|
7630 |
value: -0.21879166839686814
|
7631 |
+
task:
|
7632 |
type: STS
|
7633 |
+
- dataset:
|
|
|
|
|
7634 |
config: default
|
7635 |
+
name: MTEB STSBenchmark (default)
|
7636 |
revision: b0fddb56ed78048fa8b90373c8a3cfc37b684831
|
7637 |
+
split: test
|
7638 |
+
type: mteb/stsbenchmark-sts
|
7639 |
metrics:
|
7640 |
- type: cosine_pearson
|
7641 |
value: 66.80057012864974
|
|
|
7655 |
value: 66.80057012864974
|
7656 |
- type: spearman
|
7657 |
value: 66.52235871936412
|
7658 |
+
task:
|
7659 |
+
type: STS
|
7660 |
+
- dataset:
|
|
|
|
|
7661 |
config: default
|
7662 |
+
name: MTEB SciDocsRR (default)
|
7663 |
revision: d3c5e1fc0b855ab6097bf1cda04dd73947d7caab
|
7664 |
+
split: test
|
7665 |
+
type: mteb/scidocs-reranking
|
7666 |
metrics:
|
7667 |
- type: main_score
|
7668 |
value: 71.11074203128574
|
|
|
7682 |
value: 72.52180244204617
|
7683 |
- type: nAUC_mrr_std
|
7684 |
value: 64.6185657337566
|
7685 |
+
task:
|
7686 |
+
type: Reranking
|
7687 |
+
- dataset:
|
|
|
|
|
7688 |
config: default
|
7689 |
+
name: MTEB SciFact (default)
|
7690 |
revision: 0228b52cf27578f30900b9e5271d331663a030d7
|
7691 |
+
split: test
|
7692 |
+
type: mteb/scifact
|
7693 |
metrics:
|
7694 |
- type: main_score
|
7695 |
value: 50.663000000000004
|
|
|
7973 |
value: 52.983000000000004
|
7974 |
- type: recall_at_5
|
7975 |
value: 58.428000000000004
|
7976 |
+
task:
|
7977 |
+
type: Retrieval
|
7978 |
+
- dataset:
|
|
|
|
|
7979 |
config: default
|
7980 |
+
name: MTEB SprintDuplicateQuestions (default)
|
7981 |
revision: d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46
|
7982 |
+
split: test
|
7983 |
+
type: mteb/sprintduplicatequestions-pairclassification
|
7984 |
metrics:
|
7985 |
- type: cosine_accuracy
|
7986 |
value: 99.6029702970297
|
|
|
8064 |
value: 78.80597014925374
|
8065 |
- type: similarity_recall
|
8066 |
value: 79.2
|
8067 |
+
task:
|
8068 |
+
type: PairClassification
|
8069 |
+
- dataset:
|
|
|
|
|
8070 |
config: default
|
8071 |
+
name: MTEB StackExchangeClustering (default)
|
8072 |
revision: 6cbc1f7b2bc0622f2e39d2c77fa502909748c259
|
8073 |
+
split: test
|
8074 |
+
type: mteb/stackexchange-clustering
|
8075 |
metrics:
|
8076 |
- type: main_score
|
8077 |
value: 40.01875953666112
|
|
|
8079 |
value: 40.01875953666112
|
8080 |
- type: v_measure_std
|
8081 |
value: 4.519991014119391
|
8082 |
+
task:
|
8083 |
type: Clustering
|
8084 |
+
- dataset:
|
|
|
|
|
8085 |
config: default
|
8086 |
+
name: MTEB StackExchangeClusteringP2P (default)
|
8087 |
revision: 815ca46b2622cec33ccafc3735d572c266efdb44
|
8088 |
+
split: test
|
8089 |
+
type: mteb/stackexchange-clustering-p2p
|
8090 |
metrics:
|
8091 |
- type: main_score
|
8092 |
value: 28.81354037080584
|
|
|
8094 |
value: 28.81354037080584
|
8095 |
- type: v_measure_std
|
8096 |
value: 1.4144350664362755
|
8097 |
+
task:
|
8098 |
+
type: Clustering
|
8099 |
+
- dataset:
|
|
|
|
|
8100 |
config: default
|
8101 |
+
name: MTEB StackOverflowDupQuestions (default)
|
8102 |
revision: e185fbe320c72810689fc5848eb6114e1ef5ec69
|
8103 |
+
split: test
|
8104 |
+
type: mteb/stackoverflowdupquestions-reranking
|
8105 |
metrics:
|
8106 |
- type: main_score
|
8107 |
value: 44.09716409649705
|
|
|
8121 |
value: 17.207604918830953
|
8122 |
- type: nAUC_mrr_std
|
8123 |
value: 6.664790258906265
|
8124 |
+
task:
|
8125 |
+
type: Reranking
|
8126 |
+
- dataset:
|
|
|
|
|
8127 |
config: default
|
8128 |
+
name: MTEB SummEval (default)
|
8129 |
revision: cda12ad7615edc362dbf25a00fdd61d3b1eaf93c
|
8130 |
+
split: test
|
8131 |
+
type: mteb/summeval
|
8132 |
metrics:
|
8133 |
- type: cosine_pearson
|
8134 |
value: 29.294245469087553
|
|
|
8144 |
value: 29.294245469087553
|
8145 |
- type: spearman
|
8146 |
value: 30.080488918284974
|
8147 |
+
task:
|
8148 |
+
type: Summarization
|
8149 |
+
- dataset:
|
|
|
|
|
8150 |
config: default
|
8151 |
+
name: MTEB TRECCOVID (default)
|
8152 |
revision: bb9466bac8153a0349341eb1b22e06409e78ef4e
|
8153 |
+
split: test
|
8154 |
+
type: mteb/trec-covid
|
8155 |
metrics:
|
8156 |
- type: main_score
|
8157 |
value: 39.983999999999995
|
|
|
8435 |
value: 0.302
|
8436 |
- type: recall_at_5
|
8437 |
value: 0.482
|
8438 |
+
task:
|
8439 |
type: Retrieval
|
8440 |
+
- dataset:
|
|
|
|
|
8441 |
config: default
|
8442 |
+
name: MTEB Touche2020 (default)
|
8443 |
revision: a34f9a33db75fa0cbb21bb5cfc3dae8dc8bec93f
|
8444 |
+
split: test
|
8445 |
+
type: mteb/touche2020
|
8446 |
metrics:
|
8447 |
- type: main_score
|
8448 |
value: 13.753000000000002
|
|
|
8726 |
value: 3.024
|
8727 |
- type: recall_at_5
|
8728 |
value: 4.776
|
8729 |
+
task:
|
8730 |
+
type: Retrieval
|
8731 |
+
- dataset:
|
|
|
|
|
8732 |
config: default
|
8733 |
+
name: MTEB ToxicConversationsClassification (default)
|
8734 |
revision: edfaf9da55d3dd50d43143d90c1ac476895ae6de
|
8735 |
+
split: test
|
8736 |
+
type: mteb/toxic_conversations_50k
|
8737 |
metrics:
|
8738 |
- type: accuracy
|
8739 |
value: 65.6884765625
|
|
|
8747 |
value: 73.34420433686675
|
8748 |
- type: main_score
|
8749 |
value: 65.6884765625
|
8750 |
+
task:
|
8751 |
type: Classification
|
8752 |
+
- dataset:
|
|
|
|
|
8753 |
config: default
|
8754 |
+
name: MTEB TweetSentimentExtractionClassification (default)
|
8755 |
revision: d604517c81ca91fe16a244d1248fc021f9ecee7a
|
8756 |
+
split: test
|
8757 |
+
type: mteb/tweet_sentiment_extraction
|
8758 |
metrics:
|
8759 |
- type: accuracy
|
8760 |
value: 49.83305036785513
|
|
|
8764 |
value: 49.32130156716104
|
8765 |
- type: main_score
|
8766 |
value: 49.83305036785513
|
8767 |
+
task:
|
8768 |
+
type: Classification
|
8769 |
+
- dataset:
|
|
|
|
|
8770 |
config: default
|
8771 |
+
name: MTEB TwentyNewsgroupsClustering (default)
|
8772 |
revision: 6125ec4e24fa026cec8a478383ee943acfbd5449
|
8773 |
+
split: test
|
8774 |
+
type: mteb/twentynewsgroups-clustering
|
8775 |
metrics:
|
8776 |
- type: main_score
|
8777 |
value: 25.27920179659098
|
|
|
8779 |
value: 25.27920179659098
|
8780 |
- type: v_measure_std
|
8781 |
value: 2.092324622279832
|
8782 |
+
task:
|
8783 |
+
type: Clustering
|
8784 |
+
- dataset:
|
|
|
|
|
8785 |
config: default
|
8786 |
+
name: MTEB TwitterSemEval2015 (default)
|
8787 |
revision: 70970daeab8776df92f5ea462b6173c0b46fd2d1
|
8788 |
+
split: test
|
8789 |
+
type: mteb/twittersemeval2015-pairclassification
|
8790 |
metrics:
|
8791 |
- type: cosine_accuracy
|
8792 |
value: 82.19586338439531
|
|
|
8870 |
value: 54.3767840152236
|
8871 |
- type: similarity_recall
|
8872 |
value: 60.31662269129288
|
8873 |
+
task:
|
8874 |
type: PairClassification
|
8875 |
+
- dataset:
|
|
|
|
|
8876 |
config: default
|
8877 |
+
name: MTEB TwitterURLCorpus (default)
|
8878 |
revision: 8b6510b0b1fa4e4c4f879467980e9be563ec1cdf
|
8879 |
+
split: test
|
8880 |
+
type: mteb/twitterurlcorpus-pairclassification
|
8881 |
metrics:
|
8882 |
- type: cosine_accuracy
|
8883 |
value: 85.86758256684907
|
|
|
8961 |
value: 69.39391707784078
|
8962 |
- type: similarity_recall
|
8963 |
value: 72.55158607945796
|
8964 |
+
task:
|
8965 |
+
type: PairClassification
|
8966 |
+
model_name: minishlab/M2V_base_glove_subword
|
8967 |
+
tags:
|
8968 |
+
- embeddings
|
8969 |
+
- static-embeddings
|
8970 |
+
- mteb
|
8971 |
---
|
8972 |
|
8973 |
# M2V_base_glove_subword Model Card
|
|
|
8988 |
from model2vec import StaticModel
|
8989 |
|
8990 |
# Load a pretrained Model2Vec model
|
8991 |
+
model = StaticModel.from_pretrained("minishlab/M2V_base_glove_subword")
|
8992 |
|
8993 |
# Compute text embeddings
|
8994 |
embeddings = model.encode(["Example sentence"])
|