OwenElliott
commited on
Commit
•
0136308
1
Parent(s):
cda2920
Upload README.md
Browse files
README.md
CHANGED
@@ -2,6 +2,114 @@
|
|
2 |
model-index:
|
3 |
- name: marqo-chimera-arctic-bge-m
|
4 |
results:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
- dataset:
|
6 |
config: default
|
7 |
name: MTEB ArguAna (default)
|
|
|
2 |
model-index:
|
3 |
- name: marqo-chimera-arctic-bge-m
|
4 |
results:
|
5 |
+
- dataset:
|
6 |
+
config: default
|
7 |
+
name: MTEB AskUbuntuDupQuestions (default)
|
8 |
+
revision: 2000358ca161889fa9c082cb41daa8dcfb161a54
|
9 |
+
split: test
|
10 |
+
type: mteb/askubuntudupquestions-reranking
|
11 |
+
metrics:
|
12 |
+
- type: main_score
|
13 |
+
value: 61.52568874736682
|
14 |
+
- type: map
|
15 |
+
value: 61.52568874736682
|
16 |
+
- type: mrr
|
17 |
+
value: 73.62001494965483
|
18 |
+
- type: nAUC_map_diff1
|
19 |
+
value: 11.157737453725963
|
20 |
+
- type: nAUC_map_max
|
21 |
+
value: 30.29323031997606
|
22 |
+
- type: nAUC_map_std
|
23 |
+
value: 21.95143338261497
|
24 |
+
- type: nAUC_mrr_diff1
|
25 |
+
value: 26.35768059823056
|
26 |
+
- type: nAUC_mrr_max
|
27 |
+
value: 44.19057408461639
|
28 |
+
- type: nAUC_mrr_std
|
29 |
+
value: 32.32130341448545
|
30 |
+
task:
|
31 |
+
type: Reranking
|
32 |
+
- dataset:
|
33 |
+
config: default
|
34 |
+
name: MTEB MindSmallReranking (default)
|
35 |
+
revision: 59042f120c80e8afa9cdbb224f67076cec0fc9a7
|
36 |
+
split: test
|
37 |
+
type: mteb/mind_small
|
38 |
+
metrics:
|
39 |
+
- type: main_score
|
40 |
+
value: 31.052388483623112
|
41 |
+
- type: map
|
42 |
+
value: 31.052388483623112
|
43 |
+
- type: mrr
|
44 |
+
value: 31.97303748813518
|
45 |
+
- type: nAUC_map_diff1
|
46 |
+
value: 13.880201359939514
|
47 |
+
- type: nAUC_map_max
|
48 |
+
value: -23.904919601765588
|
49 |
+
- type: nAUC_map_std
|
50 |
+
value: -3.138204392900581
|
51 |
+
- type: nAUC_mrr_diff1
|
52 |
+
value: 13.148603185908703
|
53 |
+
- type: nAUC_mrr_max
|
54 |
+
value: -18.517207734097
|
55 |
+
- type: nAUC_mrr_std
|
56 |
+
value: -1.3667482921703122
|
57 |
+
task:
|
58 |
+
type: Reranking
|
59 |
+
- dataset:
|
60 |
+
config: default
|
61 |
+
name: MTEB SciDocsRR (default)
|
62 |
+
revision: d3c5e1fc0b855ab6097bf1cda04dd73947d7caab
|
63 |
+
split: test
|
64 |
+
type: mteb/scidocs-reranking
|
65 |
+
metrics:
|
66 |
+
- type: main_score
|
67 |
+
value: 87.00318874885207
|
68 |
+
- type: map
|
69 |
+
value: 87.00318874885207
|
70 |
+
- type: mrr
|
71 |
+
value: 96.21491616589655
|
72 |
+
- type: nAUC_map_diff1
|
73 |
+
value: -2.7942161554105844
|
74 |
+
- type: nAUC_map_max
|
75 |
+
value: 49.58195147676295
|
76 |
+
- type: nAUC_map_std
|
77 |
+
value: 64.58288470555894
|
78 |
+
- type: nAUC_mrr_diff1
|
79 |
+
value: 35.81189666282723
|
80 |
+
- type: nAUC_mrr_max
|
81 |
+
value: 82.72754486008768
|
82 |
+
- type: nAUC_mrr_std
|
83 |
+
value: 78.20855667229367
|
84 |
+
task:
|
85 |
+
type: Reranking
|
86 |
+
- dataset:
|
87 |
+
config: default
|
88 |
+
name: MTEB StackOverflowDupQuestions (default)
|
89 |
+
revision: e185fbe320c72810689fc5848eb6114e1ef5ec69
|
90 |
+
split: test
|
91 |
+
type: mteb/stackoverflowdupquestions-reranking
|
92 |
+
metrics:
|
93 |
+
- type: main_score
|
94 |
+
value: 52.94628709332918
|
95 |
+
- type: map
|
96 |
+
value: 52.94628709332918
|
97 |
+
- type: mrr
|
98 |
+
value: 53.8088474025974
|
99 |
+
- type: nAUC_map_diff1
|
100 |
+
value: 37.831491121972284
|
101 |
+
- type: nAUC_map_max
|
102 |
+
value: 9.954672687599786
|
103 |
+
- type: nAUC_map_std
|
104 |
+
value: 7.151680447940397
|
105 |
+
- type: nAUC_mrr_diff1
|
106 |
+
value: 37.94530402756348
|
107 |
+
- type: nAUC_mrr_max
|
108 |
+
value: 11.10510095877716
|
109 |
+
- type: nAUC_mrr_std
|
110 |
+
value: 7.335464870025126
|
111 |
+
task:
|
112 |
+
type: Reranking
|
113 |
- dataset:
|
114 |
config: default
|
115 |
name: MTEB ArguAna (default)
|