arnocandel
commited on
Commit
•
6b732b5
1
Parent(s):
d1aea17
commit files to HF hub
Browse files
README.md
CHANGED
@@ -11,8 +11,7 @@ tags:
|
|
11 |
- large language model
|
12 |
- open-source
|
13 |
datasets:
|
14 |
-
- h2oai/
|
15 |
-
- h2oai/openassistant_oasst1_h2ogpt
|
16 |
---
|
17 |
# h2oGPT Model Card
|
18 |
## Summary
|
@@ -20,9 +19,9 @@ datasets:
|
|
20 |
H2O.ai's `h2ogpt-oasst1-512-12b` is a 12 billion parameter instruction-following large language model licensed for commercial use.
|
21 |
|
22 |
- Base model: [EleutherAI/pythia-12b](https://huggingface.co/EleutherAI/pythia-12b)
|
23 |
-
- Fine-tuning dataset: [h2oai/
|
24 |
- Data-prep and fine-tuning code: [H2O.ai GitHub](https://github.com/h2oai/h2ogpt)
|
25 |
-
- Training logs: [zip](https://huggingface.co/h2oai/h2ogpt-oasst1-512-12b/blob/main/pythia-12b
|
26 |
|
27 |
## Chatbot
|
28 |
|
@@ -107,12 +106,6 @@ GPTNeoXConfig {
|
|
107 |
"pt": "AutoModelForCausalLM"
|
108 |
}
|
109 |
},
|
110 |
-
"custom_pipelines": {
|
111 |
-
"text-generation": {
|
112 |
-
"impl": "h2oai_pipeline.H2OTextGenerationPipeline",
|
113 |
-
"pt": "AutoModelForCausalLM"
|
114 |
-
}
|
115 |
-
},
|
116 |
"eos_token_id": 0,
|
117 |
"hidden_act": "gelu",
|
118 |
"hidden_size": 5120,
|
@@ -140,22 +133,22 @@ GPTNeoXConfig {
|
|
140 |
Model validation results using [EleutherAI lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness).
|
141 |
|
142 |
|
143 |
-
[eval source code](https://github.com/h2oai/h2ogpt/issues/
|
144 |
|
145 |
| Task |Version| Metric |Value | |Stderr|
|
146 |
|-------------|------:|--------|-----:|---|-----:|
|
|
|
|
|
147 |
|arc_easy | 0|acc |0.6932|± |0.0095|
|
148 |
| | |acc_norm|0.6225|± |0.0099|
|
149 |
-
|
|
150 |
-
| | |acc_norm|0.3740|± |0.0217|
|
151 |
-
|winogrande | 0|acc |0.6369|± |0.0135|
|
152 |
|hellaswag | 0|acc |0.5140|± |0.0050|
|
153 |
| | |acc_norm|0.6803|± |0.0047|
|
|
|
|
|
154 |
|piqa | 0|acc |0.7682|± |0.0098|
|
155 |
| | |acc_norm|0.7661|± |0.0099|
|
156 |
-
|
|
157 |
-
|arc_challenge| 0|acc |0.3157|± |0.0136|
|
158 |
-
| | |acc_norm|0.3507|± |0.0139|
|
159 |
|
160 |
|
161 |
## Disclaimer
|
|
|
11 |
- large language model
|
12 |
- open-source
|
13 |
datasets:
|
14 |
+
- h2oai/openassistant_oasst1_h2ogpt_graded
|
|
|
15 |
---
|
16 |
# h2oGPT Model Card
|
17 |
## Summary
|
|
|
19 |
H2O.ai's `h2ogpt-oasst1-512-12b` is a 12 billion parameter instruction-following large language model licensed for commercial use.
|
20 |
|
21 |
- Base model: [EleutherAI/pythia-12b](https://huggingface.co/EleutherAI/pythia-12b)
|
22 |
+
- Fine-tuning dataset: [h2oai/openassistant_oasst1_h2ogpt_graded](https://huggingface.co/datasets/h2oai/openassistant_oasst1_h2ogpt_graded)
|
23 |
- Data-prep and fine-tuning code: [H2O.ai GitHub](https://github.com/h2oai/h2ogpt)
|
24 |
+
- Training logs: [zip](https://huggingface.co/h2oai/h2ogpt-oasst1-512-12b/blob/main/pythia-12b-deduped.h2oaiopenassistant_oasst1_h2ogpt_graded.3_epochs.2ccf687ea3f3f3775a501838e81c1a0066430455.4.zip)
|
25 |
|
26 |
## Chatbot
|
27 |
|
|
|
106 |
"pt": "AutoModelForCausalLM"
|
107 |
}
|
108 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
109 |
"eos_token_id": 0,
|
110 |
"hidden_act": "gelu",
|
111 |
"hidden_size": 5120,
|
|
|
133 |
Model validation results using [EleutherAI lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness).
|
134 |
|
135 |
|
136 |
+
[eval source code](https://github.com/h2oai/h2ogpt/issues/125#issuecomment-1548239108)
|
137 |
|
138 |
| Task |Version| Metric |Value | |Stderr|
|
139 |
|-------------|------:|--------|-----:|---|-----:|
|
140 |
+
|arc_challenge| 0|acc |0.3157|± |0.0136|
|
141 |
+
| | |acc_norm|0.3507|± |0.0139|
|
142 |
|arc_easy | 0|acc |0.6932|± |0.0095|
|
143 |
| | |acc_norm|0.6225|± |0.0099|
|
144 |
+
|boolq | 1|acc |0.6685|± |0.0082|
|
|
|
|
|
145 |
|hellaswag | 0|acc |0.5140|± |0.0050|
|
146 |
| | |acc_norm|0.6803|± |0.0047|
|
147 |
+
|openbookqa | 0|acc |0.2900|± |0.0203|
|
148 |
+
| | |acc_norm|0.3740|± |0.0217|
|
149 |
|piqa | 0|acc |0.7682|± |0.0098|
|
150 |
| | |acc_norm|0.7661|± |0.0099|
|
151 |
+
|winogrande | 0|acc |0.6369|± |0.0135|
|
|
|
|
|
152 |
|
153 |
|
154 |
## Disclaimer
|