Librarian Bot: Add moe tag to model
#2
by
librarian-bot
- opened
README.md
CHANGED
@@ -1,5 +1,7 @@
|
|
1 |
---
|
2 |
-
|
|
|
|
|
3 |
tags:
|
4 |
- mixtral
|
5 |
- instruct
|
@@ -8,12 +10,11 @@ tags:
|
|
8 |
- gpt4
|
9 |
- synthetic data
|
10 |
- distillation
|
|
|
|
|
11 |
model-index:
|
12 |
- name: OpenHermes-Mixtral-8x7B
|
13 |
results: []
|
14 |
-
license: apache-2.0
|
15 |
-
language:
|
16 |
-
- en
|
17 |
---
|
18 |
|
19 |
# OpenHermes - Mixtral 8x7B
|
|
|
1 |
---
|
2 |
+
language:
|
3 |
+
- en
|
4 |
+
license: apache-2.0
|
5 |
tags:
|
6 |
- mixtral
|
7 |
- instruct
|
|
|
10 |
- gpt4
|
11 |
- synthetic data
|
12 |
- distillation
|
13 |
+
- moe
|
14 |
+
base_model: mistralai/Mixtral-8x7B-Instruct-v0.1
|
15 |
model-index:
|
16 |
- name: OpenHermes-Mixtral-8x7B
|
17 |
results: []
|
|
|
|
|
|
|
18 |
---
|
19 |
|
20 |
# OpenHermes - Mixtral 8x7B
|