Update README.md
Browse files
README.md
CHANGED
@@ -15,6 +15,7 @@ This is the un-quantized fp16 version for training and merging. If you want the
|
|
15 |
This model is a TIES merger of Mixtral-8x7B-Instruct-v0.1 and bagel-dpo-8x7b-v0.2 with MixtralOrochi8x7B being the Base model.
|
16 |
|
17 |
|
|
|
18 |
I was very impressed with MixtralOrochi8x7B performance and multifaceted usecases as it is already a merger of many usefull Mixtral models such as Mixtral instruct,
|
19 |
Noromaid-v0.1-mixtral, openbuddy-mixtral and possibly other models that were not named. My goal was to expand the models capabilities and make it even more useful of a model, maybe even competitive with closed source models like Gpt-4. But for that more testing is required. I hope the community can help me determine if its deserving of its name. 😊
|
20 |
|
@@ -56,4 +57,5 @@ parameters:
|
|
56 |
dtype: float16
|
57 |
|
58 |
|
|
|
59 |
```
|
|
|
15 |
This model is a TIES merger of Mixtral-8x7B-Instruct-v0.1 and bagel-dpo-8x7b-v0.2 with MixtralOrochi8x7B being the Base model.
|
16 |
|
17 |
|
18 |
+
|
19 |
I was very impressed with MixtralOrochi8x7B performance and multifaceted usecases as it is already a merger of many usefull Mixtral models such as Mixtral instruct,
|
20 |
Noromaid-v0.1-mixtral, openbuddy-mixtral and possibly other models that were not named. My goal was to expand the models capabilities and make it even more useful of a model, maybe even competitive with closed source models like Gpt-4. But for that more testing is required. I hope the community can help me determine if its deserving of its name. 😊
|
21 |
|
|
|
57 |
dtype: float16
|
58 |
|
59 |
|
60 |
+
|
61 |
```
|