File size: 1,269 Bytes
b4b0ef9 49c9d38 5eadf18 b4b0ef9 49c9d38 537fbe0 49c9d38 10de8b0 49c9d38 ea4bf73 14552c8 49c9d38 3feef72 91b790a 3feef72 49c9d38 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 |
---
license: apache-2.0
tags:
- moe
- merge
- mergekit
- lazymergekit
- mlabonne/NeuralBeagle14-7B
- mlabonne/NeuralDaredevil-7B
- text-generation-inference
- Text Generation
---
---
**This is a repository of GGUF Quants for DareBeagel-2x7B**
---
Original Model Available Here: https://huggingface.co/shadowml/DareBeagel-2x7B
**Available Quants**
* Q8_0
* Q6_K
* Q5_K_M
* Q5_K_S
* Q4_K_M
* Q4_K_S
* Q3_K_M
* Q3_K_S
* Q2_K
# Beyonder-2x7B-v2
Beyonder-2x7B-v2 is a Mixure of Experts (MoE) made with the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
* [mlabonne/NeuralBeagle14-7B](https://huggingface.co/mlabonne/NeuralBeagle14-7B)
* [mlabonne/NeuralDaredevil-7B](https://huggingface.co/mlabonne/NeuralDaredevil-7B)
## 🧩 Configuration
```yaml
base_model: mlabonne/NeuralBeagle14-7B
gate_mode: random
experts:
- source_model: mlabonne/NeuralBeagle14-7B
positive_prompts: [""]
- source_model: mlabonne/NeuralDaredevil-7B
positive_prompts: [""]
```
## 💻 Usage
```
Load in Kobold.cpp or whatever.
I found Alpaca (and Alpaca-ish) prompts worked well.
Settings that worked good for me are:
Min P - 0.1
Dynamic Temperature Min 0 Max 3
Rep Pen 1.03
Rep Pen Range 1000
``` |