Triangle104 commited on
Commit
9f9d7a3
1 Parent(s): 5bf400a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -1
README.md CHANGED
@@ -6,12 +6,46 @@ tags:
6
  - merge
7
  - llama-cpp
8
  - gguf-my-repo
 
9
  ---
10
 
11
  # Triangle104/MN-SlushoMix-Q8_0-GGUF
12
  This model was converted to GGUF format from [`crestf411/MN-SlushoMix`](https://huggingface.co/crestf411/MN-SlushoMix) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
13
  Refer to the [original model card](https://huggingface.co/crestf411/MN-SlushoMix) for more details on the model.
14
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  ## Use with llama.cpp
16
  Install llama.cpp through brew (works on Mac and Linux)
17
 
@@ -50,4 +84,4 @@ Step 3: Run inference through the main binary.
50
  or
51
  ```
52
  ./llama-server --hf-repo Triangle104/MN-SlushoMix-Q8_0-GGUF --hf-file mn-slushomix-q8_0.gguf -c 2048
53
- ```
 
6
  - merge
7
  - llama-cpp
8
  - gguf-my-repo
9
+ license: apache-2.0
10
  ---
11
 
12
  # Triangle104/MN-SlushoMix-Q8_0-GGUF
13
  This model was converted to GGUF format from [`crestf411/MN-SlushoMix`](https://huggingface.co/crestf411/MN-SlushoMix) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
14
  Refer to the [original model card](https://huggingface.co/crestf411/MN-SlushoMix) for more details on the model.
15
 
16
+ ---
17
+ Model details:
18
+ -
19
+ MN-Slush + NemoMix Unleashed.
20
+
21
+ models:
22
+ - model: slush-stage1
23
+ parameters:
24
+ weight: 1
25
+ density: 1
26
+ - model: slush-stage2
27
+ parameters:
28
+ weight: 0.7
29
+ density: 1
30
+ - model: MarinaraSpaghetti/NemoMix-Unleashed-12B
31
+ parameters:
32
+ weight: 0.9
33
+ density: 1
34
+ - model: mistralai/Mistral-Nemo-Instruct-2407
35
+ parameters:
36
+ weight: 1
37
+ density: 1
38
+ merge_method: ties
39
+ base_model: mistralai/Mistral-Nemo-Base-2407
40
+ parameters:
41
+ weight: 1
42
+ density: 1
43
+ normalize: true
44
+ int8_mask: true
45
+ tokenizer_source: mistralai/Mistral-Nemo-Instruct-2407
46
+ dtype: bfloat16
47
+
48
+ ---
49
  ## Use with llama.cpp
50
  Install llama.cpp through brew (works on Mac and Linux)
51
 
 
84
  or
85
  ```
86
  ./llama-server --hf-repo Triangle104/MN-SlushoMix-Q8_0-GGUF --hf-file mn-slushomix-q8_0.gguf -c 2048
87
+ ```