asiansoul commited on
Commit
b8bc3fe
1 Parent(s): 632998f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +66 -2
README.md CHANGED
@@ -1,5 +1,69 @@
1
  ---
2
  license: other
3
- license_name: other
4
- license_link: LICENSE
 
 
 
 
 
 
 
 
 
 
 
5
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: other
3
+ base_model:
4
+ - beomi/Llama-3-Open-Ko-8B-Instruct-preview
5
+ - cognitivecomputations/dolphin-2.9-llama3-8b
6
+ - NousResearch/Meta-Llama-3-8B-Instruct
7
+ - NousResearch/Meta-Llama-3-8B
8
+ - abacusai/Llama-3-Smaug-8B
9
+ - Locutusque/Llama-3-Orca-1.0-8B
10
+ library_name: transformers
11
+ tags:
12
+ - mergekit
13
+ - merge
14
+ - llama
15
+
16
  ---
17
+ # KoDolph
18
+
19
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
20
+
21
+ ## Merge Details
22
+ ### Merge Method
23
+
24
+ This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [NousResearch/Meta-Llama-3-8B](https://huggingface.co/NousResearch/Meta-Llama-3-8B) as a base.
25
+
26
+ ### Models Merged
27
+
28
+ The following models were included in the merge:
29
+ * [beomi/Llama-3-Open-Ko-8B-Instruct-preview](https://huggingface.co/beomi/Llama-3-Open-Ko-8B-Instruct-preview)
30
+ * [cognitivecomputations/dolphin-2.9-llama3-8b](https://huggingface.co/cognitivecomputations/dolphin-2.9-llama3-8b)
31
+ * [NousResearch/Meta-Llama-3-8B-Instruct](https://huggingface.co/NousResearch/Meta-Llama-3-8B-Instruct)
32
+ * [abacusai/Llama-3-Smaug-8B](https://huggingface.co/abacusai/Llama-3-Smaug-8B)
33
+ * [Locutusque/Llama-3-Orca-1.0-8B](https://huggingface.co/Locutusque/Llama-3-Orca-1.0-8B)
34
+
35
+ ### Configuration
36
+
37
+ The following YAML configuration was used to produce this model:
38
+
39
+ ```yaml
40
+ models:
41
+ - model: NousResearch/Meta-Llama-3-8B
42
+ # Base model providing a general foundation without specific parameters
43
+ - model: NousResearch/Meta-Llama-3-8B-Instruct
44
+ parameters:
45
+ density: 0.58
46
+ weight: 0.25
47
+ - model: cognitivecomputations/dolphin-2.9-llama3-8b
48
+ parameters:
49
+ density: 0.52
50
+ weight: 0.15
51
+ - model: Locutusque/Llama-3-Orca-1.0-8B
52
+ parameters:
53
+ density: 0.52
54
+ weight: 0.15
55
+ - model: abacusai/Llama-3-Smaug-8B
56
+ parameters:
57
+ density: 0.52
58
+ weight: 0.15
59
+ - model: beomi/Llama-3-Open-Ko-8B-Instruct-preview
60
+ parameters:
61
+ density: 0.53
62
+ weight: 0.2
63
+ merge_method: dare_ties
64
+ base_model: NousResearch/Meta-Llama-3-8B
65
+ parameters:
66
+ int8_mask: true
67
+ dtype: bfloat16
68
+ ```
69
+