grimjim commited on
Commit
b4a6ce0
1 Parent(s): 865e8dc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -45
README.md CHANGED
@@ -1,45 +1,45 @@
1
- ---
2
- base_model:
3
- - KoboldAI/Mistral-7B-Holodeck-1
4
- - grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B
5
- library_name: transformers
6
- tags:
7
- - mergekit
8
- - merge
9
- license: apache-2.0
10
- pipeline_tag: text-generation
11
- ---
12
- # koboldai-holodeck-extended-32k-7B -7B
13
-
14
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
15
-
16
- Effective context length has been extended.
17
-
18
- ## Merge Details
19
- ### Merge Method
20
-
21
- This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B](https://huggingface.co/grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B) as a base.
22
-
23
- ### Models Merged
24
-
25
- The following models were included in the merge:
26
- * [KoboldAI/Mistral-7B-Holodeck-1](https://huggingface.co/KoboldAI/Mistral-7B-Holodeck-1)
27
-
28
- ### Configuration
29
-
30
- The following YAML configuration was used to produce this model:
31
-
32
- ```yaml
33
- base_model: grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B
34
- dtype: bfloat16
35
- merge_method: task_arithmetic
36
- slices:
37
- - sources:
38
- - layer_range: [0, 32]
39
- model: grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B
40
- - layer_range: [0, 32]
41
- model: KoboldAI/Mistral-7B-Holodeck-1
42
- parameters:
43
- weight: 0.4
44
-
45
- ```
 
1
+ ---
2
+ base_model:
3
+ - KoboldAI/Mistral-7B-Holodeck-1
4
+ - grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B
5
+ library_name: transformers
6
+ tags:
7
+ - mergekit
8
+ - merge
9
+ license: apache-2.0
10
+ pipeline_tag: text-generation
11
+ ---
12
+ # koboldai-holodeck-extended-32k-7B
13
+
14
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
15
+
16
+ Effective context length has been extended to at least 16k.
17
+
18
+ ## Merge Details
19
+ ### Merge Method
20
+
21
+ This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B](https://huggingface.co/grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B) as a base.
22
+
23
+ ### Models Merged
24
+
25
+ The following models were included in the merge:
26
+ * [KoboldAI/Mistral-7B-Holodeck-1](https://huggingface.co/KoboldAI/Mistral-7B-Holodeck-1)
27
+
28
+ ### Configuration
29
+
30
+ The following YAML configuration was used to produce this model:
31
+
32
+ ```yaml
33
+ base_model: grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B
34
+ dtype: bfloat16
35
+ merge_method: task_arithmetic
36
+ slices:
37
+ - sources:
38
+ - layer_range: [0, 32]
39
+ model: grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B
40
+ - layer_range: [0, 32]
41
+ model: KoboldAI/Mistral-7B-Holodeck-1
42
+ parameters:
43
+ weight: 0.4
44
+
45
+ ```