grimjim commited on
Commit
c8899f3
1 Parent(s): 1d24f38

Initial release

Browse files
.gitattributes CHANGED
@@ -4,6 +4,7 @@
4
  *.bz2 filter=lfs diff=lfs merge=lfs -text
5
  *.ckpt filter=lfs diff=lfs merge=lfs -text
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
 
7
  *.gz filter=lfs diff=lfs merge=lfs -text
8
  *.h5 filter=lfs diff=lfs merge=lfs -text
9
  *.joblib filter=lfs diff=lfs merge=lfs -text
 
4
  *.bz2 filter=lfs diff=lfs merge=lfs -text
5
  *.ckpt filter=lfs diff=lfs merge=lfs -text
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gguf filter=lfs diff=lfs merge=lfs -text
8
  *.gz filter=lfs diff=lfs merge=lfs -text
9
  *.h5 filter=lfs diff=lfs merge=lfs -text
10
  *.joblib filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,52 @@
1
- ---
2
- license: cc-by-nc-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B
4
+ - grimjim/kunoichi-lemon-royale-7B
5
+ library_name: transformers
6
+ tags:
7
+ - mergekit
8
+ - merge
9
+ license: cc-by-nc-4.0
10
+ pipeline_tag: text-generation
11
+ ---
12
+ # kunoichi-lemon-royale-v2-32K-7B-GGUF
13
+
14
+ This repository contains GGUF quantizations of kunoichi-lemon-royale-v2-32K-7B.
15
+
16
+ This merge amounts to the grafting of a model derived from Mistral v0.1 (4K sliding window context, to a maximum of 8K practical context length) onto a model derived from Mistral v0.2 (32K context length). It appears to work, although rope_theta in config.json was lowered from 1000000.0 to 100000.0, which works well enough to 16K.
17
+
18
+ In light testing, this model appears to follow formatting very well, with temperature 1.0 and minP 0.01, using ChatML prompts, even though the underlying model claims to follow Alpaca prompts.
19
+
20
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
21
+
22
+ - Full weights: [kunoichi-lemon-royale-v2-32K-7B](https://huggingface.co/grimjim/kunoichi-lemon-royale-v2-32K-7B)
23
+ - GGUFs: [kunoichi-lemon-royale-v2-32K-7B-GGUF](https://huggingface.co/grimjim/kunoichi-lemon-royale-v2-32K-7B-GGUF)
24
+
25
+ ## Merge Details
26
+ ### Merge Method
27
+
28
+ This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B](https://huggingface.co/grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B) as a base.
29
+
30
+ ### Models Merged
31
+
32
+ The following models were included in the merge:
33
+ * [grimjim/kunoichi-lemon-royale-7B](https://huggingface.co/grimjim/kunoichi-lemon-royale-7B)
34
+
35
+ ### Configuration
36
+
37
+ The following YAML configuration was used to produce this model:
38
+
39
+ ```yaml
40
+ base_model: grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B
41
+ dtype: bfloat16
42
+ merge_method: task_arithmetic
43
+ slices:
44
+ - sources:
45
+ - layer_range: [0, 32]
46
+ model: grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B
47
+ - layer_range: [0, 32]
48
+ model: grimjim/kunoichi-lemon-royale-7B
49
+ parameters:
50
+ weight: 1.00
51
+
52
+ ```
kunoichi-lemon-royale-v2-32K-7B.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8457b5e629f0f4096bb80ddbab23f893ad00f7e46e132ffad29fd35a70353fb
3
+ size 4368439552
kunoichi-lemon-royale-v2-32K-7B.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c42e60e4ba5f9d098b351efcbf884b3f960d9eae3276df910f5bb773cc27bd3
3
+ size 5131409664
kunoichi-lemon-royale-v2-32K-7B.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:389074a0a90c1e009843a82191a6d804f85a82831563e1c25db7888bc65858e5
3
+ size 5942065408
kunoichi-lemon-royale-v2-32K-7B.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:181433e3f25ecf34f0ddcbe81ccd11c3325aa249b902d3bb68d721b8ce058d13
3
+ size 7695857888