TheBloke commited on
Commit
17c1094
1 Parent(s): cc24292

Initial FP16 model commit

Browse files
Files changed (1) hide show
  1. README.md +196 -0
README.md ADDED
@@ -0,0 +1,196 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ inference: false
3
+ license: other
4
+ model_creator: nRuaif
5
+ model_link: https://huggingface.co/nRuaif/Kimiko_13B
6
+ model_name: Kimiko 13B
7
+ model_type: llama
8
+ quantized_by: TheBloke
9
+ ---
10
+
11
+ <!-- header start -->
12
+ <div style="width: 100%;">
13
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
14
+ </div>
15
+ <div style="display: flex; justify-content: space-between; width: 100%;">
16
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
17
+ <p><a href="https://discord.gg/theblokeai">Chat & support: my new Discord server</a></p>
18
+ </div>
19
+ <div style="display: flex; flex-direction: column; align-items: flex-end;">
20
+ <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p>
21
+ </div>
22
+ </div>
23
+ <!-- header end -->
24
+
25
+ # Kimiko 13B - FP16
26
+ - Model creator: [nRuaif](https://huggingface.co/nRuaif)
27
+ - Original model: [Kimiko 13B](nRuaif/Kimiko_13B)
28
+
29
+ ## Description
30
+
31
+ This repo contains pytorch format fp16 model files for [none](nRuaif/Kimiko_13B).
32
+
33
+ It is the result of merging and/or converting the source repository to float16.
34
+
35
+ ## Repositories available
36
+
37
+ * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/Kimiko-13B-GPTQ)
38
+ * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/Kimiko-13B-GGML)
39
+ * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/Kimiko-13B-fp16)
40
+ * [nRuaif's original LoRA adapter, which can be merged on to the base model.](https://huggingface.co/nRuaif/Kimiko_13B)
41
+
42
+ ## Prompt template: %%PROMPT_TEMPLATE_TITLE
43
+
44
+ ```
45
+ <<HUMAN>>
46
+ {prompt}
47
+
48
+ <<AIBOT>>
49
+ ```
50
+
51
+ <!-- footer start -->
52
+ ## Discord
53
+
54
+ For further support, and discussions on these models and AI in general, join us at:
55
+
56
+ [TheBloke AI's Discord server](https://discord.gg/theblokeai)
57
+
58
+ ## Thanks, and how to contribute.
59
+
60
+ Thanks to the [chirper.ai](https://chirper.ai) team!
61
+
62
+ I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training.
63
+
64
+ If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects.
65
+
66
+ Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits.
67
+
68
+ * Patreon: https://patreon.com/TheBlokeAI
69
+ * Ko-Fi: https://ko-fi.com/TheBlokeAI
70
+
71
+ **Special thanks to**: Luke from CarbonQuill, Aemon Algiz.
72
+
73
+ **Patreon special mentions**: Slarti, Chadd, John Detwiler, Pieter, zynix, K, Mano Prime, ReadyPlayerEmma, Ai Maven, Leonard Tan, Edmond Seymore, Joseph William Delisle, Luke @flexchar, Fred von Graf, Viktor Bowallius, Rishabh Srivastava, Nikolai Manek, Matthew Berman, Johann-Peter Hartmann, ya boyyy, Greatston Gnanesh, Femi Adebogun, Talal Aujan, Jonathan Leane, terasurfer, David Flickinger, William Sang, Ajan Kanaga, Vadim, Artur Olbinski, Raven Klaugh, Michael Levine, Oscar Rangel, Randy H, Cory Kujawski, RoA, Dave, Alex, Alexandros Triantafyllidis, Fen Risland, Eugene Pentland, vamX, Elle, Nathan LeClaire, Khalefa Al-Ahmad, Rainer Wilmers, subjectnull, Junyu Yang, Daniel P. Andersen, SuperWojo, LangChain4j, Mandus, Kalila, Illia Dulskyi, Trenton Dambrowitz, Asp the Wyvern, Derek Yates, Jeffrey Morgan, Deep Realms, Imad Khwaja, Pyrater, Preetika Verma, biorpg, Gabriel Tamborski, Stephen Murray, Spiking Neurons AB, Iucharbius, Chris Smitley, Willem Michiel, Luke Pendergrass, Sebastain Graf, senxiiz, Will Dee, Space Cruiser, Karl Bernard, Clay Pascal, Lone Striker, transmissions 11, webtim, WelcomeToTheClub, Sam, theTransient, Pierre Kircher, chris gileta, John Villwock, Sean Connelly, Willian Hasse
74
+
75
+
76
+ Thank you to all my generous patrons and donaters!
77
+
78
+ <!-- footer end -->
79
+
80
+ # Original model card: none
81
+
82
+
83
+ # Model Card for Kimiko_13B
84
+
85
+ <!-- Provide a quick summary of what the model is/does. -->
86
+
87
+ This is my new Kimiko models, trained with LLaMA2-13B for...purpose
88
+
89
+ ## Model Details
90
+
91
+ ### Model Description
92
+
93
+ <!-- Provide a longer summary of what this model is. -->
94
+
95
+
96
+
97
+ - **Developed by:** nRuaif
98
+ - **Model type:** Decoder only
99
+ - **License:** CC BY-NC-SA
100
+ - **Finetuned from model [optional]:** LLaMA 2
101
+
102
+ ### Model Sources [optional]
103
+
104
+ <!-- Provide the basic links for the model. -->
105
+
106
+ - **Repository:** https://github.com/OpenAccess-AI-Collective/axolotl
107
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
108
+ ## Uses
109
+
110
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
111
+
112
+
113
+ ### Direct Use
114
+
115
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
116
+
117
+ This model is trained on 3k examples of instructions dataset, high quality roleplay, for best result follow this format
118
+ ```
119
+ <<HUMAN>>
120
+ How to do abc
121
+
122
+ <<AIBOT>>
123
+ Here is how
124
+
125
+ Or with system prompting for roleplay
126
+
127
+ <<SYSTEM>>
128
+ A's Persona:
129
+ B's Persona:
130
+ Scenario:
131
+ Add some instruction here on how you want your RP to go.
132
+ ```
133
+
134
+
135
+ ## Bias, Risks, and Limitations
136
+
137
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
138
+
139
+ All bias of this model come from LlaMA2 with an exception of NSFW bias.....
140
+
141
+
142
+
143
+
144
+ ## Training Details
145
+
146
+ ### Training Data
147
+
148
+ <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
149
+
150
+ 3000 examples from LIMAERP, LIMA and I sample 1000 good instruction from Airboro
151
+
152
+ ### Training Procedure
153
+
154
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
155
+
156
+ Model is trained with 1 L4 from GCP costing a whooping 2.5USD
157
+
158
+
159
+
160
+
161
+
162
+ #### Training Hyperparameters
163
+
164
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
165
+
166
+ 3 epochs with 0.0002 lr, full 4096 ctx token, QLoRA
167
+
168
+ #### Speeds, Sizes, Times [optional]
169
+
170
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
171
+
172
+ It takes 18 hours to train this model with xformers enable
173
+
174
+ [More Information Needed]
175
+
176
+
177
+
178
+
179
+
180
+
181
+
182
+ [More Information Needed]
183
+
184
+ ## Environmental Impact
185
+
186
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
187
+
188
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
189
+
190
+ - **Hardware Type:** L4 with 12CPUs 48gb ram
191
+ - **Hours used:** 5
192
+ - **Cloud Provider:** GCP
193
+ - **Compute Region:** US
194
+ - **Carbon Emitted:** 0.5KG
195
+
196
+