ruslandev commited on
Commit
22a845c
1 Parent(s): e059b14

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -2
README.md CHANGED
@@ -9,6 +9,8 @@ tags:
9
  - llama
10
  - trl
11
  base_model: unsloth/llama-3-8b-bnb-4bit
 
 
12
  ---
13
 
14
  # Uploaded model
@@ -17,6 +19,22 @@ base_model: unsloth/llama-3-8b-bnb-4bit
17
  - **License:** apache-2.0
18
  - **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
19
 
20
- This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
 
 
 
21
 
22
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  - llama
10
  - trl
11
  base_model: unsloth/llama-3-8b-bnb-4bit
12
+ datasets:
13
+ - cognitivecomputations/samantha-data
14
  ---
15
 
16
  # Uploaded model
 
19
  - **License:** apache-2.0
20
  - **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
21
 
22
+ This model is finetuned on the data of [Samantha](https://erichartford.com/meet-samantha).
23
+ Prompt format is Alpaca. I used the same system prompt as the original Samantha.
24
+ ```
25
+ """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
26
 
27
+ ### Instruction:
28
+ {SYSTEM_PROMPT}
29
+
30
+ ### Input:
31
+ {QUESTION}
32
+
33
+ ### Response:
34
+ """
35
+ ```
36
+ [Training code is here](https://github.com/RuslanPeresy/gptchain)
37
+
38
+ 2 epoch finetuning from llama-3-8b took 1 hour on a single A100 with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
39
+
40
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)