Update README.md
Browse files
README.md
CHANGED
@@ -1,9 +1,8 @@
|
|
1 |
---
|
2 |
-
base_model:
|
3 |
tags:
|
4 |
- text-generation-inference
|
5 |
- transformers
|
6 |
-
- unsloth
|
7 |
- qwen2
|
8 |
- trl
|
9 |
- sft
|
@@ -53,5 +52,8 @@ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
|
|
53 |
```
|
54 |
|
55 |
# Notes:
|
56 |
-
-
|
57 |
- Fine-tuned lora with rank = 16 and alpha = 32, epoch = 1
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
base_model: Qwen2.5-3B-Instruct
|
3 |
tags:
|
4 |
- text-generation-inference
|
5 |
- transformers
|
|
|
6 |
- qwen2
|
7 |
- trl
|
8 |
- sft
|
|
|
52 |
```
|
53 |
|
54 |
# Notes:
|
55 |
+
- For small datasets with narrow content which the model already has well, and doesn't want the model to forget the knowledge by focusing on o.
|
56 |
- Fine-tuned lora with rank = 16 and alpha = 32, epoch = 1
|
57 |
+
|
58 |
+
# Improvement
|
59 |
+
- Increasing rank can help the model do better at your robust structure.
|