Update README.md
Browse files
README.md
CHANGED
@@ -9,7 +9,27 @@ tags:
|
|
9 |
- mistral
|
10 |
- trl
|
11 |
base_model: augmxnt/shisa-base-7b-v1
|
|
|
|
|
|
|
12 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
|
14 |
# Uploaded model
|
15 |
|
|
|
9 |
- mistral
|
10 |
- trl
|
11 |
base_model: augmxnt/shisa-base-7b-v1
|
12 |
+
datasets:
|
13 |
+
- NilanE/ParallelFiction-Ja_En-100k
|
14 |
+
- mpasila/ParallelFiction-Ja_En-100k-alpaca
|
15 |
---
|
16 |
+
Experimental model, may not perform that well. Dataset used is [a modified](https://huggingface.co/datasets/mpasila/ParallelFiction-Ja_En-100k-alpaca) version of [NilanE/ParallelFiction-Ja_En-100k](https://huggingface.co/datasets/NilanE/ParallelFiction-Ja_En-100k).
|
17 |
+
|
18 |
+
After training with an 8k context length it didn't appear to improve performance much at all. Not sure if I should keep training it (which is costly) or if I should fix some issues with the dataset (like it starting with Ch or Chapter) or I go back to finetuning Finnish models.
|
19 |
+
|
20 |
+
### Prompt format: Alpaca
|
21 |
+
```
|
22 |
+
Below is a translation task, paired with an input that provides further context. Write a response that appropriately completes the request.
|
23 |
+
|
24 |
+
### Instruction:
|
25 |
+
{}
|
26 |
+
|
27 |
+
### Input:
|
28 |
+
{}
|
29 |
+
|
30 |
+
### Response:
|
31 |
+
{}
|
32 |
+
```
|
33 |
|
34 |
# Uploaded model
|
35 |
|