Finetuning
Browse files- README.md +1 -22
- adapter_config.json +2 -0
- adapter_model.bin +1 -1
README.md
CHANGED
@@ -1,27 +1,6 @@
|
|
1 |
---
|
2 |
library_name: peft
|
3 |
-
license: cc-by-nc-2.0
|
4 |
-
language:
|
5 |
-
- fr
|
6 |
-
- en
|
7 |
-
tags:
|
8 |
-
- pytorch
|
9 |
-
- llama
|
10 |
-
- code
|
11 |
---
|
12 |
-
|
13 |
-
## Aria 7B V3
|
14 |
-
|
15 |
-
We decided to build a V3 of Aria 7B based on Mistral instruct instead of LLAMA 2. The base model has been quantized with Qlora to reduce the model size and trained on a high quality french dataset.
|
16 |
-
|
17 |
-
## Base Model : Mistral-7B-Instruct-v0.1
|
18 |
-
|
19 |
-
## Technical issues Fixed & Limits of base model
|
20 |
-
|
21 |
-
We noticed that the base model had a common issue of mixing french and english when the request was done in french in some cases,not all of them. This issue was more visible for
|
22 |
-
prompts over 1000 tokens. By training the base model on our dataset, we fixed this issue and allow the model to reply in the same specific language used for the question to answer.
|
23 |
-
This pain-point is a valuable upgrade for corporate users in non-english areas willing to deploy a model with an increased quality and accuracy in french language.
|
24 |
-
|
25 |
## Training procedure
|
26 |
|
27 |
|
@@ -39,4 +18,4 @@ The following `bitsandbytes` quantization config was used during training:
|
|
39 |
### Framework versions
|
40 |
|
41 |
|
42 |
-
- PEFT 0.
|
|
|
1 |
---
|
2 |
library_name: peft
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
## Training procedure
|
5 |
|
6 |
|
|
|
18 |
### Framework versions
|
19 |
|
20 |
|
21 |
+
- PEFT 0.6.0.dev0
|
adapter_config.json
CHANGED
@@ -1,4 +1,5 @@
|
|
1 |
{
|
|
|
2 |
"auto_mapping": null,
|
3 |
"base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
|
4 |
"bias": "none",
|
@@ -12,6 +13,7 @@
|
|
12 |
"modules_to_save": null,
|
13 |
"peft_type": "LORA",
|
14 |
"r": 16,
|
|
|
15 |
"revision": null,
|
16 |
"target_modules": [
|
17 |
"q_proj",
|
|
|
1 |
{
|
2 |
+
"alpha_pattern": {},
|
3 |
"auto_mapping": null,
|
4 |
"base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
|
5 |
"bias": "none",
|
|
|
13 |
"modules_to_save": null,
|
14 |
"peft_type": "LORA",
|
15 |
"r": 16,
|
16 |
+
"rank_pattern": {},
|
17 |
"revision": null,
|
18 |
"target_modules": [
|
19 |
"q_proj",
|
adapter_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 27308941
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4d39ab534317fcef11f07327bbd422a5d23e196697a85bccfdb51abd000b5070
|
3 |
size 27308941
|