Spaces:
Sleeping
Sleeping
Update config/config.json
Browse files- config/config.json +6 -11
config/config.json
CHANGED
@@ -1,14 +1,14 @@
|
|
1 |
{
|
2 |
-
"model":{
|
3 |
-
"params":{
|
4 |
"output_dir": "outputs",
|
5 |
"num_train_epochs": 3,
|
6 |
"per_device_train_batch_size": 4,
|
7 |
"gradient_accumulation_steps": 8,
|
8 |
"learning_rate": 5e-5,
|
9 |
"warmup_steps": 10,
|
10 |
-
"fp16":
|
11 |
-
"bf16":
|
12 |
"logging_steps": 1,
|
13 |
"optim": "adamw_8bit",
|
14 |
"weight_decay": 0.01,
|
@@ -16,13 +16,8 @@
|
|
16 |
"seed": 42,
|
17 |
"max_seq_length": 128,
|
18 |
"dataset_num_proc": 2,
|
19 |
-
"packing":
|
20 |
},
|
21 |
-
|
22 |
}
|
23 |
}
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
}
|
|
|
1 |
{
|
2 |
+
"model": {
|
3 |
+
"params": {
|
4 |
"output_dir": "outputs",
|
5 |
"num_train_epochs": 3,
|
6 |
"per_device_train_batch_size": 4,
|
7 |
"gradient_accumulation_steps": 8,
|
8 |
"learning_rate": 5e-5,
|
9 |
"warmup_steps": 10,
|
10 |
+
"fp16": true,
|
11 |
+
"bf16": false,
|
12 |
"logging_steps": 1,
|
13 |
"optim": "adamw_8bit",
|
14 |
"weight_decay": 0.01,
|
|
|
16 |
"seed": 42,
|
17 |
"max_seq_length": 128,
|
18 |
"dataset_num_proc": 2,
|
19 |
+
"packing": false
|
20 |
},
|
21 |
+
"choices": ["gpt2", "bert-base-uncased", "llama3-8b"]
|
22 |
}
|
23 |
}
|
|
|
|
|
|
|
|
|
|