tykiww commited on
Commit
d116aef
1 Parent(s): 569f592

Update config/config.json

Browse files
Files changed (1) hide show
  1. config/config.json +6 -11
config/config.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
- "model":{
3
- "params":{
4
  "output_dir": "outputs",
5
  "num_train_epochs": 3,
6
  "per_device_train_batch_size": 4,
7
  "gradient_accumulation_steps": 8,
8
  "learning_rate": 5e-5,
9
  "warmup_steps": 10,
10
- "fp16": True,
11
- "bf16": False,
12
  "logging_steps": 1,
13
  "optim": "adamw_8bit",
14
  "weight_decay": 0.01,
@@ -16,13 +16,8 @@
16
  "seed": 42,
17
  "max_seq_length": 128,
18
  "dataset_num_proc": 2,
19
- "packing": False
20
  },
21
- "choices": ["gpt2", "bert-base-uncased", "llama3-8b"]
22
  }
23
  }
24
-
25
-
26
-
27
-
28
- }
 
1
  {
2
+ "model": {
3
+ "params": {
4
  "output_dir": "outputs",
5
  "num_train_epochs": 3,
6
  "per_device_train_batch_size": 4,
7
  "gradient_accumulation_steps": 8,
8
  "learning_rate": 5e-5,
9
  "warmup_steps": 10,
10
+ "fp16": true,
11
+ "bf16": false,
12
  "logging_steps": 1,
13
  "optim": "adamw_8bit",
14
  "weight_decay": 0.01,
 
16
  "seed": 42,
17
  "max_seq_length": 128,
18
  "dataset_num_proc": 2,
19
+ "packing": false
20
  },
21
+ "choices": ["gpt2", "bert-base-uncased", "llama3-8b"]
22
  }
23
  }