File size: 721 Bytes
b8b0aaa
d116aef
dc813e2
18ea001
 
 
 
dc813e2
18ea001
 
dc813e2
 
b8b0aaa
 
 
 
 
 
d116aef
 
b8b0aaa
 
 
 
 
 
 
d116aef
b8b0aaa
d116aef
b8b0aaa
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
{
  "model": {
    "peft": {
      "r": 64,
      "alpha": 16,
      "dropout": 0,
      "bias": "none",
      "seed": 3407,
      "rslora": false,
      "loftq_config": null
    },
    "sft": {
      "output_dir": "outputs",
      "num_train_epochs": 3,
      "per_device_train_batch_size": 4,
      "gradient_accumulation_steps": 8,
      "learning_rate": 5e-5,
      "warmup_steps": 10,
      "fp16": true,
      "bf16": false,
      "logging_steps": 1,
      "optim": "adamw_8bit",
      "weight_decay": 0.01,
      "lr_scheduler_type": "linear",
      "seed": 42,
      "max_seq_length": 128,
      "dataset_num_proc": 2,
      "packing": false
    },
    "choices": ["gpt2", "bert-base-uncased", "llama3-8b"]
  }
}