willtensora commited on
Commit
f183e36
·
verified ·
1 Parent(s): 49e6963

Training in progress, step 3, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -10,23 +10,23 @@
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
- "lora_alpha": 8,
14
- "lora_dropout": 0.1,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 4,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "down_proj",
24
- "up_proj",
25
  "v_proj",
26
- "k_proj",
27
- "gate_proj",
28
  "o_proj",
29
- "q_proj"
 
 
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
+ "lora_alpha": 16,
14
+ "lora_dropout": 0.05,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
+ "r": 8,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "v_proj",
 
 
24
  "o_proj",
25
+ "gate_proj",
26
+ "q_proj",
27
+ "up_proj",
28
+ "down_proj",
29
+ "k_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7e711b7b4034b2f3bd9799e3738e807081a18d9dd9d4402e426cefbb04d49c9
3
- size 9048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab90251ab6e27a57801a12d18af11f2f9f07fcab223af845d95de1a0b405501c
3
+ size 14696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e58e24af97e29d770d88764fcd1061468363d04933ab5bc770d00cb8fc8b69f1
3
- size 28134
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b12248ecbc96221e2afdb2d474ebe9b42802236a73a19ebb91f18881b9c511f
3
+ size 39398
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b789b44e35b8f0004dd2e31bfb1af056b8528eeb7ffa2f97d314e99f11187ccd
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:464e6cdff3f8548ba79ace3859c72e6b06d1e045e69932eb75c584ad0e8918c8
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6da16eefab22daee3fb0999c7651be192feb36b73810bf419e9165c7023ec02
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffd8c58e5d02492554dbaa495f8cf80dff41fabc0e1288cb2fd18a103d402219
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,34 +1,56 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.0002666666666666667,
5
- "eval_steps": 500,
6
- "global_step": 1,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.0002666666666666667,
13
- "grad_norm": 0.014549034647643566,
14
- "learning_rate": 0.0,
15
- "loss": 11.9308,
16
  "step": 1
17
  },
18
  {
19
- "epoch": 0.0002666666666666667,
20
- "eval_loss": 11.933904647827148,
21
- "eval_runtime": 0.0581,
22
- "eval_samples_per_second": 275.597,
23
- "eval_steps_per_second": 68.899,
24
  "step": 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
  }
26
  ],
27
  "logging_steps": 1,
28
- "max_steps": 1,
29
  "num_input_tokens_seen": 0,
30
  "num_train_epochs": 1,
31
- "save_steps": 500,
32
  "stateful_callbacks": {
33
  "TrainerControl": {
34
  "args": {
@@ -36,13 +58,13 @@
36
  "should_evaluate": false,
37
  "should_log": false,
38
  "should_save": true,
39
- "should_training_stop": true
40
  },
41
  "attributes": {}
42
  }
43
  },
44
- "total_flos": 10444800.0,
45
- "train_batch_size": 4,
46
  "trial_name": null,
47
  "trial_params": null
48
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.0016825574873808188,
5
+ "eval_steps": 3,
6
+ "global_step": 3,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.0005608524957936063,
13
+ "grad_norm": 0.014698871411383152,
14
+ "learning_rate": 2e-05,
15
+ "loss": 11.9315,
16
  "step": 1
17
  },
18
  {
19
+ "epoch": 0.0005608524957936063,
20
+ "eval_loss": 11.931306838989258,
21
+ "eval_runtime": 2.6652,
22
+ "eval_samples_per_second": 281.781,
23
+ "eval_steps_per_second": 141.078,
24
  "step": 1
25
+ },
26
+ {
27
+ "epoch": 0.0011217049915872126,
28
+ "grad_norm": 0.009716438129544258,
29
+ "learning_rate": 4e-05,
30
+ "loss": 11.9304,
31
+ "step": 2
32
+ },
33
+ {
34
+ "epoch": 0.0016825574873808188,
35
+ "grad_norm": 0.013599561527371407,
36
+ "learning_rate": 6e-05,
37
+ "loss": 11.9319,
38
+ "step": 3
39
+ },
40
+ {
41
+ "epoch": 0.0016825574873808188,
42
+ "eval_loss": 11.931304931640625,
43
+ "eval_runtime": 2.7015,
44
+ "eval_samples_per_second": 277.993,
45
+ "eval_steps_per_second": 139.182,
46
+ "step": 3
47
  }
48
  ],
49
  "logging_steps": 1,
50
+ "max_steps": 10,
51
  "num_input_tokens_seen": 0,
52
  "num_train_epochs": 1,
53
+ "save_steps": 3,
54
  "stateful_callbacks": {
55
  "TrainerControl": {
56
  "args": {
 
58
  "should_evaluate": false,
59
  "should_log": false,
60
  "should_save": true,
61
+ "should_training_stop": false
62
  },
63
  "attributes": {}
64
  }
65
  },
66
+ "total_flos": 384024576.0,
67
+ "train_batch_size": 2,
68
  "trial_name": null,
69
  "trial_params": null
70
  }
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a34c87415c9e639533cb4d7b940fdf4014edaec46abd1ff3438eb94417519db3
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:525897d6fc230472c14ab0a00d3e93c0465a5d3840498e66ad09e713ff0af362
3
  size 6776