ardaspear commited on
Commit
86d1b42
1 Parent(s): 69a457d

Training in progress, step 8, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:879311a4aebc17b4fafcb549aa97630288d831fabf9f2b7c29c0be0e3f4d6cd9
3
  size 864368280
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f6a55ad0be753e29269e2f185df91f0be0e84bbc37972098c62a3986236bb65
3
  size 864368280
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3307f49c2895af38b46962aea4ec72240ac1ecffff0a114178a9d05df44a3d28
3
  size 1729074742
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c095a9c2de308e133a689b07cc4cdf84b91e74a79283411211fad3e64d594cc4
3
  size 1729074742
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:99b98c21a6c7c3c92fc68e617bba8fd566712396b2134a20b6aff6e47927480b
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd01e813cc343537008d77625d57e124afc6747a7069b071d8b63f835a36e499
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d44c0ecfb16f05885a8478afdaa40e988069bb6891bbc6c82185a9b1533b4c7
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a3d18e607a27cfab9cec7cfaa7384cdb877a2330c4bcd4e1efcae25be9908cb
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.6363636363636362,
5
  "eval_steps": 1,
6
- "global_step": 7,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -77,6 +77,14 @@
77
  "eval_samples_per_second": 7.791,
78
  "eval_steps_per_second": 3.116,
79
  "step": 7
 
 
 
 
 
 
 
 
80
  }
81
  ],
82
  "logging_steps": 3,
@@ -91,12 +99,12 @@
91
  "should_evaluate": false,
92
  "should_log": false,
93
  "should_save": true,
94
- "should_training_stop": false
95
  },
96
  "attributes": {}
97
  }
98
  },
99
- "total_flos": 1.1753603576365056e+16,
100
  "train_batch_size": 4,
101
  "trial_name": null,
102
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 3.0454545454545454,
5
  "eval_steps": 1,
6
+ "global_step": 8,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
77
  "eval_samples_per_second": 7.791,
78
  "eval_steps_per_second": 3.116,
79
  "step": 7
80
+ },
81
+ {
82
+ "epoch": 3.0454545454545454,
83
+ "eval_loss": 0.00037194820470176637,
84
+ "eval_runtime": 0.6402,
85
+ "eval_samples_per_second": 7.811,
86
+ "eval_steps_per_second": 3.124,
87
+ "step": 8
88
  }
89
  ],
90
  "logging_steps": 3,
 
99
  "should_evaluate": false,
100
  "should_log": false,
101
  "should_save": true,
102
+ "should_training_stop": true
103
  },
104
  "attributes": {}
105
  }
106
  },
107
+ "total_flos": 1.3432689801560064e+16,
108
  "train_batch_size": 4,
109
  "trial_name": null,
110
  "trial_params": null