willtensora commited on
Commit
31368fc
·
verified ·
1 Parent(s): 128df71

Training in progress, epoch 0, checkpoint

Browse files
last-checkpoint/README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: oopsung/llama2-7b-n-ox-test-v1
3
  library_name: peft
4
  ---
5
 
 
1
  ---
2
+ base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
3
  library_name: peft
4
  ---
5
 
last-checkpoint/adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "oopsung/llama2-7b-n-ox-test-v1",
5
  "bias": "none",
6
  "fan_in_fan_out": null,
7
  "inference_mode": true,
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "o_proj",
24
- "v_proj",
25
- "up_proj",
26
  "q_proj",
27
- "down_proj",
 
28
  "gate_proj",
29
- "k_proj"
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
5
  "bias": "none",
6
  "fan_in_fan_out": null,
7
  "inference_mode": true,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "q_proj",
24
+ "v_proj",
25
+ "o_proj",
26
  "gate_proj",
27
+ "up_proj",
28
+ "k_proj",
29
+ "down_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e72b4a3235c259913087ac0799deb5f0a9ce95fa5a2731f32072330633125a2
3
- size 80013120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91a9ea1223ba85bbea9348c544be11652f89e0e757979feed1bcb01f6c918891
3
+ size 25271744
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:102d34f3f27b40ee77ec12e016d974012f097b4b96ba45b8d940f87321cddd34
3
- size 41119636
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:336c1293365b64ddb78543f3bc45d81537f442d08bc1806978f0ed670257d124
3
+ size 13685516
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b7bbb9e6b1248c3560ef0c4d2a18853bb3b1d23161e5fbf5d045daee6827a56f
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a833f22f25ea3834727974efdfc06b3927711900cd9acb75c7c0258b248c9673
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bb578e75c11a81e85dda67a691f96ba4793a02960f1409fd3e1511aac873491a
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f883fd9393081a5c2e54691be1749bef50e284275a9e7a1c589cf11727a2e22b
3
  size 1064
last-checkpoint/tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
last-checkpoint/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
last-checkpoint/tokenizer_config.json CHANGED
@@ -26,46 +26,6 @@
26
  "rstrip": false,
27
  "single_word": false,
28
  "special": true
29
- },
30
- "46331": {
31
- "content": "<|sep|>",
32
- "lstrip": false,
33
- "normalized": false,
34
- "rstrip": false,
35
- "single_word": false,
36
- "special": true
37
- },
38
- "46332": {
39
- "content": "<|endoftext|>",
40
- "lstrip": false,
41
- "normalized": false,
42
- "rstrip": false,
43
- "single_word": false,
44
- "special": true
45
- },
46
- "46333": {
47
- "content": "<|acc|>",
48
- "lstrip": false,
49
- "normalized": false,
50
- "rstrip": false,
51
- "single_word": false,
52
- "special": true
53
- },
54
- "46334": {
55
- "content": "<|rrn|>",
56
- "lstrip": false,
57
- "normalized": false,
58
- "rstrip": false,
59
- "single_word": false,
60
- "special": true
61
- },
62
- "46335": {
63
- "content": "<|tel|>",
64
- "lstrip": false,
65
- "normalized": false,
66
- "rstrip": false,
67
- "single_word": false,
68
- "special": true
69
  }
70
  },
71
  "bos_token": "<s>",
@@ -73,8 +33,9 @@
73
  "clean_up_tokenization_spaces": false,
74
  "eos_token": "</s>",
75
  "legacy": false,
76
- "model_max_length": 1000000000000000019884624838656,
77
  "pad_token": "</s>",
 
78
  "sp_model_kwargs": {},
79
  "tokenizer_class": "LlamaTokenizer",
80
  "unk_token": "<unk>",
 
26
  "rstrip": false,
27
  "single_word": false,
28
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  }
30
  },
31
  "bos_token": "<s>",
 
33
  "clean_up_tokenization_spaces": false,
34
  "eos_token": "</s>",
35
  "legacy": false,
36
+ "model_max_length": 2048,
37
  "pad_token": "</s>",
38
+ "padding_side": "right",
39
  "sp_model_kwargs": {},
40
  "tokenizer_class": "LlamaTokenizer",
41
  "unk_token": "<unk>",
last-checkpoint/trainer_state.json CHANGED
@@ -1,121 +1,41 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.005608524957936063,
5
- "eval_steps": 3,
6
- "global_step": 10,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.0005608524957936063,
13
- "grad_norm": NaN,
14
- "learning_rate": 2e-05,
15
- "loss": 0.0,
16
- "step": 1
17
- },
18
- {
19
- "epoch": 0.0005608524957936063,
20
- "eval_loss": NaN,
21
- "eval_runtime": 23.6156,
22
- "eval_samples_per_second": 31.801,
23
- "eval_steps_per_second": 15.922,
24
  "step": 1
25
  },
26
  {
27
- "epoch": 0.0011217049915872126,
28
- "grad_norm": NaN,
29
- "learning_rate": 4e-05,
30
- "loss": 0.0,
31
  "step": 2
32
  },
33
  {
34
- "epoch": 0.0016825574873808188,
35
- "grad_norm": NaN,
36
- "learning_rate": 6e-05,
37
- "loss": 0.0,
38
- "step": 3
39
- },
40
- {
41
- "epoch": 0.0016825574873808188,
42
- "eval_loss": NaN,
43
- "eval_runtime": 22.074,
44
- "eval_samples_per_second": 34.022,
45
- "eval_steps_per_second": 17.034,
46
- "step": 3
47
- },
48
- {
49
- "epoch": 0.002243409983174425,
50
- "grad_norm": NaN,
51
- "learning_rate": 8e-05,
52
- "loss": 0.0,
53
- "step": 4
54
- },
55
- {
56
- "epoch": 0.0028042624789680315,
57
- "grad_norm": NaN,
58
- "learning_rate": 0.0001,
59
- "loss": 0.0,
60
- "step": 5
61
- },
62
- {
63
- "epoch": 0.0033651149747616375,
64
- "grad_norm": NaN,
65
- "learning_rate": 0.00012,
66
- "loss": 0.0,
67
- "step": 6
68
- },
69
- {
70
- "epoch": 0.0033651149747616375,
71
- "eval_loss": NaN,
72
- "eval_runtime": 22.2246,
73
- "eval_samples_per_second": 33.791,
74
- "eval_steps_per_second": 16.918,
75
- "step": 6
76
- },
77
- {
78
- "epoch": 0.003925967470555244,
79
- "grad_norm": NaN,
80
- "learning_rate": 0.00014,
81
- "loss": 0.0,
82
- "step": 7
83
- },
84
- {
85
- "epoch": 0.00448681996634885,
86
- "grad_norm": NaN,
87
- "learning_rate": 0.00016,
88
- "loss": 0.0,
89
- "step": 8
90
- },
91
- {
92
- "epoch": 0.005047672462142457,
93
- "grad_norm": NaN,
94
- "learning_rate": 0.00018,
95
- "loss": 0.0,
96
- "step": 9
97
- },
98
- {
99
- "epoch": 0.005047672462142457,
100
- "eval_loss": NaN,
101
- "eval_runtime": 22.2714,
102
- "eval_samples_per_second": 33.72,
103
- "eval_steps_per_second": 16.883,
104
- "step": 9
105
- },
106
- {
107
- "epoch": 0.005608524957936063,
108
- "grad_norm": NaN,
109
- "learning_rate": 0.0002,
110
- "loss": 0.0,
111
- "step": 10
112
  }
113
  ],
114
  "logging_steps": 1,
115
- "max_steps": 10,
116
  "num_input_tokens_seen": 0,
117
  "num_train_epochs": 1,
118
- "save_steps": 3,
119
  "stateful_callbacks": {
120
  "TrainerControl": {
121
  "args": {
@@ -128,8 +48,8 @@
128
  "attributes": {}
129
  }
130
  },
131
- "total_flos": 1725322437328896.0,
132
- "train_batch_size": 2,
133
  "trial_name": null,
134
  "trial_params": null
135
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.00013455328310010763,
5
+ "eval_steps": 500,
6
+ "global_step": 2,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 6.727664155005382e-05,
13
+ "grad_norm": 17.90272331237793,
14
+ "learning_rate": 0.0001,
15
+ "loss": 7.986,
 
 
 
 
 
 
 
 
16
  "step": 1
17
  },
18
  {
19
+ "epoch": 0.00013455328310010763,
20
+ "grad_norm": 3.7075886726379395,
21
+ "learning_rate": 0.0002,
22
+ "loss": 2.0306,
23
  "step": 2
24
  },
25
  {
26
+ "epoch": 0.00013455328310010763,
27
+ "eval_loss": 2.1211705207824707,
28
+ "eval_runtime": 5.4013,
29
+ "eval_samples_per_second": 27.956,
30
+ "eval_steps_per_second": 27.956,
31
+ "step": 2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
  }
33
  ],
34
  "logging_steps": 1,
35
+ "max_steps": 2,
36
  "num_input_tokens_seen": 0,
37
  "num_train_epochs": 1,
38
+ "save_steps": 500,
39
  "stateful_callbacks": {
40
  "TrainerControl": {
41
  "args": {
 
48
  "attributes": {}
49
  }
50
  },
51
+ "total_flos": 1598699864064.0,
52
+ "train_batch_size": 1,
53
  "trial_name": null,
54
  "trial_params": null
55
  }
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c0a0205286a7bd4fca07ce15cfa80889162bf3c590f02085703f2bd7a8a716c0
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f7cf9b3f4a9b6aee0af9bb352b3071b1f748975269b94023838baf634daee8c
3
  size 6776