Minbyul commited on
Commit
dca49ad
1 Parent(s): 1df69d1

Model save

Browse files
README.md CHANGED
@@ -2,15 +2,11 @@
2
  license: apache-2.0
3
  base_model: BioMistral/BioMistral-7B
4
  tags:
5
- - alignment-handbook
6
- - trl
7
- - sft
8
- - generated_from_trainer
9
  - trl
10
  - sft
11
  - generated_from_trainer
12
  datasets:
13
- - HuggingFaceH4/deita-10k-v0-sft
14
  model-index:
15
  - name: biomistral-7b-wo-kqa_golden-iter-sft-step1
16
  results: []
@@ -21,9 +17,9 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  # biomistral-7b-wo-kqa_golden-iter-sft-step1
23
 
24
- This model is a fine-tuned version of [BioMistral/BioMistral-7B](https://huggingface.co/BioMistral/BioMistral-7B) on the HuggingFaceH4/deita-10k-v0-sft dataset.
25
  It achieves the following results on the evaluation set:
26
- - Loss: 1.4295
27
 
28
  ## Model description
29
 
@@ -60,14 +56,14 @@ The following hyperparameters were used during training:
60
 
61
  | Training Loss | Epoch | Step | Validation Loss |
62
  |:-------------:|:-----:|:----:|:---------------:|
63
- | 0.8334 | 0.92 | 9 | 1.4131 |
64
- | 0.6303 | 1.95 | 19 | 1.4060 |
65
- | 0.4647 | 2.77 | 27 | 1.4295 |
66
 
67
 
68
  ### Framework versions
69
 
70
- - Transformers 4.39.0.dev0
71
- - Pytorch 2.1.2
72
  - Datasets 2.14.6
73
  - Tokenizers 0.15.2
 
2
  license: apache-2.0
3
  base_model: BioMistral/BioMistral-7B
4
  tags:
 
 
 
 
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
  datasets:
9
+ - generator
10
  model-index:
11
  - name: biomistral-7b-wo-kqa_golden-iter-sft-step1
12
  results: []
 
17
 
18
  # biomistral-7b-wo-kqa_golden-iter-sft-step1
19
 
20
+ This model is a fine-tuned version of [BioMistral/BioMistral-7B](https://huggingface.co/BioMistral/BioMistral-7B) on the generator dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 1.8395
23
 
24
  ## Model description
25
 
 
56
 
57
  | Training Loss | Epoch | Step | Validation Loss |
58
  |:-------------:|:-----:|:----:|:---------------:|
59
+ | 1.7589 | 1.0 | 7 | 1.7180 |
60
+ | 1.4041 | 2.0 | 14 | 1.7635 |
61
+ | 0.6946 | 3.0 | 21 | 1.8395 |
62
 
63
 
64
  ### Framework versions
65
 
66
+ - Transformers 4.38.2
67
+ - Pytorch 2.1.2+cu121
68
  - Datasets 2.14.6
69
  - Tokenizers 0.15.2
all_results.json CHANGED
@@ -1,13 +1,8 @@
1
  {
2
- "epoch": 2.77,
3
- "eval_loss": 1.4294962882995605,
4
- "eval_runtime": 38.9103,
5
- "eval_samples": 4044,
6
- "eval_samples_per_second": 8.841,
7
- "eval_steps_per_second": 0.565,
8
- "train_loss": 0.6124309632513258,
9
- "train_runtime": 676.0028,
10
  "train_samples": 4750,
11
- "train_samples_per_second": 2.72,
12
- "train_steps_per_second": 0.04
13
  }
 
1
  {
2
+ "epoch": 3.0,
3
+ "train_loss": 1.1876622920944577,
4
+ "train_runtime": 427.6941,
 
 
 
 
 
5
  "train_samples": 4750,
6
+ "train_samples_per_second": 3.065,
7
+ "train_steps_per_second": 0.049
8
  }
config.json CHANGED
@@ -20,7 +20,7 @@
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
- "transformers_version": "4.39.0.dev0",
24
- "use_cache": true,
25
  "vocab_size": 32000
26
  }
 
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.38.2",
24
+ "use_cache": false,
25
  "vocab_size": 32000
26
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.39.0.dev0"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.38.2"
6
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da75fcbb729830fc5bffe2a3dde01be33068b3c5d1138b6f9e6f17fd94b3dff8
3
  size 4943162336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de60edc1d447a43d526f0a7cd8b5394d4fb25f2e8b54c99a2aaa483fbcea90f6
3
  size 4943162336
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:71e80be36b549dbfc42cbfe4fc54e3b51d0f03a4dc48a7f47646010492d2f314
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3152f1d3f5d1010b71b73360fed6f92fff5c249a609e464b53e61bb8ceeb2250
3
  size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:867985b25124dc090645cae581460d25989b30e7b2f668fc2769f6e68ded1b79
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ff759b5b480a27ee7333f8a507cf8cf48f8d382938602357de1c39a583c5847
3
  size 4540516344
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 2.77,
3
- "train_loss": 0.6124309632513258,
4
- "train_runtime": 676.0028,
5
  "train_samples": 4750,
6
- "train_samples_per_second": 2.72,
7
- "train_steps_per_second": 0.04
8
  }
 
1
  {
2
+ "epoch": 3.0,
3
+ "train_loss": 1.1876622920944577,
4
+ "train_runtime": 427.6941,
5
  "train_samples": 4750,
6
+ "train_samples_per_second": 3.065,
7
+ "train_steps_per_second": 0.049
8
  }
trainer_state.json CHANGED
@@ -1,95 +1,88 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.769230769230769,
5
  "eval_steps": 500,
6
- "global_step": 27,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.1,
13
- "grad_norm": 24.191700961175112,
14
  "learning_rate": 6.666666666666667e-06,
15
- "loss": 0.5581,
16
  "step": 1
17
  },
18
  {
19
- "epoch": 0.51,
20
- "grad_norm": 12.576350745768869,
21
- "learning_rate": 1.9659258262890683e-05,
22
- "loss": 0.8334,
23
  "step": 5
24
  },
25
  {
26
- "epoch": 0.92,
27
- "eval_loss": 1.4131343364715576,
28
- "eval_runtime": 38.4185,
29
- "eval_samples_per_second": 8.954,
30
- "eval_steps_per_second": 0.573,
31
- "step": 9
32
  },
33
  {
34
- "epoch": 1.03,
35
- "grad_norm": 17.837211332799143,
36
- "learning_rate": 1.608761429008721e-05,
37
- "loss": 0.738,
38
  "step": 10
39
  },
40
  {
41
- "epoch": 1.54,
42
- "grad_norm": 3.797877129136344,
43
- "learning_rate": 1e-05,
44
- "loss": 0.6303,
45
- "step": 15
 
46
  },
47
  {
48
- "epoch": 1.95,
49
- "eval_loss": 1.4060215950012207,
50
- "eval_runtime": 39.241,
51
- "eval_samples_per_second": 8.766,
52
- "eval_steps_per_second": 0.561,
53
- "step": 19
54
  },
55
  {
56
- "epoch": 2.05,
57
- "grad_norm": 7.485764754081964,
58
- "learning_rate": 3.912385709912794e-06,
59
- "loss": 0.5348,
60
  "step": 20
61
  },
62
  {
63
- "epoch": 2.56,
64
- "grad_norm": 3.0173543034517323,
65
- "learning_rate": 3.4074173710931804e-07,
66
- "loss": 0.4647,
67
- "step": 25
68
- },
69
- {
70
- "epoch": 2.77,
71
- "eval_loss": 1.4294962882995605,
72
- "eval_runtime": 38.6474,
73
- "eval_samples_per_second": 8.901,
74
- "eval_steps_per_second": 0.569,
75
- "step": 27
76
  },
77
  {
78
- "epoch": 2.77,
79
- "step": 27,
80
- "total_flos": 5600905789440.0,
81
- "train_loss": 0.6124309632513258,
82
- "train_runtime": 676.0028,
83
- "train_samples_per_second": 2.72,
84
- "train_steps_per_second": 0.04
85
  }
86
  ],
87
  "logging_steps": 5,
88
- "max_steps": 27,
89
  "num_input_tokens_seen": 0,
90
  "num_train_epochs": 3,
91
  "save_steps": 500,
92
- "total_flos": 5600905789440.0,
93
  "train_batch_size": 4,
94
  "trial_name": null,
95
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 3.0,
5
  "eval_steps": 500,
6
+ "global_step": 21,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.14,
13
+ "grad_norm": 10.547944577355347,
14
  "learning_rate": 6.666666666666667e-06,
15
+ "loss": 1.7384,
16
  "step": 1
17
  },
18
  {
19
+ "epoch": 0.71,
20
+ "grad_norm": 13.625247977537818,
21
+ "learning_rate": 1.9396926207859085e-05,
22
+ "loss": 1.7589,
23
  "step": 5
24
  },
25
  {
26
+ "epoch": 1.0,
27
+ "eval_loss": 1.7180463075637817,
28
+ "eval_runtime": 1.7598,
29
+ "eval_samples_per_second": 7.387,
30
+ "eval_steps_per_second": 0.568,
31
+ "step": 7
32
  },
33
  {
34
+ "epoch": 1.43,
35
+ "grad_norm": 5.744275140398176,
36
+ "learning_rate": 1.342020143325669e-05,
37
+ "loss": 1.4041,
38
  "step": 10
39
  },
40
  {
41
+ "epoch": 2.0,
42
+ "eval_loss": 1.763495683670044,
43
+ "eval_runtime": 1.6856,
44
+ "eval_samples_per_second": 7.712,
45
+ "eval_steps_per_second": 0.593,
46
+ "step": 14
47
  },
48
  {
49
+ "epoch": 2.14,
50
+ "grad_norm": 7.096816156714056,
51
+ "learning_rate": 5.000000000000003e-06,
52
+ "loss": 1.0218,
53
+ "step": 15
 
54
  },
55
  {
56
+ "epoch": 2.86,
57
+ "grad_norm": 4.77063469960506,
58
+ "learning_rate": 1.519224698779198e-07,
59
+ "loss": 0.6946,
60
  "step": 20
61
  },
62
  {
63
+ "epoch": 3.0,
64
+ "eval_loss": 1.8394839763641357,
65
+ "eval_runtime": 1.7001,
66
+ "eval_samples_per_second": 7.647,
67
+ "eval_steps_per_second": 0.588,
68
+ "step": 21
 
 
 
 
 
 
 
69
  },
70
  {
71
+ "epoch": 3.0,
72
+ "step": 21,
73
+ "total_flos": 4344627855360.0,
74
+ "train_loss": 1.1876622920944577,
75
+ "train_runtime": 427.6941,
76
+ "train_samples_per_second": 3.065,
77
+ "train_steps_per_second": 0.049
78
  }
79
  ],
80
  "logging_steps": 5,
81
+ "max_steps": 21,
82
  "num_input_tokens_seen": 0,
83
  "num_train_epochs": 3,
84
  "save_steps": 500,
85
+ "total_flos": 4344627855360.0,
86
  "train_batch_size": 4,
87
  "trial_name": null,
88
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9605dab41a6f0bcc6542eda4f5de7efa4b0c431b8dd3c26e5d4c64263bed48ba
3
  size 6200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b19306777988d00a289fd79b99cc19d76222c0c39bf75362ee3f92ab44f2161
3
  size 6200