Maelstrome commited on
Commit
5202b39
1 Parent(s): 577d99b

End of training

Browse files
README.md CHANGED
@@ -13,23 +13,24 @@ model-index:
13
  results: []
14
  ---
15
 
 
 
 
16
  # gemma-2b-storytelling
17
 
18
  This model is a fine-tuned version of [google/gemma-2b](https://huggingface.co/google/gemma-2b) on the generator dataset.
19
- It achieves the following results on the evaluation set:
20
- - Loss: nan
21
 
22
  ## Model description
23
 
24
- This model has been fine-tuned specifically for the task of text generation, focusing on various storytelling themes. It utilizes advanced language modeling techniques to produce coherent and contextually relevant narratives based on user prompts.
25
 
26
  ## Intended uses & limitations
27
 
28
- This model is intended for use in applications requiring high-quality narrative text generation, such as content creation, interactive storytelling, or game design. Users should be aware of potential limitations in the model's understanding of complex contexts or subtleties in language, which may affect the output quality.
29
 
30
  ## Training and evaluation data
31
 
32
- The model was trained using the `PocketDoc/RUCAIBox-Story-Generation-Alpaca` dataset, which contains diverse storytelling prompts and responses, ensuring a robust ability to generate varied narrative content.
33
 
34
  ## Training procedure
35
 
@@ -42,21 +43,15 @@ The following hyperparameters were used during training:
42
  - seed: 42
43
  - gradient_accumulation_steps: 8
44
  - total_train_batch_size: 32
45
- - optimizer: Adam with betas=(0.9, 0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_ratio: 0.05
48
  - training_steps: 154
49
 
50
- ### Training results
51
-
52
- | Training Loss | Epoch | Step | Validation Loss |
53
- |:----------------:|:------:|:----:|:---------------:|
54
- | 1454737970954.24 | 0.9164 | 100 | nan |
55
-
56
  ### Framework versions
57
 
58
  - PEFT 0.10.0
59
  - Transformers 4.40.1
60
  - Pytorch 2.2.2+cu121
61
  - Datasets 2.19.0
62
- - Tokenizers 0.19.1
 
13
  results: []
14
  ---
15
 
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
  # gemma-2b-storytelling
20
 
21
  This model is a fine-tuned version of [google/gemma-2b](https://huggingface.co/google/gemma-2b) on the generator dataset.
 
 
22
 
23
  ## Model description
24
 
25
+ More information needed
26
 
27
  ## Intended uses & limitations
28
 
29
+ More information needed
30
 
31
  ## Training and evaluation data
32
 
33
+ More information needed
34
 
35
  ## Training procedure
36
 
 
43
  - seed: 42
44
  - gradient_accumulation_steps: 8
45
  - total_train_batch_size: 32
46
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
  - lr_scheduler_warmup_ratio: 0.05
49
  - training_steps: 154
50
 
 
 
 
 
 
 
51
  ### Framework versions
52
 
53
  - PEFT 0.10.0
54
  - Transformers 4.40.1
55
  - Pytorch 2.2.2+cu121
56
  - Datasets 2.19.0
57
+ - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "v_proj",
24
- "q_proj",
25
- "down_proj",
26
  "up_proj",
27
- "k_proj",
28
- "o_proj",
29
- "gate_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "gate_proj",
24
+ "o_proj",
25
  "v_proj",
 
 
26
  "up_proj",
27
+ "down_proj",
28
+ "q_proj",
29
+ "k_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d54182c3d91897e6633fa14116f4381c64fdfa26ccac18e902cc979f1fd30946
3
  size 156926880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd8288e0c2887913ec48ae015aaf5524afb0102d4347650cbd097232b8250fcc
3
  size 156926880
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd7ce1c110ddfa0d013b2a7ebf59c8f67be1b117ca955599b1a6362c8e19d33f
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fe5c200ad09b1c5099fc17a8d8e79cf6638956c5bd4e9a7c17b62099f41039e
3
  size 4984