Heralax commited on
Commit
2eaec3e
1 Parent(s): 3a56d37

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +27 -101
README.md CHANGED
@@ -9,108 +9,34 @@ model-index:
9
  results: []
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
14
 
15
- [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
16
- <details><summary>See axolotl config</summary>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
 
18
- axolotl version: `0.4.1`
19
- ```yaml
20
- base_model: Heralax/army-pretrain-1
21
- tokenizer_type: AutoTokenizer
22
- is_mistral_derived_model: true
23
- load_in_8bit: false
24
- load_in_4bit: false
25
- strict: false
26
-
27
- datasets:
28
- - path: json
29
- data_files: us_army_plain_qa_list_open.jsonl
30
- ds_type: json
31
- type: sharegpt
32
- conversation: chatml
33
- - path: json
34
- data_files: us_army_plain_qa_list_vanilla.jsonl
35
- ds_type: json
36
- type: sharegpt
37
- conversation: chatml
38
- - path: json
39
- data_files: us_army_plain_qa_list_negative.jsonl
40
- ds_type: json
41
- type: sharegpt
42
- conversation: chatml
43
-
44
- dataset_prepared_path: last_run_prepared
45
- output_dir: ./us-army-finetune-1
46
-
47
- sequence_len: 4096
48
- sample_packing: true
49
- pad_to_sequence_len: true
50
- shuffle_merged_datasets: true
51
-
52
- wandb_project: mistral-usarmy
53
- wandb_entity:
54
- wandb_watch:
55
- wandb_run_id:
56
- wandb_log_model:
57
-
58
- gradient_accumulation_steps: 6
59
- micro_batch_size: 2
60
- eval_batch_size: 1
61
- num_epochs: 6
62
- optimizer: paged_adamw_8bit
63
- lr_scheduler: cosine
64
- learning_rate: 0.000020
65
- weight_decay: 0
66
- # Gradient clipping max norm
67
- max_grad_norm: 1.0
68
- noisy_embedding_alpha: 0
69
- train_on_inputs: false
70
- group_by_length: false
71
- bf16: true
72
- fp16: false
73
- tf32: false
74
-
75
- gradient_checkpointing: unsloth
76
- early_stopping_patience:
77
- resume_from_checkpoint:
78
- logging_steps: 1
79
- xformers_attention:
80
- flash_attention: true
81
-
82
- chat_template: chatml
83
-
84
- warmup_ratio: 0.5
85
- auto_resume_from_checkpoints: false
86
- #warmup_ratio: 0.5
87
- eval_steps: 10
88
- saves_per_epoch: 1
89
- eval_sample_packing: false
90
- save_total_limit: 3
91
- debug:
92
- deepspeed: deepspeed_configs/zero2.json
93
- special_tokens:
94
- pad_token: "<|end_of_text|>"
95
- ```
96
-
97
- </details><br>
98
-
99
- # us-army-finetune-1
100
-
101
- This model is a fine-tuned version of [Heralax/army-pretrain-1](https://huggingface.co/Heralax/army-pretrain-1) on the None dataset.
102
-
103
- ## Model description
104
-
105
- More information needed
106
-
107
- ## Intended uses & limitations
108
-
109
- More information needed
110
-
111
- ## Training and evaluation data
112
-
113
- More information needed
114
 
115
  ## Training procedure
116
 
@@ -133,7 +59,7 @@ The following hyperparameters were used during training:
133
 
134
  ### Training results
135
 
136
-
137
 
138
  ### Framework versions
139
 
 
9
  results: []
10
  ---
11
 
12
+ Was torn between calling it MiLLM and Mistrillitary. *Sigh* naming is one of the two great problems in computer science...
 
13
 
14
+ This is a domain-expert finetune based on the US Army field manuals (the ones that are published and available for civvies like me). It's focused on factual question answer only, but seems to be able to answer slightly deeper questions in a pinch.
15
+
16
+ ## Model Quirks
17
+
18
+ - I had to focus on the army field manuals because the armed forces publishes a truly massive amount of text.
19
+ - No generalist assistant data was included, which means this is very very very focused on QA, and may be inflexible.
20
+ - Experimental change: data was mostly generated by a smaller model, Mistral NeMo. Quality seems unaffected, costs are much lower. Had problems with the open-ended questions not being in the right format.
21
+ - Low temperture recommended. Screenshots use 0.
22
+ - ChatML
23
+ - No special tokens added.
24
+
25
+ Examples:
26
+
27
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64825ebceb4befee377cf8ac/KakWvjSMwSHkISPGoB0RH.png))
28
+
29
+
30
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64825ebceb4befee377cf8ac/7rlJxcjGECqFuEFmYC3aV.png)
31
 
32
+
33
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64825ebceb4befee377cf8ac/mzxk9Qa9cveFx7PArnAmB.png)
34
+
35
+
36
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64825ebceb4befee377cf8ac/2KtpGhqReVPj4Wh3fles5.png)
37
+
38
+
39
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64825ebceb4befee377cf8ac/Pz70D922utg5ZZCqYiGpT.png)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
 
41
  ## Training procedure
42
 
 
59
 
60
  ### Training results
61
 
62
+ It answers questions alright.
63
 
64
  ### Framework versions
65