salmanshahid commited on
Commit
a5b53b4
1 Parent(s): 22303a9

Upload output/training_config.yml with huggingface_hub

Browse files
Files changed (1) hide show
  1. output/training_config.yml +70 -0
output/training_config.yml ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ _component_: models.lora_mmllama3_8b
3
+ lora_attn_modules:
4
+ - q_proj
5
+ - v_proj
6
+ apply_lora_to_mlp: false
7
+ apply_lora_to_output: false
8
+ lora_rank: 8
9
+ lora_alpha: 16
10
+ perception_tokens: 2
11
+ tokenizer:
12
+ _component_: models.a2a_tokenizer
13
+ path: checkpoints/Meta-Llama-3-8B-Instruct/tokenizer.model
14
+ checkpointer:
15
+ _component_: torchtune.utils.FullModelMetaCheckpointer
16
+ checkpoint_dir: checkpoints/Meta-Llama-3-8B-Instruct/
17
+ checkpoint_files:
18
+ - consolidated.00.pth
19
+ adapter_checkpoint: null
20
+ recipe_checkpoint: null
21
+ output_dir: output
22
+ model_type: LLAMA3
23
+ resume_from_checkpoint: false
24
+ dataset:
25
+ _component_: ds.EvenBatcher
26
+ dataset:
27
+ _component_: ds.RoundRobinDataset
28
+ datasets:
29
+ - _component_: ds.LlavaInstructDataset
30
+ ib_embed_path: ds/coco_llava_instruct/*.ib_embed.pt
31
+ clip_embed_path: ds/coco_llava_instruct/*.clip_embed.pt
32
+ caption_path: ds/coco_llava_instruct/*.caption.pt
33
+ train_on_input: false
34
+ - _component_: ds.LlavaInstructDataset
35
+ ib_embed_path: ds/vision_flan/*.ib_embed.pt
36
+ clip_embed_path: ds/vision_flan/*.clip_embed.pt
37
+ caption_path: ds/vision_flan/*.caption.pt
38
+ train_on_input: false
39
+ - _component_: ds.CaptionInstructDataset
40
+ ib_embed_path: ds/sam_llava/00.ib_embed.pt
41
+ clip_embed_path: ds/sam_llava/00.clip_embed.pt
42
+ caption_path: ds/sam_llava/00.caption.pt
43
+ train_on_input: false
44
+ seed: null
45
+ shuffle: true
46
+ batch_size: 4
47
+ optimizer:
48
+ _component_: torch.optim.AdamW
49
+ weight_decay: 0.01
50
+ lr: 0.0003
51
+ lr_scheduler:
52
+ _component_: torchtune.modules.get_cosine_schedule_with_warmup
53
+ num_warmup_steps: 100
54
+ loss:
55
+ _component_: torch.nn.CrossEntropyLoss
56
+ epochs: 1
57
+ max_steps_per_epoch: null
58
+ gradient_accumulation_steps: 64
59
+ compile: false
60
+ output_dir: /tmp/lora_finetune_output
61
+ metric_logger:
62
+ _component_: torchtune.utils.metric_logging.DiskLogger
63
+ log_dir: ${output_dir}
64
+ log_every_n_steps: null
65
+ device: cuda
66
+ dtype: bf16
67
+ enable_activation_checkpointing: false
68
+ profiler:
69
+ _component_: torchtune.utils.profiler
70
+ enabled: false