File size: 3,139 Bytes
597ae55
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
{
  "accelerator_kwargs": {},
  "adap_kl_ctrl": true,
  "backward_batch_size": 64,
  "batch_size": 64,
  "cliprange": 0.2,
  "cliprange_value": 0.2,
  "compare_steps": 1,
  "dataset_num_proc": null,
  "early_stopping": false,
  "exp_name": "stego_trainer",
  "forward_batch_size": null,
  "gamma": 1,
  "global_backward_batch_size": 64,
  "global_batch_size": 64,
  "gradient_accumulation_steps": 4,
  "gradient_checkpointing": false,
  "horizon": 10000,
  "init_kl_coef": 0.05,
  "is_encoder_decoder": false,
  "is_peft_model": true,
  "kl_penalty": "kl",
  "lam": 0.95,
  "learning_rate": 2e-05,
  "log_with": "wandb",
  "max_grad_norm": null,
  "mini_batch_size": 16,
  "model_name": "unsloth/gemma-2-2b-it",
  "optimize_cuda_cache": true,
  "optimize_device_cache": false,
  "ppo_epochs": 4,
  "project_kwargs": {},
  "push_to_hub_if_best_kwargs": {},
  "query_dataset": "imdb",
  "ratio_threshold": 10.0,
  "remove_unused_columns": true,
  "reward_model": "sentiment-analysis:lvwerra/distilbert-imdb",
  "score_clip": null,
  "seed": 0,
  "steps": 20000,
  "target": 12.0,
  "target_kl": 1,
  "task_name": null,
  "tracker_kwargs": {
    "wandb": {
      "name": "cv_gemma-2-2b-it_to_distilbert-base-uncased_EBS64_Joan",
      "notes": "Dataset: cv\n    Same Prompt:  \n    Payload Prefixes: ['Movie Review: This movie was really amazing!', 'Movie Review: This movie was really terrible!']\n    Payload Template: Movie Review: This movie was really {payload}!\n    Separate Enc/Dec Data: True\n\n    Encoder: gemma-2-2b-it (LR: 2e-05)\n    Decoder: distilbert-base-uncased (LR: 1e-05)\n    Train Loop: v2_dylan\n\n    Effective Batch Sizes:\n    - Encoder: 64\n    - Decoder: 512\n\n    Training Iterations:\n    - Encoder updates: 100\n    - Decoder updates: 400\n    - Update Encoder First: False\n\n    Temperatures:\n    - Decoder Training: 1.0\n    - Encoder Training: 1.0\n    - Evaluation: 1.0\n\n    Encoder Parameters:\n    - KL Coefficient: 0.05\n    - LoRA: True\n    - Quantization: False\n    - Output Length: {'min': 42, 'max': 51}\n\n    Decoder Parameters:\n    - New Classification Head: True\n    - Use Probs Reward: False\n    - Weight Decay: 0.01\n    - Update Parameters: {'head': True, 'body': True}\n\n    Training Configuration:\n    - Update Encoder: True\n    - Update Decoder: True\n    - Paraphrase: False\n    - Leak Password: False\n    - WandB Logging: True\n    - Eval Every N: 50\n    - Number of Epochs: 100000\n\n    Debug:\n    - Override Dec Batch: False",
      "tags": [
        "cv",
        "gemma-2-2b-it",
        "distilbert-base-uncased",
        "v2_dylan",
        "enc_lr_2e-05",
        "dec_lr_1e-05",
        "enc_eff_bs_64",
        "dec_eff_bs_512",
        "enc_updates_100",
        "dec_updates_400",
        "LoRA",
        "Full_Precision",
        "same_prompt_ ",
        "Separate_Enc_Dec_Data",
        "Update_Enc",
        "Update_Dec",
        "No_Paraphrase",
        "No_Leak",
        "1-bit"
      ]
    }
  },
  "tracker_project_name": "trl",
  "use_score_norm": false,
  "use_score_scaling": false,
  "vf_coef": 0.1,
  "whiten_rewards": false,
  "world_size": 1
}