Hzfinfdu commited on
Commit
b6113db
1 Parent(s): f62820f

Rename folders

Browse files
Llama3_1Base-L5TC-8x/checkpoints/final.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a43b60771fe6dc71414daf8b28a5d5e73359e52f2be7d1420924f7110e3dcd6
3
- size 536945024
 
 
 
 
Llama3_1Base-L5TC-8x/hyperparams.json DELETED
@@ -1,35 +0,0 @@
1
- {
2
- "device": "cuda:0",
3
- "seed": 42,
4
- "dtype": "torch.bfloat16",
5
- "hook_point_in": "blocks.5.ln2.hook_normalized",
6
- "hook_point_out": "blocks.5.hook_mlp_out",
7
- "use_decoder_bias": true,
8
- "apply_decoder_bias_to_pre_encoder": false,
9
- "expansion_factor": 8,
10
- "d_model": 4096,
11
- "d_sae": 32768,
12
- "bias_init_method": "all_zero",
13
- "act_fn": "jumprelu",
14
- "jump_relu_threshold": 0.0732421875,
15
- "norm_activation": "dataset-wise",
16
- "dataset_average_activation_norm": {
17
- "in": 64.0,
18
- "out": 2.765625
19
- },
20
- "decoder_exactly_fixed_norm": false,
21
- "sparsity_include_decoder_norm": true,
22
- "use_glu_encoder": false,
23
- "init_decoder_norm": 0.5,
24
- "init_encoder_norm": 0.5,
25
- "init_encoder_with_decoder_transpose": false,
26
- "lp": 1,
27
- "l1_coefficient": 8e-05,
28
- "l1_coefficient_warmup_steps": 39062,
29
- "top_k": 50,
30
- "k_warmup_steps": 39062,
31
- "use_batch_norm_mse": true,
32
- "use_ghost_grads": false,
33
- "tp_size": 1,
34
- "ddp_size": 1
35
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Llama3_1Base-L5TC-8x/lm_config.json DELETED
@@ -1,10 +0,0 @@
1
- {
2
- "device": "cuda",
3
- "seed": 42,
4
- "dtype": "torch.bfloat16",
5
- "model_name": "meta-llama/Meta-Llama-3.1-8B",
6
- "use_flash_attn": false,
7
- "cache_dir": null,
8
- "d_model": 4096,
9
- "local_files_only": false
10
- }