5cp commited on
Commit
b7704e4
1 Parent(s): 2701c0e

Synchronizing local compiler cache.

Browse files
.gitattributes CHANGED
@@ -5703,3 +5703,5 @@ neuronxcc-2.15.128.0+56dc5a86/MODULE_4765382095544640324+0662544f/model.neff fil
5703
  neuronxcc-2.15.128.0+56dc5a86/MODULE_4825501270026396480+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
5704
  neuronxcc-2.15.128.0+56dc5a86/MODULE_5352243160435722256+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
5705
  neuronxcc-2.15.128.0+56dc5a86/MODULE_9004138607823884889+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
 
 
 
5703
  neuronxcc-2.15.128.0+56dc5a86/MODULE_4825501270026396480+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
5704
  neuronxcc-2.15.128.0+56dc5a86/MODULE_5352243160435722256+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
5705
  neuronxcc-2.15.128.0+56dc5a86/MODULE_9004138607823884889+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
5706
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_355980366824992736+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
5707
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_6220973826920381074+0662544f/model.neff filter=lfs diff=lfs merge=lfs -text
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25/training/llama/meta-llama/Meta-Llama-3-8B/3d277a0c99a976229120.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128001, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 8192, "mlp_bias": false, "model_type": "llama", "neuron": {"compiler_version": "2.15.128.0+56dc5a86", "input_specs": {"attention_mask": [4, 1024], "input_ids": [4, 1024], "labels": [4, 1024]}, "model_class": "PeftModelForCausalLM", "num_neuron_cores_per_node": 2, "pipeline_parallel_size": 1, "precision": "bfloat16", "tensor_parallel_size": 2, "training": true}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
neuronxcc-2.15.128.0+56dc5a86/MODULE_355980366824992736+0662544f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--model-type=transformer"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_355980366824992736+0662544f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_355980366824992736+0662544f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a49a47f8e9b2af9ab4e85cdb8a535d195445721cb61b7b1a87f1935b1245a862
3
+ size 3276014
neuronxcc-2.15.128.0+56dc5a86/MODULE_355980366824992736+0662544f/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af3ce18c2f30af576c121eca80e5fb605fa1376c790ac554532a1d3a1d09f808
3
+ size 53720064
neuronxcc-2.15.128.0+56dc5a86/MODULE_6220973826920381074+0662544f/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--model-type=transformer", "--model-type=transformer"]
neuronxcc-2.15.128.0+56dc5a86/MODULE_6220973826920381074+0662544f/model.done ADDED
File without changes
neuronxcc-2.15.128.0+56dc5a86/MODULE_6220973826920381074+0662544f/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:928a43a6554541427c5aa68405f8e39aa56b4bd1f8cf7044c34245debb54d8d0
3
+ size 3275630
neuronxcc-2.15.128.0+56dc5a86/MODULE_6220973826920381074+0662544f/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:190b9f83fea2d1e8a8f432d6cb591c3474e127fd7baa3023f0cf32375c67b703
3
+ size 53699584