Upload folder using huggingface_hub

#4
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/tmp/tmpu4qxllsm",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -11,7 +11,7 @@
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 14336,
14
- "max_position_embeddings": 8192,
15
  "model_type": "llama",
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
@@ -27,7 +27,7 @@
27
  },
28
  "rms_norm_eps": 1e-05,
29
  "rope_scaling": null,
30
- "rope_theta": 500000.0,
31
  "tie_word_embeddings": false,
32
  "torch_dtype": "float16",
33
  "transformers_version": "4.40.0",
 
1
  {
2
+ "_name_or_path": "/tmp/tmpetnsv95z",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 14336,
14
+ "max_position_embeddings": 524288,
15
  "model_type": "llama",
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
 
27
  },
28
  "rms_norm_eps": 1e-05,
29
  "rope_scaling": null,
30
+ "rope_theta": 2000000.0,
31
  "tie_word_embeddings": false,
32
  "torch_dtype": "float16",
33
  "transformers_version": "4.40.0",
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0042d699577ceedaa1a41e7fd771030684da8f5ee475b025a5590c20869cafe5
3
  size 4677281680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55137d8fe28ef15407cb4fe5a31717527894c8b0860d9283a34768b060722b47
3
  size 4677281680
plots.png CHANGED
results.json CHANGED
@@ -1,30 +1,32 @@
1
  {
2
  "base_current_gpu_type": "NVIDIA A100-PCIE-40GB",
3
  "base_current_gpu_total_memory": 40339.3125,
4
- "base_token_generation_latency_sync": 54.04956970214844,
5
- "base_token_generation_latency_async": 54.09220550209284,
6
- "base_token_generation_throughput_sync": 0.01850153489677552,
7
- "base_token_generation_throughput_async": 0.018486951876297775,
 
8
  "base_token_generation_CO2_emissions": null,
9
  "base_token_generation_energy_consumption": null,
10
- "base_inference_latency_sync": 53.23417663574219,
11
- "base_inference_latency_async": 52.190279960632324,
12
- "base_inference_throughput_sync": 0.018784924708097873,
13
- "base_inference_throughput_async": 0.01916065598334231,
14
  "base_inference_CO2_emissions": null,
15
  "base_inference_energy_consumption": null,
16
  "smashed_current_gpu_type": "NVIDIA A100-PCIE-40GB",
17
  "smashed_current_gpu_total_memory": 40339.3125,
18
- "smashed_token_generation_latency_sync": 40.57583160400391,
19
- "smashed_token_generation_latency_async": 40.6801862642169,
20
- "smashed_token_generation_throughput_sync": 0.024645212691126284,
21
- "smashed_token_generation_throughput_async": 0.024581991672924564,
 
22
  "smashed_token_generation_CO2_emissions": null,
23
  "smashed_token_generation_energy_consumption": null,
24
- "smashed_inference_latency_sync": 51.8927360534668,
25
- "smashed_inference_latency_async": 41.21739864349365,
26
- "smashed_inference_throughput_sync": 0.019270519846355125,
27
- "smashed_inference_throughput_async": 0.02426159905552056,
28
  "smashed_inference_CO2_emissions": null,
29
  "smashed_inference_energy_consumption": null
30
  }
 
1
  {
2
  "base_current_gpu_type": "NVIDIA A100-PCIE-40GB",
3
  "base_current_gpu_total_memory": 40339.3125,
4
+ "base_perplexity": 13.74486255645752,
5
+ "base_token_generation_latency_sync": 37.35860481262207,
6
+ "base_token_generation_latency_async": 37.739571928977966,
7
+ "base_token_generation_throughput_sync": 0.026767594909276098,
8
+ "base_token_generation_throughput_async": 0.026497385870775065,
9
  "base_token_generation_CO2_emissions": null,
10
  "base_token_generation_energy_consumption": null,
11
+ "base_inference_latency_sync": 122.0685821533203,
12
+ "base_inference_latency_async": 43.52927207946777,
13
+ "base_inference_throughput_sync": 0.008192116123246046,
14
+ "base_inference_throughput_async": 0.022973046693139806,
15
  "base_inference_CO2_emissions": null,
16
  "base_inference_energy_consumption": null,
17
  "smashed_current_gpu_type": "NVIDIA A100-PCIE-40GB",
18
  "smashed_current_gpu_total_memory": 40339.3125,
19
+ "smashed_perplexity": 14.672628402709961,
20
+ "smashed_token_generation_latency_sync": 43.5498722076416,
21
+ "smashed_token_generation_latency_async": 43.305750377476215,
22
+ "smashed_token_generation_throughput_sync": 0.02296217989417044,
23
+ "smashed_token_generation_throughput_async": 0.02309162158104783,
24
  "smashed_token_generation_CO2_emissions": null,
25
  "smashed_token_generation_energy_consumption": null,
26
+ "smashed_inference_latency_sync": 164.25011291503907,
27
+ "smashed_inference_latency_async": 69.53554153442383,
28
+ "smashed_inference_throughput_sync": 0.006088275875446525,
29
+ "smashed_inference_throughput_async": 0.014381134854683576,
30
  "smashed_inference_CO2_emissions": null,
31
  "smashed_inference_energy_consumption": null
32
  }
smash_config.json CHANGED
@@ -14,7 +14,7 @@
14
  "controlnet": "None",
15
  "unet_dim": 4,
16
  "device": "cuda",
17
- "cache_dir": "/ceph/hdd/staff/charpent/.cache/modelsa_klkits",
18
  "batch_size": 1,
19
  "model_name": "cognitivecomputations/dolphin-2.9-llama3-8b-256k",
20
  "task": "text_text_generation",
 
14
  "controlnet": "None",
15
  "unet_dim": 4,
16
  "device": "cuda",
17
+ "cache_dir": "/ceph/hdd/staff/charpent/.cache/modelsgmyikxqu",
18
  "batch_size": 1,
19
  "model_name": "cognitivecomputations/dolphin-2.9-llama3-8b-256k",
20
  "task": "text_text_generation",
tokenizer_config.json CHANGED
@@ -2069,6 +2069,7 @@
2069
  "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
2070
  "clean_up_tokenization_spaces": true,
2071
  "eos_token": "<|im_end|>",
 
2072
  "model_input_names": [
2073
  "input_ids",
2074
  "attention_mask"
 
2069
  "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
2070
  "clean_up_tokenization_spaces": true,
2071
  "eos_token": "<|im_end|>",
2072
+ "legacy": false,
2073
  "model_input_names": [
2074
  "input_ids",
2075
  "attention_mask"