File size: 458 Bytes
d44d431
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
{
  "base_model_name": "Llama-2-13B-fp16-padded",
  "base_model_class": "LlamaForCausalLM",
  "base_loaded_in_4bit": true,
  "base_loaded_in_8bit": false,
  "projections": "q, v",
  "loss": 1.07,
  "learning_rate": 1.7080653142008908e-05,
  "epoch": 0.74,
  "current_steps": 9678,
  "train_runtime": 20262.2951,
  "train_samples_per_second": 12.817,
  "train_steps_per_second": 0.1,
  "total_flos": 4.119249176808653e+17,
  "train_loss": 1.1322209582423532
}