Text Generation
Transformers
Safetensors
llama
text-generation-inference
Inference Endpoints
jonabur commited on
Commit
efee9d2
1 Parent(s): 6e93406

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -109,10 +109,10 @@ Training Checkpoints are available as branches in the repository. Checkpoints w
109
  The transformers library allows you to load a checkpoint from a branch as follows:
110
 
111
  ```python
112
- branch = "200B"
113
  model = transformers.AutoModelForCausalLM.from_pretrained(
114
  "LumiOpen/Viking-7B",
115
- torch_dtype=torch.bfloat16,
116
  revision=branch,
117
  )
118
  ```
 
109
  The transformers library allows you to load a checkpoint from a branch as follows:
110
 
111
  ```python
112
+ branch = "2000B"
113
  model = transformers.AutoModelForCausalLM.from_pretrained(
114
  "LumiOpen/Viking-7B",
115
+ torch_dtype=torch.bfloat16 if torch.cuda.is_bf16_supported() else torch.float16,
116
  revision=branch,
117
  )
118
  ```