Text Generation
Transformers
Safetensors
zamba2
Inference Endpoints
pglo commited on
Commit
3f5315e
1 Parent(s): 6b57bf2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -42,7 +42,7 @@ import torch
42
  tokenizer = AutoTokenizer.from_pretrained("Zyphra/Zamba2-2.7B")
43
  model = AutoModelForCausalLM.from_pretrained("Zyphra/Zamba2-2.7B", device_map="cuda", torch_dtype=torch.bfloat16)
44
 
45
- input_text = "A funny prompt would be "
46
  input_ids = tokenizer(input_text, return_tensors="pt").to("cuda")
47
 
48
  outputs = model.generate(**input_ids, max_new_tokens=100)
 
42
  tokenizer = AutoTokenizer.from_pretrained("Zyphra/Zamba2-2.7B")
43
  model = AutoModelForCausalLM.from_pretrained("Zyphra/Zamba2-2.7B", device_map="cuda", torch_dtype=torch.bfloat16)
44
 
45
+ input_text = "What factors contributed to the fall of the Roman Empire?"
46
  input_ids = tokenizer(input_text, return_tensors="pt").to("cuda")
47
 
48
  outputs = model.generate(**input_ids, max_new_tokens=100)