metadata
library_name: transformers
datasets:
- jondurbin/truthy-dpo-v0.1
MBX-7B-v3-DPO
This model is a finetune of flemmingmiguel/MBX-7B-v3 using jondurbin/truthy-dpo-v0.1
Code Example
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
tokenizer = AutoTokenizer.from_pretrained("macadeliccc/MBX-7B-v3-DPO")
model = AutoModelForCausalLM.from_pretrained("macadeliccc/MBX-7B-v3-DPO")
messages = [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Can you write me a creative haiku?"}
]
gen_input = tokenizer.apply_chat_template(messages, return_tensors="pt")
response = model.generate(**gen_input)
decoded_response = tokenizer.decode(response[0], skip_special_tokens=True)
print(decoded_response)
GGUF
TODO
Evaluations
TODO