MBX-7B-v3-DPO / README.md
macadeliccc's picture
Update README.md
88575bb verified
|
raw
history blame
914 Bytes
metadata
library_name: transformers
datasets:
  - jondurbin/truthy-dpo-v0.1

MBX-7B-v3-DPO

This model is a finetune of flemmingmiguel/MBX-7B-v3 using jondurbin/truthy-dpo-v0.1

MBX-v3-orca

Code Example

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

tokenizer = AutoTokenizer.from_pretrained("macadeliccc/MBX-7B-v3-DPO")
model = AutoModelForCausalLM.from_pretrained("macadeliccc/MBX-7B-v3-DPO")

messages = [
    {"role": "system", "content": "You are a helpful assistant."},
    {"role": "user", "content": "Can you write me a creative haiku?"}
]
gen_input = tokenizer.apply_chat_template(messages, return_tensors="pt")

response = model.generate(**gen_input)
decoded_response = tokenizer.decode(response[0], skip_special_tokens=True)

print(decoded_response)

GGUF

TODO

Evaluations

TODO