--- library_name: transformers datasets: - jondurbin/truthy-dpo-v0.1 --- # MBX-7B-v3-DPO This model is a finetune of [flemmingmiguel/MBX-7B-v3](https://huggingface.co/flemmingmiguel/MBX-7B-v3) using jondurbin/truthy-dpo-v0.1 ![MBX-v3-orca](MBX-v3-orca.png) ## Code Example ```python from transformers import AutoModelForCausalLM, AutoTokenizer import torch tokenizer = AutoTokenizer.from_pretrained("macadeliccc/MBX-7B-v3-DPO") model = AutoModelForCausalLM.from_pretrained("macadeliccc/MBX-7B-v3-DPO") messages = [ {"role": "system", "content": "You are a helpful assistant."}, {"role": "user", "content": "Can you write me a creative haiku?"} ] gen_input = tokenizer.apply_chat_template(messages, return_tensors="pt") response = model.generate(**gen_input) decoded_response = tokenizer.decode(response[0], skip_special_tokens=True) print(decoded_response) ``` ## GGUF TODO ## Evaluations TODO