Hugging Face
Models
Datasets
Spaces
Posts
Docs
Solutions
Pricing
Log In
Sign Up
DUAL-GPO
/
zephyr-7b-dpo-new-lora-v1-merged
like
0
Follow
DUAL Group
2
Safetensors
mistral
License:
apache-2.0
Model card
Files
Files and versions
Community
Train
lole25
commited on
Sep 22
Commit
564d269
•
1 Parent(s):
064883f
Create README.md
Browse files
Files changed (1)
hide
show
README.md
+5
-0
README.md
ADDED
Viewed
@@ -0,0 +1,5 @@
1
+
---
2
+
license: apache-2.0
3
+
---
4
+
5
+
This model is a fine-tuned version of Zephyr-7B using DPO on the HuggingFaceH4/ultrafeedback_binarized dataset.