File size: 6,400 Bytes
f08255e 379db0e f08255e 379db0e f08255e 8367fb3 f08255e 8367fb3 f08255e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 |
---
base_model: SanjiWatsuki/Kunoichi-DPO-v2-7B
inference: false
language:
- en
license: cc-by-nc-4.0
model_creator: SanjiWatsuki
model_name: Kunoichi-DPO-v2-7B
model_type: mistral
pipeline_tag: text-generation
prompt_template: "{{system_message}}
### Instruction:
{{prompt}}
### Response:
"
quantized_by: brittlewis12
---
# Kunoichi-DPO-v2-7B GGUF
![Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B/resolve/main/assets/kunoichi.png)
Original model: [Kunoichi-DPO-v2-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-DPO-v2-7B)
Model creator: [SanjiWatsuki](https://huggingface.co/SanjiWatsuki)
This repo contains GGUF format model files for SanjiWatsuki’s Kunoichi-DPO-v2-7B. Updated as of 2024-05-01.
### What is GGUF?
GGUF is a file format for representing AI models. It is the third version of the format, introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp.
Converted using llama.cpp build 2780 (revision [b0d943de](https://github.com/ggerganov/llama.cpp/commit/b0d943de))
### Prompt template: Unknown (Alpaca)
[Alpaca-style](https://huggingface.co/SanjiWatsuki/Kunoichi-7B#prompt-template-custom-format-or-alpaca) was the prompt format for the original [Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B).
```
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{{prompt}}
### Response:
```
---
## Download & run with [cnvrs](https://twitter.com/cnvrsai) on iPhone, iPad, and Mac!
![cnvrs.ai](https://pbs.twimg.com/profile_images/1744049151241797632/0mIP-P9e_400x400.jpg)
[cnvrs](https://testflight.apple.com/join/sFWReS7K) is the best app for private, local AI on your device:
- create & save **Characters** with custom system prompts & temperature settings
- download and experiment with any **GGUF model** you can [find on HuggingFace](https://huggingface.co/models?library=gguf)!
- make it your own with custom **Theme colors**
- powered by Metal ⚡️ & [Llama.cpp](https://github.com/ggerganov/llama.cpp), with **haptics** during response streaming!
- **try it out** yourself today, on [Testflight](https://testflight.apple.com/join/sFWReS7K)!
- follow [cnvrs on twitter](https://twitter.com/cnvrsai) to stay up to date
---
## Original Model Evaluations:
| Model | MT Bench | EQ Bench | MMLU | Logic Test |
|----------------------|----------|----------|---------|-------------|
| GPT-4-Turbo | 9.32 | - | - | - |
| GPT-4 | 8.99 | 62.52 | 86.4 | 0.86 |
| **Kunoichi-DPO-v2-7B** | **8.51** | **42.18** | - | **0.58** |
| Mixtral-8x7B-Instruct| 8.30 | 44.81 | 70.6 | 0.75 |
| **Kunoichi-DPO-7B** | **8.29** | **41.60** | **64.83** | **0.59** |
| **Kunoichi-7B** | **8.14** | **44.32** | **64.9** | **0.58** |
| Starling-7B | 8.09 | - | 63.9 | 0.51 |
| Claude-2 | 8.06 | 52.14 | 78.5 | - |
| Silicon-Maid-7B | 7.96 | 40.44 | 64.7 | 0.54 |
| Loyal-Macaroni-Maid-7B | 7.95 | 38.66 | 64.9 | 0.57 |
| GPT-3.5-Turbo | 7.94 | 50.28 | 70 | 0.57 |
| Claude-1 | 7.9 | - | 77 | - |
| Openchat-3.5 | 7.81 | 37.08 | 64.3 | 0.39 |
| Dolphin-2.6-DPO | 7.74 | 42.88 | 61.9 | 0.53 |
| Zephyr-7B-beta | 7.34 | 38.71 | 61.4 | 0.30 |
| Llama-2-70b-chat-hf | 6.86 | 51.56 | 63 | - |
| Neural-chat-7b-v3-1 | 6.84 | 43.61 | 62.4 | 0.30 |
| Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
|---|---:|---:|---:|---:|---:|
| **Kunoichi-DPO-7B**|**58.4**| 45.08 | 74| 66.99| 47.52|
| **Kunoichi-DPO-v2-7B**|**58.31**| 44.85| 75.05| 65.69| 47.65|
| [Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B)|57.54| 44.99| 74.86| 63.72| 46.58|
| [OpenPipe/mistral-ft-optimized-1218](https://huggingface.co/OpenPipe/mistral-ft-optimized-1218)| 56.85 | 44.74 | 75.6 | 59.89 | 47.17 |
| [Silicon-Maid-7B](https://huggingface.co/SanjiWatsuki/Silicon-Maid-7B) | 56.45| 44.74| 74.26| 61.5| 45.32|
| [mlabonne/NeuralHermes-2.5-Mistral-7B](https://huggingface.co/mlabonne/NeuralHermes-2.5-Mistral-7B) | 53.51 | 43.67 | 73.24 | 55.37 | 41.76 |
| [teknium/OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B) | 52.42 | 42.75 | 72.99 | 52.99 | 40.94 |
| [openchat/openchat_3.5](https://huggingface.co/openchat/openchat_3.5) | 51.34 | 42.67 | 72.92 | 47.27 | 42.51 |
| [berkeley-nest/Starling-LM-7B-alpha](https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha) | 51.16 | 42.06 | 72.72 | 47.33 | 42.53 |
| [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) | 50.99 | 37.33 | 71.83 | 55.1 | 39.7 |
| Model | AlpacaEval2 | Length |
| --------------------------- | ----------- | ------ |
| GPT-4 | 23.58% | 1365 |
| GPT-4 0314 | 22.07% | 1371 |
| Mistral Medium | 21.86% | 1500 |
| Mixtral 8x7B v0.1 | 18.26% | 1465 |
| **Kunoichi-DPO-v2** | **17.19%** | 1785 |
| Claude 2 | 17.19% | 1069 |
| Claude | 16.99% | 1082 |
| Gemini Pro | 16.85% | 1315 |
| GPT-4 0613 | 15.76% | 1140 |
| Claude 2.1 | 15.73% | 1096 |
| Mistral 7B v0.2 | 14.72% | 1676 |
| GPT 3.5 Turbo 0613 | 14.13% | 1328 |
| LLaMA2 Chat 70B | 13.87% | 1790 |
| LMCocktail-10.7B-v1 | 13.15% | 1203 |
| WizardLM 13B V1.1 | 11.23% | 1525 |
| Zephyr 7B Beta | 10.99% | 1444 |
| OpenHermes-2.5-Mistral (7B) | 10.34% | 1107 |
| GPT 3.5 Turbo 0301 | 9.62% | 827 |
| **Kunoichi-7B** | **9.38%** | 1492 |
| GPT 3.5 Turbo 1106 | 9.18% | 796 |
| GPT-3.5 | 8.56% | 1018 |
| Phi-2 DPO | 7.76% | 1687 |
| LLaMA2 Chat 13B | 7.70% | 1513 | |