File size: 2,103 Bytes
cc2fc7f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
---
license: cc-by-nc-4.0
tags:
- not-for-all-audiences
- nsfw
---

## MiquMaid v2 DPO

Check out our blogpost about this model series [Here!](https://ikaridevgit.github.io/index.html?blog=blogid-6&bo=true#Miqu-base) - Join our Discord server [Here!](https://discord.gg/Bb8pRUXy3Z)

<center>[<a href="https://huggingface.co/NeverSleep/MiquMaid-v2-70B">V2-70B</a> - <a href="https://huggingface.co/NeverSleep/MiquMaid-v2-70B-DPO">V2-70B-DPO</a> - <a href="https://huggingface.co/NeverSleep/MiquMaid-v2-2x70B">V2-2x70B</a> - <a href="https://huggingface.co/NeverSleep/MiquMaid-v2-2x70B-DPO">V2-2x70B-DPO</a>]
</br>
<div style="width: 100%;">
    <img src="https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/tPFdudSae6SCDNvhe1lC9.png" style="display: block; margin: auto;">
</div></center>

This model uses the Alpaca **prompting format**

Model trained for RP conversation on Miqu-70B with our magic sauce, then trained on DPO for uncensoring.

## Credits:
- Undi
- IkariDev

## Description

This repo contains FP16 files of MiquMaid-v2-70B-DPO.

Switch: [FP16](https://huggingface.co/NeverSleep/MiquMaid-v2-70B-DPO) - [GGUF](https://huggingface.co/NeverSleep/MiquMaid-v2-70B-DPO-GGUF)

## Training data used:
- [Aesir datasets](https://huggingface.co/MinervaAI)
- [NoRobots](https://huggingface.co/datasets/Doctor-Shotgun/no-robots-sharegpt)
- [limarp](https://huggingface.co/datasets/lemonilia/LimaRP)
- [toxic-dpo-v0.1-sharegpt](https://huggingface.co/datasets/Undi95/toxic-dpo-v0.1-sharegpt)
- [ToxicQAFinal](https://huggingface.co/datasets/NobodyExistsOnTheInternet/ToxicQAFinal)

## DPO training data used:
- [ToxicDPOqa](https://huggingface.co/datasets/NobodyExistsOnTheInternet/ToxicDPOqa)
- [toxic-dpo-v0.1-NoWarning](https://huggingface.co/datasets/Undi95/toxic-dpo-v0.1-NoWarning)

### Custom format:
```
### Instruction:
{system prompt}

### Input:
{input}

### Response:
{reply}
```

## Others

Undi: If you want to support us, you can [here](https://ko-fi.com/undiai).

IkariDev: Visit my [retro/neocities style website](https://ikaridevgit.github.io/) please kek