File size: 6,841 Bytes
0d7ee2f 7b5b62b 8bb897c 7b5b62b f0c7f36 3380b02 7b5b62b f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 e092060 f0c7f36 e092060 f0c7f36 e092060 f0c7f36 3380b02 19de066 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 f0c7f36 3380b02 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 |
---
pipeline_tag: text-generation
library_name: transformers
language:
- en
license: llama3
tags:
- mergekit
- merge
- multi-step merge
- rp
- roleplay
- role-play
- chain-of-thoughts
- summarization
- emotion classification
- biology
- psychology
base_model:
- nothingiisreal/L3-8B-Celeste-v1
- Nitral-AI/Hathor_Tahsin-L3-8B-v0.85
- Sao10K/L3-8B-Stheno-v3.2
- ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B
- Sao10K/L3-8B-Lunaris-v1
- turboderp/llama3-turbcat-instruct-8b
- ChaoticNeutrals/Domain-Fusion-L3-8B
- migtissera/Llama-3-8B-Synthia-v3.5
- TheDrummer/Llama-3SOME-8B-v2
- ChaoticNeutrals/Hathor_RP-v.01-L3-8B
- TheSkullery/llama-3-cat-8b-instruct-v1
- FPHam/L3-8B-Everything-COT
- Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-24K-LoRANET-Merged
- OEvortex/Emotional-llama-8B
- lighteternal/Llama3-merge-biomed-8b
- Casual-Autopsy/Llama3-merge-psychotherapy-8b
---
| <img src="https://huggingface.co/Casual-Autopsy/L3-Super-Nova-RP-8B/resolve/main/Card-Assets/NovaKid-Girl.jpeg" width="50%" height="50%" style="display: block; margin: auto;"> |
|:---:|
| Image generated by [mayonays_on_toast](https://civitai.com/user/mayonays_on_toast) - [Sauce](https://civitai.com/images/10153472) |
***
***
***
# L3-Super-Nova-RP-8B
***
***
## Presets
I've(or anyone else) yet to find good Textgen Preset so here's the starting point preset I use instead, It should get you by for now.
```yaml
Top K: 50
Top P: 0.85
Repetition Penalty: 1.01
# Don't make this higher, DRY handles the bulk of Squashing Repetition.
# This is justs to lightly nudge the bot to move the plot forward
Rep Pen Range: 2048 # Don't make this higher either.
Presence Penalty: 0.03 # Minor encouragement to use synonyms.
Smoothing Factor: 0.3
DRY Repetition Penalty:
Multiplier: 0.8
Base: 1.75
Allowed Length: 2
Penalty Range: 4096
Dynamic Temperature:
Min Temp: 0.5
Max Temp: 1.25
Exponent: 0.85
```
***
***
## Usage Info
Some of the **INT** models were chosen with some of SillyTavern's features in mind, such as emotion based sprites, dynamic music, and pretty much any feature, extension, or STscript that uses sumarization. With that said, it's recommended to use SillyTavern as your front-end.
While not required, I'd recommend building the story string prompt with Lorebooks rather than using the Advance Formatting menu. The only thing you really need in the Story String prompt within Advance Formatting is the system prompt. Doing it this way tends to keep the character more consistent as the RP goes on as all character card info is locked to a certain depth rather than getting further and further away within the context.
***
***
## Quants
***
***
## Merge Info
The merge methods used were **Ties**, **Dare Ties**, **Breadcrumbs Ties**, **SLERP**, and **Task Arithmetic**.
The model was finished off with both **Merge Densification**, and **Negative Weighting** tenchniques to boost creativity.
All merging steps had the merge calculations done in **float32** and were output as **bfloat16**.
***
### Models Merged
The following models were used to make this merge:
* [nothingiisreal/L3-8B-Celeste-v1](https://huggingface.co/nothingiisreal/L3-8B-Celeste-v1)
* [Nitral-AI/Hathor_Tahsin-L3-8B-v0.85](https://huggingface.co/Nitral-AI/Hathor_Tahsin-L3-8B-v0.85)
* [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2)
* [ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B](https://huggingface.co/ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B)
* [Sao10K/L3-8B-Lunaris-v1](https://huggingface.co/Sao10K/L3-8B-Lunaris-v1)
* [turboderp/llama3-turbcat-instruct-8b](https://huggingface.co/turboderp/llama3-turbcat-instruct-8b)
* [ChaoticNeutrals/Domain-Fusion-L3-8B](https://huggingface.co/ChaoticNeutrals/Domain-Fusion-L3-8B)
* [migtissera/Llama-3-8B-Synthia-v3.5](https://huggingface.co/migtissera/Llama-3-8B-Synthia-v3.5)
* [TheDrummer/Llama-3SOME-8B-v2](https://huggingface.co/TheDrummer/Llama-3SOME-8B-v2)
* [ChaoticNeutrals/Hathor_RP-v.01-L3-8B](https://huggingface.co/ChaoticNeutrals/Hathor_RP-v.01-L3-8B)
* [TheSkullery/llama-3-cat-8b-instruct-v1](https://huggingface.co/TheSkullery/llama-3-cat-8b-instruct-v1)
* [FPHam/L3-8B-Everything-COT](https://huggingface.co/FPHam/L3-8B-Everything-COT)
* [Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-24K-LoRANET-Merged](https://huggingface.co/Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-24K-LoRANET-Merged)
* [OEvortex/Emotional-llama-8B](https://huggingface.co/OEvortex/Emotional-llama-8B)
* [lighteternal/Llama3-merge-biomed-8b](https://huggingface.co/lighteternal/Llama3-merge-biomed-8b)
* [Casual-Autopsy/Llama3-merge-psychotherapy-8b](https://huggingface.co/Casual-Autopsy/Llama3-merge-psychotherapy-8b)
***
***
## Evaluation Results
***
### [Open LLM Leaderboard](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Casual-Autopsy__L3-Umbral-Mind-RP-v2.0-8B)
**Explaination for AI RP newbies:** IFEval is the most important evaluation for RP AIs as it determines how well it can follow OOC, Lorebooks, and most importantly character cards.
The rest don't matter. At least not nearly as much as IFEval.
|Metric | Value|
|:------------------|------:|
|Avg. |N/A|
|IFEval (0-Shot) |N/A|
|BBH (3-Shot) |N/A|
|MATH Lvl 5 (4-Shot)|N/A|
|GPQA (0-shot) |N/A|
|MuSR (0-shot) |N/A|
|MMLU-PRO (5-shot) |N/A|
***
### [UGI Leaderboard](https://huggingface.co/spaces/DontPlanToEnd/UGI-Leaderboard)
Information about the metrics can be found at the bottom of the [UGI Leaderboard](https://huggingface.co/spaces/DontPlanToEnd/UGI-Leaderboard) in the respective tabs.
|Metric(UGI-Leaderboard) | Value | Value | Metric(Writing Style)|
|:------------------------|:-----:|:-----:|----------------------:|
|UGI(Avg.) |N/A|N/A|RegV1 |
|W/10 |N/A|N/A|RegV2 |
|Unruly |N/A|N/A|MyScore |
|Internet |N/A|N/A|ASSS |
|Stats |N/A|N/A|SMOG |
|Writing |N/A|N/A|Yule |
|PolContro |N/A| | |
***
***
## Secret Sauce
The following YAML configs were used to make this merge.
***
### Super-Nova-CRE_pt.1
```yaml
```
***
### Super-Nova-CRE_pt.2
```yaml
```
***
### Super-Nova-UNC_pt.1
```yaml
```
***
### Super-Nova-UNC_pt.2
```yaml
```
***
### Super-Nova-INT_pt.1
```yaml
```
***
### Super-Nova-INT_pt.2
```yaml
```
***
### Super-Nova-CRE
```yaml
```
***
### Super-Nova-UNC
```yaml
```
***
### Super-Nova-INT
```yaml
```
***
### Super-Nova-RP_pt.1
```yaml
```
***
### Super-Nova-RP_pt.2
```yaml
```
***
### L3-Super-Nova-RP-8B
```yaml
``` |