Casual-Autopsy commited on
Commit
70d033a
1 Parent(s): b14e98f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +153 -15
README.md CHANGED
@@ -1,35 +1,174 @@
1
  ---
2
- base_model:
3
- - Hastagaras/Halu-8B-Llama3-Blackroot
4
- - Casual-Autopsy/Umbral-Mind
5
- - Casual-Autopsy/Llama-3-Mopeyfied-Psychology-8B
6
- - Sao10K/L3-8B-Stheno-v3.3-32K
7
- library_name: transformers
8
  tags:
9
- - mergekit
10
  - merge
11
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  ---
13
- # merge
 
 
 
 
 
14
 
15
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
 
17
  ## Merge Details
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  ### Merge Method
19
 
20
- This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [Casual-Autopsy/Umbral-Mind](https://huggingface.co/Casual-Autopsy/Umbral-Mind) as a base.
21
 
22
  ### Models Merged
23
 
24
  The following models were included in the merge:
25
- * [Hastagaras/Halu-8B-Llama3-Blackroot](https://huggingface.co/Hastagaras/Halu-8B-Llama3-Blackroot)
26
- * [Casual-Autopsy/Llama-3-Mopeyfied-Psychology-8B](https://huggingface.co/Casual-Autopsy/Llama-3-Mopeyfied-Psychology-8B)
27
  * [Sao10K/L3-8B-Stheno-v3.3-32K](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.3-32K)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
 
29
- ### Configuration
30
 
31
  The following YAML configuration was used to produce this model:
32
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  ```yaml
34
  slices:
35
  - sources:
@@ -50,5 +189,4 @@ slices:
50
  merge_method: task_arithmetic
51
  base_model: Casual-Autopsy/Umbral-Mind
52
  dtype: bfloat16
53
-
54
- ```
 
1
  ---
 
 
 
 
 
 
2
  tags:
 
3
  - merge
4
+ - mergekit
5
+ - lazymergekit
6
+ - not-for-all-audiences
7
+ - nsfw
8
+ - rp
9
+ - roleplay
10
+ - role-play
11
+ license: llama3
12
+ language:
13
+ - en
14
+ library_name: transformers
15
+ pipeline_tag: text-generation
16
+ base_model:
17
+ - Sao10K/L3-8B-Stheno-v3.3-32K
18
+ - Hastagaras/Jamet-8B-L3-MK.V-Blackroot
19
+ - grimjim/Llama-3-Oasis-v1-OAS-8B
20
+ - Casual-Autopsy/SOVL-MopeyMule-8B
21
+ - Casual-Autopsy/MopeyMule-Blackroot-8B
22
+ - ResplendentAI/Theory_of_Mind_Llama3
23
+ - ResplendentAI/RP_Format_QuoteAsterisk_Llama3
24
+ - ResplendentAI/Smarts_Llama3
25
+ - Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B
26
+ - Casual-Autopsy/Llama-3-Mopeyfied-Psychology-8B
27
+ - Hastagaras/Halu-8B-Llama3-Blackroot
28
  ---
29
+ <img src="https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v3-8B/resolve/main/63073798_p0_master1200.jpg" style="display: block; margin: auto;">
30
+ Image by ろ47
31
+
32
+ **Highest ranked 8B model on the [UGI Leaderboard](https://huggingface.co/spaces/DontPlanToEnd/UGI-Leaderboard) as of writing this!**
33
+
34
+ # Merge
35
 
36
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
37
 
38
  ## Merge Details
39
+
40
+ The goal of this merge was to make an RP model better suited for role-plays with heavy themes such as but not limited to:
41
+ - Mental illness
42
+ - Self-harm
43
+ - Trauma
44
+ - Suicide
45
+
46
+ I hated how RP models tended to be overly positive and hopeful with role-plays involving such themes,
47
+ but thanks to [failspy/Llama-3-8B-Instruct-MopeyMule](https://huggingface.co/failspy/Llama-3-8B-Instruct-MopeyMule) this problem has been lessened considerably.
48
+
49
+ If you're an enjoyer of savior/reverse savior type role-plays like myself, then this model is for you.
50
+
51
+ ### Usage Info
52
+
53
+ This model is meant to be used with asterisks/quotes RPing formats, any other format that isn't asterisks/quotes is likely to cause issues
54
+
55
+ ### Quants
56
+
57
+
58
+
59
  ### Merge Method
60
 
61
+ This model was merged using several Task Arithmetic merges and then tied together with a Model Stock merge, followed by another Task Arithmetic merge with a model containing psychology data.
62
 
63
  ### Models Merged
64
 
65
  The following models were included in the merge:
 
 
66
  * [Sao10K/L3-8B-Stheno-v3.3-32K](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.3-32K)
67
+ * [Hastagaras/Halu-8B-Llama3-Blackroot](Hastagaras/Halu-8B-Llama3-Blackroot)
68
+ * [Casual-Autopsy/Llama-3-Mopeyfied-Psychology-8B](https://huggingface.co/Casual-Autopsy/Llama-3-Mopeyfied-Psychology-8B)
69
+ * [Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B](https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B)
70
+ * Casual-Autopsy/Umbral-v3-1 + [ResplendentAI/Theory_of_Mind_Llama3](https://huggingface.co/ResplendentAI/Theory_of_Mind_Llama3)
71
+ * [Sao10K/L3-8B-Stheno-v3.3-32K](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.3-32K)
72
+ * [Casual-Autopsy/SOVL-MopeyMule-8B](https://huggingface.co/Casual-Autopsy/SOVL-MopeyMule-8B)
73
+ * [Casual-Autopsy/MopeyMule-Blackroot-8B](https://huggingface.co/Casual-Autopsy/MopeyMule-Blackroot-8B)
74
+
75
+ * Casual-Autopsy/Umbral-v3-2 + [ResplendentAI/Smarts_Llama3](https://huggingface.co/ResplendentAI/Smarts_Llama3)
76
+ * [Hastagaras/Jamet-8B-L3-MK.V-Blackroot](https://huggingface.co/Hastagaras/Jamet-8B-L3-MK.V-Blackroot)
77
+ * [Casual-Autopsy/SOVL-MopeyMule-8B](https://huggingface.co/Casual-Autopsy/SOVL-MopeyMule-8B)
78
+ * [Casual-Autopsy/MopeyMule-Blackroot-8B](https://huggingface.co/Casual-Autopsy/MopeyMule-Blackroot-8B)
79
+
80
+ * Casual-Autopsy/Umbral-v3-3 + [ResplendentAI/RP_Format_QuoteAsterisk_Llama3](https://huggingface.co/ResplendentAI/RP_Format_QuoteAsterisk_Llama3)
81
+ * [grimjim/Llama-3-Oasis-v1-OAS-8B](https://huggingface.co/grimjim/Llama-3-Oasis-v1-OAS-8B)
82
+ * [Casual-Autopsy/SOVL-MopeyMule-8B](https://huggingface.co/Casual-Autopsy/SOVL-MopeyMule-8B)
83
+ * [Casual-Autopsy/MopeyMule-Blackroot-8B](https://huggingface.co/Casual-Autopsy/MopeyMule-Blackroot-8B)
84
 
85
+ ## Secret Sauce
86
 
87
  The following YAML configuration was used to produce this model:
88
 
89
+ ### Umbral-1
90
+
91
+ ```yaml
92
+ slices:
93
+ - sources:
94
+ - model: Sao10K/L3-8B-Stheno-v3.3-32K
95
+ layer_range: [0, 32]
96
+ parameters:
97
+ weight: 0.65
98
+ - model: Casual-Autopsy/SOVL-MopeyMule-8B
99
+ layer_range: [0, 32]
100
+ parameters:
101
+ weight: 0.25
102
+ - model: Casual-Autopsy/MopeyMule-Blackroot-8B
103
+ layer_range: [0, 32]
104
+ parameters:
105
+ weight: 0.1
106
+ merge_method: task_arithmetic
107
+ base_model: Sao10K/L3-8B-Stheno-v3.2
108
+ normalize: False
109
+ dtype: bfloat16
110
+ ```
111
+
112
+ ### Umbral-2
113
+
114
+ ```yaml
115
+ slices:
116
+ - sources:
117
+ - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
118
+ layer_range: [0, 32]
119
+ parameters:
120
+ weight: 0.75
121
+ - model: Casual-Autopsy/SOVL-MopeyMule-8B
122
+ layer_range: [0, 32]
123
+ parameters:
124
+ weight: 0.15
125
+ - model: Casual-Autopsy/MopeyMule-Blackroot-8B
126
+ layer_range: [0, 32]
127
+ parameters:
128
+ weight: 0.1
129
+ merge_method: task_arithmetic
130
+ base_model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
131
+ normalize: False
132
+ dtype: bfloat16
133
+ ```
134
+
135
+ ### Umbral-3
136
+
137
+ ```yaml
138
+ slices:
139
+ - sources:
140
+ - model: grimjim/Llama-3-Oasis-v1-OAS-8B
141
+ layer_range: [0, 32]
142
+ parameters:
143
+ weight: 0.55
144
+ - model: Casual-Autopsy/SOVL-MopeyMule-8B
145
+ layer_range: [0, 32]
146
+ parameters:
147
+ weight: 0.35
148
+ - model: Casual-Autopsy/MopeyMule-Blackroot-8B
149
+ layer_range: [0, 32]
150
+ parameters:
151
+ weight: 0.1
152
+ merge_method: task_arithmetic
153
+ base_model: grimjim/Llama-3-Oasis-v1-OAS-8B
154
+ normalize: False
155
+ dtype: bfloat16
156
+ ```
157
+
158
+ ### Umbral-Mind
159
+
160
+ ```yaml
161
+ models:
162
+ - model: Casual-Autopsy/Umbral-1+ResplendentAI/Theory_of_Mind_Llama3
163
+ - model: Casual-Autopsy/Umbral-2+ResplendentAI/Smarts_Llama3
164
+ - model: Casual-Autopsy/Umbral-3+ResplendentAI/RP_Format_QuoteAsterisk_Llama3
165
+ merge_method: model_stock
166
+ base_model: Casual-Autopsy/Umbral-1
167
+ dtype: bfloat16
168
+ ```
169
+
170
+ ### L3-Umbral-Mind-RP-v1.0.1-8B
171
+
172
  ```yaml
173
  slices:
174
  - sources:
 
189
  merge_method: task_arithmetic
190
  base_model: Casual-Autopsy/Umbral-Mind
191
  dtype: bfloat16
192
+ ```