Undi95 commited on
Commit
3dc0610
1 Parent(s): 7bc510a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +233 -1
README.md CHANGED
@@ -1 +1,233 @@
1
- more info to come
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ tags:
4
+ - not-for-all-audiences
5
+ - nsfw
6
+ ---
7
+
8
+ <!-- description start -->
9
+ ## Description
10
+
11
+ This repo contains fp16 files of UtopiaXL-13B, a merge I have done with the new [layer shuffle](https://github.com/cg123/mergekit/blob/main/mergekit/scripts/layershuffle.py) method from mergekit.
12
+
13
+ This is more a proof of concept showing the following:
14
+ - Llama2 is very flexible
15
+ - Llama2 don't care about what is finetuned on the layers specifically if you keep them in the same order
16
+ - Clean merge (no ties, no SLERP, etc...) with only layer is possible without breaking something
17
+ - Deleting special tokens/using model with special token don't break the model
18
+ - Alpaca win, always. So use it.
19
+
20
+ The name "XL" come from the absurd amount of model pushed into it.
21
+
22
+ <!-- description end -->
23
+ <!-- description start -->
24
+ ## Models and loras used
25
+
26
+ - [Undi95/Utopia-13B](https://huggingface.co/Undi95/Utopia-13B)
27
+ - [KoboldAI/LLAMA2-13B-Holodeck-1](https://huggingface.co/KoboldAI/LLAMA2-13B-Holodeck-1)
28
+ - [Undi95/PsyMedRP-v1-13B](https://huggingface.co/Undi95/PsyMedRP-v1-13B)
29
+ - [PygmalionAI/pygmalion-2-13b](https://huggingface.co/PygmalionAI/pygmalion-2-13b)
30
+ - [Heralax/Cat-0.5](https://huggingface.co/Heralax/Cat-0.5)
31
+ - [KoboldAI/LLaMA2-13B-TiefighterLR](https://huggingface.co/KoboldAI/LLaMA2-13B-TiefighterLR)
32
+ - [Heralax/Augmental-13b-two-epochs](https://huggingface.co/Heralax/Augmental-13b-two-epochs)
33
+ - [Undi95/Storytelling-v2.1-13B-lora](https://huggingface.co/Undi95/Storytelling-v2.1-13B-lora)
34
+ - [Undi95/LimaRP-UtopiaXL-13B-v3-lora](https://huggingface.co/Undi95/LimaRP-UtopiaXL-13B-v3-lora)
35
+
36
+ <!-- description end -->
37
+ ## The sauce
38
+ ```
39
+ !mergekit-layershuffle ./UtopiaXL \
40
+ --model Undi95/Utopia-13B --weight 0.4 \
41
+ --model KoboldAI/LLAMA2-13B-Holodeck-1 --weight 0.1 \
42
+ --model Undi95/PsyMedRP-v1-13B --weight 0.1 \
43
+ --model PygmalionAI/pygmalion-2-13b --weight 0.25 \
44
+ --model Heralax/Cat-0.5 --weight 0.1 \
45
+ --model KoboldAI/LLaMA2-13B-TiefighterLR --weight 0.1 \
46
+ --model Heralax/Augmental-13b-two-epochs --weight 0.1 \
47
+ --write-yaml UtopiaXL.yaml
48
+
49
+ =========================
50
+
51
+ merge_method: passthrough
52
+ slices:
53
+ - sources:
54
+ - layer_range:
55
+ - 0
56
+ - 1
57
+ model: Undi95/Utopia-13B
58
+ - sources:
59
+ - layer_range:
60
+ - 1
61
+ - 2
62
+ model: Undi95/PsyMedRP-v1-13B
63
+ - sources:
64
+ - layer_range:
65
+ - 2
66
+ - 3
67
+ model: KoboldAI/LLaMA2-13B-TiefighterLR
68
+ - sources:
69
+ - layer_range:
70
+ - 3
71
+ - 4
72
+ model: Heralax/Augmental-13b-two-epochs
73
+ - sources:
74
+ - layer_range:
75
+ - 4
76
+ - 5
77
+ model: Undi95/Utopia-13B
78
+ - sources:
79
+ - layer_range:
80
+ - 5
81
+ - 7
82
+ model: KoboldAI/LLaMA2-13B-TiefighterLR
83
+ - sources:
84
+ - layer_range:
85
+ - 7
86
+ - 8
87
+ model: Heralax/Augmental-13b-two-epochs
88
+ - sources:
89
+ - layer_range:
90
+ - 8
91
+ - 11
92
+ model: Undi95/Utopia-13B
93
+ - sources:
94
+ - layer_range:
95
+ - 11
96
+ - 12
97
+ model: Undi95/PsyMedRP-v1-13B
98
+ - sources:
99
+ - layer_range:
100
+ - 12
101
+ - 13
102
+ model: Undi95/Utopia-13B
103
+ - sources:
104
+ - layer_range:
105
+ - 13
106
+ - 14
107
+ model: KoboldAI/LLaMA2-13B-TiefighterLR
108
+ - sources:
109
+ - layer_range:
110
+ - 14
111
+ - 15
112
+ model: PygmalionAI/pygmalion-2-7b
113
+ - sources:
114
+ - layer_range:
115
+ - 15
116
+ - 16
117
+ model: KoboldAI/LLAMA2-13B-Holodeck-1
118
+ - sources:
119
+ - layer_range:
120
+ - 16
121
+ - 17
122
+ model: Heralax/Augmental-13b-two-epochs
123
+ - sources:
124
+ - layer_range:
125
+ - 17
126
+ - 18
127
+ model: Heralax/Cat-0.5
128
+ - sources:
129
+ - layer_range:
130
+ - 18
131
+ - 19
132
+ model: Undi95/PsyMedRP-v1-13B
133
+ - sources:
134
+ - layer_range:
135
+ - 19
136
+ - 20
137
+ model: KoboldAI/LLAMA2-13B-Holodeck-1
138
+ - sources:
139
+ - layer_range:
140
+ - 20
141
+ - 22
142
+ model: Undi95/Utopia-13B
143
+ - sources:
144
+ - layer_range:
145
+ - 22
146
+ - 23
147
+ model: KoboldAI/LLaMA2-13B-TiefighterLR
148
+ - sources:
149
+ - layer_range:
150
+ - 23
151
+ - 25
152
+ model: Undi95/Utopia-13B
153
+ - sources:
154
+ - layer_range:
155
+ - 25
156
+ - 26
157
+ model: KoboldAI/LLAMA2-13B-Holodeck-1
158
+ - sources:
159
+ - layer_range:
160
+ - 26
161
+ - 27
162
+ model: Undi95/Utopia-13B
163
+ - sources:
164
+ - layer_range:
165
+ - 27
166
+ - 28
167
+ model: PygmalionAI/pygmalion-2-7b
168
+ - sources:
169
+ - layer_range:
170
+ - 28
171
+ - 29
172
+ model: Heralax/Cat-0.5
173
+ - sources:
174
+ - layer_range:
175
+ - 29
176
+ - 30
177
+ model: Undi95/Utopia-13B
178
+ - sources:
179
+ - layer_range:
180
+ - 30
181
+ - 32
182
+ model: Undi95/PsyMedRP-v1-13B
183
+ - sources:
184
+ - layer_range:
185
+ - 32
186
+ - 34
187
+ model: PygmalionAI/pygmalion-2-7b
188
+ - sources:
189
+ - layer_range:
190
+ - 34
191
+ - 36
192
+ model: Undi95/Utopia-13B
193
+ - sources:
194
+ - layer_range:
195
+ - 36
196
+ - 37
197
+ model: KoboldAI/LLAMA2-13B-Holodeck-1
198
+ - sources:
199
+ - layer_range:
200
+ - 37
201
+ - 38
202
+ model: Undi95/Utopia-13B
203
+ - sources:
204
+ - layer_range:
205
+ - 38
206
+ - 39
207
+ model: Undi95/PsyMedRP-v1-13B
208
+ - sources:
209
+ - layer_range:
210
+ - 39
211
+ - 40
212
+ model: KoboldAI/LLAMA2-13B-Holodeck-1
213
+
214
+ =========================
215
+
216
+ => Applying Undi95/Storytelling-v2.1-13B-lora x 0.1
217
+ => Trained on LimaRP for +2h
218
+ => Applying Undi95/LimaRP-UtopiaXL-13B-v3-lora x 0.35
219
+ ```
220
+ <!-- prompt-template start -->
221
+ ## Prompt template: Alpaca
222
+
223
+ ```
224
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
225
+
226
+ ### Instruction:
227
+ {prompt}
228
+
229
+ ### Response:
230
+
231
+ ```
232
+
233
+ If you want to support me, you can [here](https://ko-fi.com/undiai).