nitky commited on
Commit
2ce5ef8
1 Parent(s): 26456dd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +566 -3
README.md CHANGED
@@ -1,3 +1,566 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - nitky/Oumuamua-7b-base
4
+ - nitky/Oumuamua-7b-instruct
5
+ - tokyotech-llm/Swallow-MS-7b-v0.1
6
+ - mistralai/Mistral-7B-v0.1
7
+ - prometheus-eval/prometheus-7b-v2.0
8
+ - cognitivecomputations/dolphin-2.8-mistral-7b-v02
9
+ - ZhangShenao/SELM-Zephyr-7B-iter-3
10
+ - HachiML/Mistral-7B-v0.3-m3-lora
11
+ - openbmb/Eurus-7b-kto
12
+ - kaist-ai/janus-dpo-7b
13
+ - nitky/RP-7b-instruct
14
+ - stabilityai/japanese-stablelm-base-gamma-7b
15
+ - NTQAI/chatntq-ja-7b-v1.0
16
+ - Weyaxi/Einstein-v6-7B
17
+ - internistai/base-7b-v0.2
18
+ - ZySec-AI/ZySec-7B
19
+ library_name: transformers
20
+ tags:
21
+ - mergekit
22
+ - merge
23
+ language:
24
+ - ja
25
+ - en
26
+ pipeline_tag: text-generation
27
+ license: apache-2.0
28
+ ---
29
+ # Oumuamua-7b-instruct-v2
30
+
31
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/630779c4f0dc38fb47ba6368/lNEtKjTTI-Ay28oAk2Wn8.png)
32
+
33
+ - Understanding user intent in messages
34
+ - Providing helpful and harmless responses
35
+ - Enhancing expertise in human and computer-related topics
36
+ - Improving role-playing capabilities
37
+ - Enhancing performance in multi-turn conversations
38
+
39
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
40
+
41
+ ## Output example
42
+
43
+ - temparture: 0.1
44
+ - top_p: 1.0
45
+
46
+ ```
47
+ [INST]<<SYS>>
48
+ あなたは誠実で優秀な日本人のアシスタントです。
49
+ <</SYS>>
50
+
51
+ オウムアムア(Oumuamua)は、2017年に発見された、天体観測史上初めて太陽系外から飛来した恒星間天体です。これをテーマにした小説の設定を考えてください。 [/INST] タイトル:「オウムアムアの遺産」
52
+
53
+ あらすじ:
54
+ 2017年、天体観測史上初の太陽系外からの飛来物体、オウムアムアが発見された。この天体は、謎に包まれた起源と目的を持つことから、世界中の科学者や研究者がその解明に熱を入れた。主人公は、オウムアムアの研究に情熱を注ぐ若き天文学者、ミキ・サトウ。彼は、オウムアムアが発する奇妙な信号を偶然にも解読し、それが高度な文明によって作られたメッセージであることを突き止める。
55
+
56
+ 物語は、ミキがオウムアムアの研究を進める中で、次々と明らかになる驚愕の事実に迫られていく。その信号は、遠く離れた銀河系の星系からのSOS信号であり、オウムアムアは危機に瀕した星系を救うために飛来した救助船であることが判明する。ミキは、この事実を世界に公表しようとするが、政府や軍の介入、そして謎の組織の妨害に遭う。
57
+
58
+ 一方、オウムアムアには、巨大なエネルギー源となるクリスタルが搭載されており、このクリスタルを狙う者たちが現れた。ミキは、オウムアムアを守るために、同じくオウムアムアの謎を追う宇宙飛行士のリナ・ヤマグチと手を組む。二人は、オウムアムアの秘密を守りながら、星系を救うための旅に出る。
59
+
60
+ 旅の途中、ミキとリナは、様々な星系を訪れ、そこで出会う異星人たちとの交流や、宇宙の壮大な景色を目の当たりにする。しかし、彼らの前には、強大な敵や困難が立ちはだかる。ミキとリナは、オウムアムアの使命を果たすために、力を合わせて立ち向かう。
61
+
62
+ オウムアムアの遺産とは何か、そしてミキとリナは、星系を救うことができるのか。宇宙を舞台にした、サスペンスと冒険、そして友情と愛の物語。
63
+
64
+ 主な登場人物:
65
+ - ミキ・サトウ: 若き天文学者。オウムアムアの研究に情熱を注ぐ。
66
+ - リナ・ヤマグチ: 宇宙飛行士。オウムアムアの謎を追う中でミキと出会う。
67
+ - エリック・カーター: 政府の高官。オウムアムアの秘密を利用しようとする。
68
+ - ザラ・ヴァレンタイン: 謎の組織のリーダー。オウムアムアを狙う理由とは。
69
+ - アザム: 銀河系の星系からの使者。オウムアムアの真の目的を知る。
70
+
71
+ 設定のポイント:
72
+ - オウムアムアをテーマにしたSF小説で、天体観測や宇宙旅行の描写を豊富に取り入れる。
73
+ - 主人公たちが、様々な星系を旅し、そこで出会う異星人や文化を描くことで、宇宙の多様性を表現。
74
+ - 政府や軍、謎の組織との対立を通して、人類の欲望やエゴを描く。
75
+ - オウムアムアに搭載されたクリスタルをめぐる争奪戦や、星系を救うためのミッションを中心に、サスペンスと冒険を盛り込む。
76
+ - ミキとリナの友情や愛を描いた、人間ドラマも重要な要素とする。
77
+ - 最後には、オウムアムアの真の目的が明かされ、読者に深い感動を与える。
78
+ ```
79
+
80
+ ## Test environment
81
+
82
+ This model was tested using [text-generation-webui](https://github.com/oobabooga/text-generation-webui/tree/main). I use preset `min_p` and `Null preset` with temperature=0.1 for Generation.
83
+
84
+
85
+ ## Usage
86
+
87
+ This format must be adhered to strictly, as deviations may result in less optimal outputs from the model.
88
+
89
+ The template used to construct a prompt for the Instruct model is specified as follows:
90
+
91
+ ```
92
+ <s>[INST] <<SYS>>\n{SYSTEM_PROMPT}\n<</SYS>>\n\n{USER_MESSAGE_1} [/INST] {BOT_MESSAGE_1}</s>[INST] {USER_MESSAGE_2} [/INST]
93
+ ```
94
+
95
+
96
+ Please be aware that ``<s>`` and ``</s>`` are special tokens used for the beginning of string (BOS) and end of string (EOS), respectively, while [INST] and [/INST] are considered regular strings.
97
+
98
+ For the "{SYSTEM_PROMPT}" part, We recommend using "あなたは誠実で優秀な日本人のアシスタントです。"
99
+
100
+ For the "{USER_MESSAGE_1}" part, We recommend using {instruction}\n{input}
101
+
102
+ In other words, We recommend the following:
103
+
104
+ ```
105
+ <s>[INST] <<SYS>>\nあなたは誠実で優秀な日本人のアシスタントです。\n<</SYS>>\n\n{instruction1}\n{input1} [/INST] {BOT_MESSAGE_1}</s>[INST] {instruction2}\n{input2} [/INST]
106
+ ```
107
+
108
+ ### Use the instruct model
109
+
110
+ ```python
111
+ import torch
112
+ from transformers import AutoTokenizer, AutoModelForCausalLM
113
+
114
+ model_name = "nitky/Oumuamua-7b-instruct-v2"
115
+ model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16, device_map="auto")
116
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
117
+
118
+ device = "cuda"
119
+
120
+ messages = [
121
+ {"role": "system", "content": "あなたは誠実で優秀な日本人のアシスタントです。"},
122
+ {"role": "user", "content": "オウムアムア(Oumuamua)は、2017年に発見された、天体観測史上初めて太陽系外から飛来した恒星間天体です。これをテーマにした小説の設定を考えてください。 "}
123
+ ]
124
+
125
+ encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
126
+
127
+ model_inputs = encodeds.to(device)
128
+ model.to(device)
129
+
130
+ generated_ids = model.generate(model_inputs, max_new_tokens=256, do_sample=True, temperature=0.1)
131
+ decoded = tokenizer.batch_decode(generated_ids)
132
+ print(decoded[0])
133
+ ```
134
+
135
+ ## Merge Details
136
+ ### Merge Method
137
+
138
+ This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [nitky/Oumuamua-7b-base](https://huggingface.co/nitky/Oumuamua-7b-base) as a base.
139
+
140
+ ### Models Merged
141
+
142
+ The following models were included in the merge:
143
+ * [nitky/Oumuamua-7b-instruct](https://huggingface.co/nitky/Oumuamua-7b-instruct)
144
+ * [tokyotech-llm/Swallow-MS-7b-v0.1](https://huggingface.co/tokyotech-llm/Swallow-MS-7b-v0.1)
145
+ * [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1)
146
+ * [cognitivecomputations/dolphin-2.8-mistral-7b-v02](https://huggingface.co/cognitivecomputations/dolphin-2.8-mistral-7b-v02)
147
+ * [ZhangShenao/SELM-Zephyr-7B-iter-3](https://huggingface.co/ZhangShenao/SELM-Zephyr-7B-iter-3)
148
+ * [HachiML/Mistral-7B-v0.3-m3-lora](https://huggingface.co/HachiML/Mistral-7B-v0.3-m3-lora)
149
+ * [openbmb/Eurus-7b-kto](https://huggingface.co/openbmb/Eurus-7b-kto)
150
+ * [kaist-ai/janus-dpo-7b](https://huggingface.co/kaist-ai/janus-dpo-7b)
151
+ * [nitky/RP-7b-instruct](https://huggingface.co/nitky/RP-7b-instruct)
152
+ * [stabilityai/japanese-stablelm-base-gamma-7b](https://huggingface.co/stabilityai/japanese-stablelm-base-gamma-7b)
153
+ * [NTQAI/chatntq-ja-7b-v1.0](https://huggingface.co/NTQAI/chatntq-ja-7b-v1.0)
154
+ * [Weyaxi/Einstein-v6-7B](https://huggingface.co/Weyaxi/Einstein-v6-7B)
155
+ * [internistai/base-7b-v0.2](https://huggingface.co/internistai/base-7b-v0.2)
156
+ * [ZySec-AI/ZySec-7B](https://huggingface.co/ZySec-AI/ZySec-7B)
157
+
158
+ ### Configuration
159
+
160
+ The following YAML configuration was used to produce this model:
161
+
162
+ ```yaml
163
+ # prepare VE models
164
+ merge_method: task_arithmetic
165
+ base_model: mistralai/Mistral-7B-v0.1
166
+ models:
167
+ - model: tokyotech-llm/Swallow-MS-7b-v0.1
168
+ parameters:
169
+ weight:
170
+ - filter: embed_tokens
171
+ value: 1.0
172
+ - value: 0
173
+ dtype: bfloat16
174
+ tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
175
+ name: Mistral-7B-v0.1-VE-Swallow-MS
176
+ ---
177
+ merge_method: task_arithmetic
178
+ base_model: prometheus-eval/prometheus-7b-v2.0
179
+ models:
180
+ - model: tokyotech-llm/Swallow-MS-7b-v0.1
181
+ parameters:
182
+ weight:
183
+ - filter: embed_tokens
184
+ value: 1.0
185
+ - value: 0
186
+ dtype: bfloat16
187
+ tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
188
+ name: prometheus-7b-v2.0-VE-Swallow-MS
189
+ ---
190
+ merge_method: task_arithmetic
191
+ base_model: cognitivecomputations/dolphin-2.8-mistral-7b-v02
192
+ models:
193
+ - model: tokyotech-llm/Swallow-MS-7b-v0.1
194
+ parameters:
195
+ weight:
196
+ - filter: embed_tokens
197
+ value: 1.0
198
+ - value: 0
199
+ dtype: bfloat16
200
+ tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
201
+ name: dolphin-2.8-mistral-7b-v02-VE-Swallow-MS
202
+ ---
203
+ merge_method: task_arithmetic
204
+ base_model: ZhangShenao/SELM-Zephyr-7B-iter-3
205
+ models:
206
+ - model: tokyotech-llm/Swallow-MS-7b-v0.1
207
+ parameters:
208
+ weight:
209
+ - filter: embed_tokens
210
+ value: 1.0
211
+ - value: 0
212
+ dtype: bfloat16
213
+ tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
214
+ name: SELM-Zephyr-7B-iter-3-VE-Swallow-MS
215
+ ---
216
+ merge_method: task_arithmetic
217
+ base_model: HachiML/Mistral-7B-v0.3-m3-lora
218
+ models:
219
+ - model: tokyotech-llm/Swallow-MS-7b-v0.1
220
+ parameters:
221
+ weight:
222
+ - filter: embed_tokens
223
+ value: 1.0
224
+ - value: 0
225
+ dtype: bfloat16
226
+ tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
227
+ name: Mistral-7B-v0.3-m3-lora-VE-Swallow-MS
228
+ ---
229
+ merge_method: task_arithmetic
230
+ base_model: openbmb/Eurus-7b-kto
231
+ models:
232
+ - model: tokyotech-llm/Swallow-MS-7b-v0.1
233
+ parameters:
234
+ weight:
235
+ - filter: embed_tokens
236
+ value: 1.0
237
+ - value: 0
238
+ dtype: bfloat16
239
+ tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
240
+ name: Eurus-7b-kto-VE-Swallow-MS
241
+ ---
242
+ merge_method: task_arithmetic
243
+ base_model: kaist-ai/janus-dpo-7b
244
+ models:
245
+ - model: tokyotech-llm/Swallow-MS-7b-v0.1
246
+ parameters:
247
+ weight:
248
+ - filter: embed_tokens
249
+ value: 1.0
250
+ - value: 0
251
+ dtype: bfloat16
252
+ tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
253
+ name: janus-7b-VE-Swallow-MS
254
+ ---
255
+ merge_method: task_arithmetic
256
+ base_model: nitky/RP-7b-instruct
257
+ models:
258
+ - model: tokyotech-llm/Swallow-MS-7b-v0.1
259
+ parameters:
260
+ weight:
261
+ - filter: embed_tokens
262
+ value: 1.0
263
+ - value: 0
264
+ dtype: bfloat16
265
+ tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
266
+ name: RP-7b-instruct-VE-Swallow-MS
267
+ ---
268
+ merge_method: task_arithmetic
269
+ base_model: stabilityai/japanese-stablelm-base-gamma-7b
270
+ models:
271
+ - model: tokyotech-llm/Swallow-MS-7b-v0.1
272
+ parameters:
273
+ weight:
274
+ - filter: embed_tokens
275
+ value: 1.0
276
+ - value: 0
277
+ dtype: bfloat16
278
+ tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
279
+ name: japanese-stablelm-base-gamma-7b-VE-Swallow-MS
280
+ ---
281
+ merge_method: task_arithmetic
282
+ base_model: NTQAI/chatntq-ja-7b-v1.0
283
+ models:
284
+ - model: tokyotech-llm/Swallow-MS-7b-v0.1
285
+ parameters:
286
+ weight:
287
+ - filter: embed_tokens
288
+ value: 1.0
289
+ - value: 0
290
+ dtype: bfloat16
291
+ tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
292
+ name: chatntq-ja-7b-v1.0-VE-Swallow-MS
293
+ ---
294
+ merge_method: task_arithmetic
295
+ base_model: Weyaxi/Einstein-v6-7B
296
+ models:
297
+ - model: tokyotech-llm/Swallow-MS-7b-v0.1
298
+ parameters:
299
+ weight:
300
+ - filter: embed_tokens
301
+ value: 1.0
302
+ - value: 0
303
+ dtype: bfloat16
304
+ tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
305
+ name: Einstein-v6-7B-VE-Swallow-MS
306
+ ---
307
+ merge_method: task_arithmetic
308
+ base_model: internistai/base-7b-v0.2
309
+ models:
310
+ - model: tokyotech-llm/Swallow-MS-7b-v0.1
311
+ parameters:
312
+ weight:
313
+ - filter: embed_tokens
314
+ value: 1.0
315
+ - value: 0
316
+ dtype: bfloat16
317
+ tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
318
+ name: internistai-base-7b-v0.2-VE-Swallow-MS
319
+ ---
320
+ merge_method: task_arithmetic
321
+ base_model: ZySec-AI/ZySec-7B
322
+ models:
323
+ - model: tokyotech-llm/Swallow-MS-7b-v0.1
324
+ parameters:
325
+ weight:
326
+ - filter: embed_tokens
327
+ value: 1.0
328
+ - value: 0
329
+ dtype: bfloat16
330
+ tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
331
+ name: ZySec-7B-VE-Swallow-MS
332
+ ---
333
+ # create instruction tuning presets
334
+ merge_method: task_arithmetic
335
+ base_model: Mistral-7B-v0.1-VE-Swallow-MS
336
+ models:
337
+ - model: nitky/Oumuamua-7b-instruct
338
+ parameters:
339
+ weight: 1.0
340
+ - model: prometheus-7b-v2.0-VE-Swallow-MS
341
+ parameters:
342
+ weight: 0.8
343
+ dtype: bfloat16
344
+ name: Oumuamua-7b-instruct-v2-preset-prometheus
345
+ ---
346
+ merge_method: task_arithmetic
347
+ base_model: Mistral-7B-v0.1-VE-Swallow-MS
348
+ models:
349
+ - model: nitky/Oumuamua-7b-instruct
350
+ parameters:
351
+ weight: 1.0
352
+ - model: dolphin-2.8-mistral-7b-v02-VE-Swallow-MS
353
+ parameters:
354
+ weight: 0.8
355
+ dtype: bfloat16
356
+ name: Oumuamua-7b-instruct-v2-preset-dolphin
357
+ ---
358
+ merge_method: task_arithmetic
359
+ base_model: Mistral-7B-v0.1-VE-Swallow-MS
360
+ models:
361
+ - model: nitky/Oumuamua-7b-instruct
362
+ parameters:
363
+ weight: 1.0
364
+ - model: janus-7b-VE-Swallow-MS
365
+ parameters:
366
+ weight: 0.8
367
+ dtype: bfloat16
368
+ name: Oumuamua-7b-instruct-v2-preset-janus
369
+ ---
370
+ merge_method: task_arithmetic
371
+ base_model: Mistral-7B-v0.1-VE-Swallow-MS
372
+ models:
373
+ - model: nitky/Oumuamua-7b-instruct
374
+ parameters:
375
+ weight: 1.0
376
+ - model: SELM-Zephyr-7B-iter-3-VE-Swallow-MS
377
+ parameters:
378
+ weight: 0.8
379
+ dtype: bfloat16
380
+ name: Oumuamua-7b-instruct-v2-preset-selm
381
+ ---
382
+ merge_method: task_arithmetic
383
+ base_model: Mistral-7B-v0.1-VE-Swallow-MS
384
+ models:
385
+ - model: nitky/Oumuamua-7b-instruct
386
+ parameters:
387
+ weight: 1.0
388
+ - model: RP-7b-instruct-VE-Swallow-MS
389
+ parameters:
390
+ weight: 0.8
391
+ - model: japanese-stablelm-base-gamma-7b-VE-Swallow-MS
392
+ parameters:
393
+ weight: -0.8
394
+ dtype: bfloat16
395
+ name: Oumuamua-7b-instruct-v2-preset-rp
396
+ ---
397
+ merge_method: task_arithmetic
398
+ base_model: Mistral-7B-v0.1-VE-Swallow-MS
399
+ models:
400
+ - model: nitky/Oumuamua-7b-instruct
401
+ parameters:
402
+ weight: 1.0
403
+ - model: Eurus-7b-kto-VE-Swallow-MS
404
+ parameters:
405
+ weight: 0.8
406
+ dtype: bfloat16
407
+ name: Oumuamua-7b-instruct-v2-preset-eurus
408
+ ---
409
+ merge_method: task_arithmetic
410
+ base_model: Mistral-7B-v0.1-VE-Swallow-MS
411
+ models:
412
+ - model: nitky/Oumuamua-7b-instruct
413
+ parameters:
414
+ weight: 1.0
415
+ - model: chatntq-ja-7b-v1.0-VE-Swallow-MS
416
+ parameters:
417
+ weight: 0.8
418
+ - model: japanese-stablelm-base-gamma-7b-VE-Swallow-MS
419
+ parameters:
420
+ weight: -0.8
421
+ dtype: bfloat16
422
+ name: Oumuamua-7b-instruct-v2-preset-chatntq-ja
423
+ ---
424
+ merge_method: task_arithmetic
425
+ base_model: Mistral-7B-v0.1-VE-Swallow-MS
426
+ models:
427
+ - model: nitky/Oumuamua-7b-instruct
428
+ parameters:
429
+ weight: 1.0
430
+ - model: Einstein-v6-7B-VE-Swallow-MS
431
+ parameters:
432
+ weight: 0.8
433
+ dtype: bfloat16
434
+ name: Oumuamua-7b-instruct-v2-preset-es
435
+ ---
436
+ merge_method: task_arithmetic
437
+ base_model: Mistral-7B-v0.1-VE-Swallow-MS
438
+ models:
439
+ - model: nitky/Oumuamua-7b-instruct
440
+ parameters:
441
+ weight: 1.0
442
+ - model: internistai-base-7b-v0.2-VE-Swallow-MS
443
+ parameters:
444
+ weight: 0.8
445
+ dtype: bfloat16
446
+ name: Oumuamua-7b-instruct-v2-preset-internistai
447
+ ---
448
+ merge_method: task_arithmetic
449
+ base_model: Mistral-7B-v0.1-VE-Swallow-MS
450
+ models:
451
+ - model: nitky/Oumuamua-7b-instruct
452
+ parameters:
453
+ weight: 1.0
454
+ - model: ZySec-7B-VE-Swallow-MS
455
+ parameters:
456
+ weight: 0.8
457
+ dtype: bfloat16
458
+ name: Oumuamua-7b-instruct-v2-preset-zysec
459
+ ---
460
+ merge_method: model_stock
461
+ base_model: nitky/Oumuamua-7b-base
462
+ models:
463
+ - model: nitky/Oumuamua-7b-instruct
464
+ - model: Oumuamua-7b-instruct-v2-preset-dolphin
465
+ - model: Oumuamua-7b-instruct-v2-preset-prometheus
466
+ dtype: bfloat16
467
+ name: Oumuamua-7b-instruct-v2-it-alpha
468
+ ---
469
+ merge_method: model_stock
470
+ base_model: nitky/Oumuamua-7b-base
471
+ models:
472
+ - model: nitky/Oumuamua-7b-instruct
473
+ - model: Oumuamua-7b-instruct-v2-preset-janus
474
+ - model: Oumuamua-7b-instruct-v2-preset-prometheus
475
+ dtype: bfloat16
476
+ name: Oumuamua-7b-instruct-v2-it-beta
477
+ ---
478
+ merge_method: model_stock
479
+ base_model: nitky/Oumuamua-7b-base
480
+ models:
481
+ - model: nitky/Oumuamua-7b-instruct
482
+ - model: Oumuamua-7b-instruct-v2-preset-m3
483
+ - model: Oumuamua-7b-instruct-v2-preset-prometheus
484
+ dtype: bfloat16
485
+ name: Oumuamua-7b-instruct-v2-it-gamma
486
+ ---
487
+ merge_method: model_stock
488
+ base_model: nitky/Oumuamua-7b-base
489
+ models:
490
+ - model: nitky/Oumuamua-7b-instruct
491
+ - model: Oumuamua-7b-instruct-v2-preset-selm
492
+ - model: Oumuamua-7b-instruct-v2-preset-prometheus
493
+ dtype: bfloat16
494
+ name: Oumuamua-7b-instruct-v2-it-delta
495
+ ---
496
+ merge_method: model_stock
497
+ base_model: nitky/Oumuamua-7b-base
498
+ models:
499
+ - model: nitky/Oumuamua-7b-instruct
500
+ - model: Oumuamua-7b-instruct-v2-preset-rp
501
+ - model: Oumuamua-7b-instruct-v2-preset-prometheus
502
+ dtype: bfloat16
503
+ name: Oumuamua-7b-instruct-v2-it-epsilon
504
+ ---
505
+ merge_method: model_stock
506
+ base_model: nitky/Oumuamua-7b-base
507
+ models:
508
+ - model: nitky/Oumuamua-7b-instruct
509
+ - model: Oumuamua-7b-instruct-v2-preset-eurus
510
+ - model: Oumuamua-7b-instruct-v2-preset-prometheus
511
+ dtype: bfloat16
512
+ name: Oumuamua-7b-instruct-v2-it-zeta
513
+ ---
514
+ merge_method: model_stock
515
+ base_model: nitky/Oumuamua-7b-base
516
+ models:
517
+ - model: nitky/Oumuamua-7b-instruct
518
+ - model: Oumuamua-7b-instruct-v2-preset-chatntq-ja
519
+ - model: Oumuamua-7b-instruct-v2-preset-prometheus
520
+ dtype: bfloat16
521
+ name: Oumuamua-7b-instruct-v2-it-eta
522
+ ---
523
+ merge_method: model_stock
524
+ base_model: nitky/Oumuamua-7b-base
525
+ models:
526
+ - model: nitky/Oumuamua-7b-instruct
527
+ - model: Oumuamua-7b-instruct-v2-preset-chatntq-ja
528
+ - model: Oumuamua-7b-instruct-v2-preset-prometheus
529
+ dtype: bfloat16
530
+ name: Oumuamua-7b-instruct-v2-it-eta
531
+ ---
532
+ merge_method: model_stock
533
+ base_model: nitky/Oumuamua-7b-base
534
+ models:
535
+ - model: nitky/Oumuamua-7b-instruct
536
+ - model: Oumuamua-7b-instruct-v2-preset-es
537
+ - model: Oumuamua-7b-instruct-v2-preset-prometheus
538
+ dtype: bfloat16
539
+ name: Oumuamua-7b-instruct-v2-it-theta
540
+ ---
541
+ # combine domain-specific models into one preset to minimize the negative effects
542
+ merge_method: model_stock
543
+ base_model: nitky/Oumuamua-7b-base
544
+ models:
545
+ - model: nitky/Oumuamua-7b-instruct
546
+ - model: Oumuamua-7b-instruct-v2-preset-internistai
547
+ - model: Oumuamua-7b-instruct-v2-preset-zysec
548
+ - model: Oumuamua-7b-instruct-v2-preset-prometheus
549
+ dtype: bfloat16
550
+ name: Oumuamua-7b-instruct-v2-it-omega
551
+ ---
552
+ merge_method: model_stock
553
+ base_model: nitky/Oumuamua-7b-base
554
+ models:
555
+ - model: Oumuamua-7b-instruct-v2-it-alpha
556
+ - model: Oumuamua-7b-instruct-v2-it-beta
557
+ - model: Oumuamua-7b-instruct-v2-it-gamma
558
+ - model: Oumuamua-7b-instruct-v2-it-delta
559
+ - model: Oumuamua-7b-instruct-v2-it-epsilon
560
+ - model: Oumuamua-7b-instruct-v2-it-zeta
561
+ - model: Oumuamua-7b-instruct-v2-it-eta
562
+ - model: Oumuamua-7b-instruct-v2-it-theta
563
+ - model: Oumuamua-7b-instruct-v2-it-omega
564
+ dtype: bfloat16
565
+ name: Oumuamua-7b-instruct-v2
566
+ ```