RichardErkhov commited on
Commit
6b0a057
1 Parent(s): b60cf66

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +325 -0
README.md ADDED
@@ -0,0 +1,325 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ multimaster-7b-v6 - GGUF
11
+ - Model creator: https://huggingface.co/ibivibiv/
12
+ - Original model: https://huggingface.co/ibivibiv/multimaster-7b-v6/
13
+
14
+
15
+ | Name | Quant method | Size |
16
+ | ---- | ---- | ---- |
17
+ | [multimaster-7b-v6.Q2_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q2_K.gguf) | Q2_K | 12.04GB |
18
+ | [multimaster-7b-v6.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.IQ3_XS.gguf) | IQ3_XS | 13.48GB |
19
+ | [multimaster-7b-v6.IQ3_S.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.IQ3_S.gguf) | IQ3_S | 14.25GB |
20
+ | [multimaster-7b-v6.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q3_K_S.gguf) | Q3_K_S | 14.23GB |
21
+ | [multimaster-7b-v6.IQ3_M.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.IQ3_M.gguf) | IQ3_M | 14.49GB |
22
+ | [multimaster-7b-v6.Q3_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q3_K.gguf) | Q3_K | 15.79GB |
23
+ | [multimaster-7b-v6.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q3_K_M.gguf) | Q3_K_M | 15.79GB |
24
+ | [multimaster-7b-v6.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q3_K_L.gguf) | Q3_K_L | 17.1GB |
25
+ | [multimaster-7b-v6.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.IQ4_XS.gguf) | IQ4_XS | 17.79GB |
26
+ | [multimaster-7b-v6.Q4_0.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q4_0.gguf) | Q4_0 | 18.6GB |
27
+ | [multimaster-7b-v6.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.IQ4_NL.gguf) | IQ4_NL | 18.77GB |
28
+ | [multimaster-7b-v6.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q4_K_S.gguf) | Q4_K_S | 18.76GB |
29
+ | [multimaster-7b-v6.Q4_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q4_K.gguf) | Q4_K | 19.96GB |
30
+ | [multimaster-7b-v6.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q4_K_M.gguf) | Q4_K_M | 19.96GB |
31
+ | [multimaster-7b-v6.Q4_1.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q4_1.gguf) | Q4_1 | 20.65GB |
32
+ | [multimaster-7b-v6.Q5_0.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q5_0.gguf) | Q5_0 | 22.7GB |
33
+ | [multimaster-7b-v6.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q5_K_S.gguf) | Q5_K_S | 22.7GB |
34
+ | [multimaster-7b-v6.Q5_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q5_K.gguf) | Q5_K | 23.41GB |
35
+ | [multimaster-7b-v6.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q5_K_M.gguf) | Q5_K_M | 23.41GB |
36
+ | [multimaster-7b-v6.Q5_1.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q5_1.gguf) | Q5_1 | 24.76GB |
37
+ | [multimaster-7b-v6.Q6_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q6_K.gguf) | Q6_K | 27.07GB |
38
+ | [multimaster-7b-v6.Q8_0.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-v6-gguf/blob/main/multimaster-7b-v6.Q8_0.gguf) | Q8_0 | 35.06GB |
39
+
40
+
41
+
42
+
43
+ Original model description:
44
+ ---
45
+ language:
46
+ - en
47
+ license: apache-2.0
48
+ library_name: transformers
49
+ model-index:
50
+ - name: multimaster-7b-v6
51
+ results:
52
+ - task:
53
+ type: text-generation
54
+ name: Text Generation
55
+ dataset:
56
+ name: AI2 Reasoning Challenge (25-Shot)
57
+ type: ai2_arc
58
+ config: ARC-Challenge
59
+ split: test
60
+ args:
61
+ num_few_shot: 25
62
+ metrics:
63
+ - type: acc_norm
64
+ value: 72.78
65
+ name: normalized accuracy
66
+ source:
67
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ibivibiv/multimaster-7b-v6
68
+ name: Open LLM Leaderboard
69
+ - task:
70
+ type: text-generation
71
+ name: Text Generation
72
+ dataset:
73
+ name: HellaSwag (10-Shot)
74
+ type: hellaswag
75
+ split: validation
76
+ args:
77
+ num_few_shot: 10
78
+ metrics:
79
+ - type: acc_norm
80
+ value: 88.77
81
+ name: normalized accuracy
82
+ source:
83
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ibivibiv/multimaster-7b-v6
84
+ name: Open LLM Leaderboard
85
+ - task:
86
+ type: text-generation
87
+ name: Text Generation
88
+ dataset:
89
+ name: MMLU (5-Shot)
90
+ type: cais/mmlu
91
+ config: all
92
+ split: test
93
+ args:
94
+ num_few_shot: 5
95
+ metrics:
96
+ - type: acc
97
+ value: 64.74
98
+ name: accuracy
99
+ source:
100
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ibivibiv/multimaster-7b-v6
101
+ name: Open LLM Leaderboard
102
+ - task:
103
+ type: text-generation
104
+ name: Text Generation
105
+ dataset:
106
+ name: TruthfulQA (0-shot)
107
+ type: truthful_qa
108
+ config: multiple_choice
109
+ split: validation
110
+ args:
111
+ num_few_shot: 0
112
+ metrics:
113
+ - type: mc2
114
+ value: 70.89
115
+ source:
116
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ibivibiv/multimaster-7b-v6
117
+ name: Open LLM Leaderboard
118
+ - task:
119
+ type: text-generation
120
+ name: Text Generation
121
+ dataset:
122
+ name: Winogrande (5-shot)
123
+ type: winogrande
124
+ config: winogrande_xl
125
+ split: validation
126
+ args:
127
+ num_few_shot: 5
128
+ metrics:
129
+ - type: acc
130
+ value: 86.42
131
+ name: accuracy
132
+ source:
133
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ibivibiv/multimaster-7b-v6
134
+ name: Open LLM Leaderboard
135
+ - task:
136
+ type: text-generation
137
+ name: Text Generation
138
+ dataset:
139
+ name: GSM8k (5-shot)
140
+ type: gsm8k
141
+ config: main
142
+ split: test
143
+ args:
144
+ num_few_shot: 5
145
+ metrics:
146
+ - type: acc
147
+ value: 70.36
148
+ name: accuracy
149
+ source:
150
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ibivibiv/multimaster-7b-v6
151
+ name: Open LLM Leaderboard
152
+ ---
153
+
154
+ # Multi Master 7Bx5 v6
155
+
156
+ ![img](./multimaster.png)
157
+
158
+ A quick multi-disciplinary moe model. This is part of a series of models built to test the gate tuning for mixtral style moe models.
159
+
160
+ # Prompting
161
+
162
+ ## Prompt Template for alpaca style
163
+
164
+ ```
165
+ ### Instruction:
166
+
167
+ <prompt> (without the <>)
168
+
169
+ ### Response:
170
+ ```
171
+
172
+ ## Sample Code
173
+
174
+ ```python
175
+ import torch
176
+ from transformers import AutoModelForCausalLM, AutoTokenizer
177
+
178
+ torch.set_default_device("cuda")
179
+
180
+ model = AutoModelForCausalLM.from_pretrained("ibivibiv/multimaster-7b-v6", torch_dtype="auto", device_config='auto')
181
+ tokenizer = AutoTokenizer.from_pretrained("ibivibiv/multimaster-7b-v6")
182
+
183
+ inputs = tokenizer("### Instruction: Who would when in an arm wrestling match between Abraham Lincoln and Chuck Norris?\nA. Abraham Lincoln \nB. Chuck Norris\n### Response:\n", return_tensors="pt", return_attention_mask=False)
184
+
185
+ outputs = model.generate(**inputs, max_length=200)
186
+ text = tokenizer.batch_decode(outputs)[0]
187
+ print(text)
188
+ ```
189
+
190
+ # Model Details
191
+ * **Trained by**: [ibivibiv](https://huggingface.co/ibivibiv)
192
+ * **Library**: [HuggingFace Transformers](https://github.com/huggingface/transformers)
193
+ * **Model type:** **multimaster-7b** is a lora tuned version of openchat/openchat-3.5-0106 with the adapter merged back into the main model
194
+ * **Language(s)**: English
195
+ * **Purpose**: This model is a focus on multi-disciplinary model tuning
196
+
197
+ # Benchmark Scores
198
+
199
+ coming soon
200
+
201
+ ## Citations
202
+
203
+ ```
204
+ @misc{open-llm-leaderboard,
205
+ author = {Edward Beeching and Clémentine Fourrier and Nathan Habib and Sheon Han and Nathan Lambert and Nazneen Rajani and Omar Sanseviero and Lewis Tunstall and Thomas Wolf},
206
+ title = {Open LLM Leaderboard},
207
+ year = {2023},
208
+ publisher = {Hugging Face},
209
+ howpublished = "\url{https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard}"
210
+ }
211
+ ```
212
+ ```
213
+ @software{eval-harness,
214
+ author = {Gao, Leo and
215
+ Tow, Jonathan and
216
+ Biderman, Stella and
217
+ Black, Sid and
218
+ DiPofi, Anthony and
219
+ Foster, Charles and
220
+ Golding, Laurence and
221
+ Hsu, Jeffrey and
222
+ McDonell, Kyle and
223
+ Muennighoff, Niklas and
224
+ Phang, Jason and
225
+ Reynolds, Laria and
226
+ Tang, Eric and
227
+ Thite, Anish and
228
+ Wang, Ben and
229
+ Wang, Kevin and
230
+ Zou, Andy},
231
+ title = {A framework for few-shot language model evaluation},
232
+ month = sep,
233
+ year = 2021,
234
+ publisher = {Zenodo},
235
+ version = {v0.0.1},
236
+ doi = {10.5281/zenodo.5371628},
237
+ url = {https://doi.org/10.5281/zenodo.5371628}
238
+ }
239
+ ```
240
+ ```
241
+ @misc{clark2018think,
242
+ title={Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge},
243
+ author={Peter Clark and Isaac Cowhey and Oren Etzioni and Tushar Khot and Ashish Sabharwal and Carissa Schoenick and Oyvind Tafjord},
244
+ year={2018},
245
+ eprint={1803.05457},
246
+ archivePrefix={arXiv},
247
+ primaryClass={cs.AI}
248
+ }
249
+ ```
250
+ ```
251
+ @misc{zellers2019hellaswag,
252
+ title={HellaSwag: Can a Machine Really Finish Your Sentence?},
253
+ author={Rowan Zellers and Ari Holtzman and Yonatan Bisk and Ali Farhadi and Yejin Choi},
254
+ year={2019},
255
+ eprint={1905.07830},
256
+ archivePrefix={arXiv},
257
+ primaryClass={cs.CL}
258
+ }
259
+ ```
260
+ ```
261
+ @misc{hendrycks2021measuring,
262
+ title={Measuring Massive Multitask Language Understanding},
263
+ author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},
264
+ year={2021},
265
+ eprint={2009.03300},
266
+ archivePrefix={arXiv},
267
+ primaryClass={cs.CY}
268
+ }
269
+ ```
270
+ ```
271
+ @misc{lin2022truthfulqa,
272
+ title={TruthfulQA: Measuring How Models Mimic Human Falsehoods},
273
+ author={Stephanie Lin and Jacob Hilton and Owain Evans},
274
+ year={2022},
275
+ eprint={2109.07958},
276
+ archivePrefix={arXiv},
277
+ primaryClass={cs.CL}
278
+ }
279
+ ```
280
+ ```
281
+ @misc{DBLP:journals/corr/abs-1907-10641,
282
+ title={{WINOGRANDE:} An Adversarial Winograd Schema Challenge at Scale},
283
+ author={Keisuke Sakaguchi and Ronan Le Bras and Chandra Bhagavatula and Yejin Choi},
284
+ year={2019},
285
+ eprint={1907.10641},
286
+ archivePrefix={arXiv},
287
+ primaryClass={cs.CL}
288
+ }
289
+ ```
290
+ ```
291
+ @misc{DBLP:journals/corr/abs-2110-14168,
292
+ title={Training Verifiers to Solve Math Word Problems},
293
+ author={Karl Cobbe and
294
+ Vineet Kosaraju and
295
+ Mohammad Bavarian and
296
+ Mark Chen and
297
+ Heewoo Jun and
298
+ Lukasz Kaiser and
299
+ Matthias Plappert and
300
+ Jerry Tworek and
301
+ Jacob Hilton and
302
+ Reiichiro Nakano and
303
+ Christopher Hesse and
304
+ John Schulman},
305
+ year={2021},
306
+ eprint={2110.14168},
307
+ archivePrefix={arXiv},
308
+ primaryClass={cs.CL}
309
+ }
310
+ ```
311
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
312
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_ibivibiv__multimaster-7b-v6)
313
+
314
+ | Metric |Value|
315
+ |---------------------------------|----:|
316
+ |Avg. |75.66|
317
+ |AI2 Reasoning Challenge (25-Shot)|72.78|
318
+ |HellaSwag (10-Shot) |88.77|
319
+ |MMLU (5-Shot) |64.74|
320
+ |TruthfulQA (0-shot) |70.89|
321
+ |Winogrande (5-shot) |86.42|
322
+ |GSM8k (5-shot) |70.36|
323
+
324
+
325
+