RichardErkhov commited on
Commit
1c514a7
1 Parent(s): 66ebb9f

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +195 -0
README.md ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ NeuralMona_MoE-4x7B - GGUF
11
+ - Model creator: https://huggingface.co/CultriX/
12
+ - Original model: https://huggingface.co/CultriX/NeuralMona_MoE-4x7B/
13
+
14
+
15
+ | Name | Quant method | Size |
16
+ | ---- | ---- | ---- |
17
+ | [NeuralMona_MoE-4x7B.Q2_K.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q2_K.gguf) | Q2_K | 8.24GB |
18
+ | [NeuralMona_MoE-4x7B.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.IQ3_XS.gguf) | IQ3_XS | 9.21GB |
19
+ | [NeuralMona_MoE-4x7B.IQ3_S.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.IQ3_S.gguf) | IQ3_S | 9.73GB |
20
+ | [NeuralMona_MoE-4x7B.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q3_K_S.gguf) | Q3_K_S | 9.72GB |
21
+ | [NeuralMona_MoE-4x7B.IQ3_M.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.IQ3_M.gguf) | IQ3_M | 9.92GB |
22
+ | [NeuralMona_MoE-4x7B.Q3_K.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q3_K.gguf) | Q3_K | 10.79GB |
23
+ | [NeuralMona_MoE-4x7B.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q3_K_M.gguf) | Q3_K_M | 10.79GB |
24
+ | [NeuralMona_MoE-4x7B.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q3_K_L.gguf) | Q3_K_L | 11.68GB |
25
+ | [NeuralMona_MoE-4x7B.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.IQ4_XS.gguf) | IQ4_XS | 12.15GB |
26
+ | [NeuralMona_MoE-4x7B.Q4_0.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q4_0.gguf) | Q4_0 | 12.69GB |
27
+ | [NeuralMona_MoE-4x7B.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.IQ4_NL.gguf) | IQ4_NL | 12.81GB |
28
+ | [NeuralMona_MoE-4x7B.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q4_K_S.gguf) | Q4_K_S | 12.8GB |
29
+ | [NeuralMona_MoE-4x7B.Q4_K.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q4_K.gguf) | Q4_K | 13.61GB |
30
+ | [NeuralMona_MoE-4x7B.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q4_K_M.gguf) | Q4_K_M | 13.61GB |
31
+ | [NeuralMona_MoE-4x7B.Q4_1.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q4_1.gguf) | Q4_1 | 14.09GB |
32
+ | [NeuralMona_MoE-4x7B.Q5_0.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q5_0.gguf) | Q5_0 | 15.48GB |
33
+ | [NeuralMona_MoE-4x7B.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q5_K_S.gguf) | Q5_K_S | 15.48GB |
34
+ | [NeuralMona_MoE-4x7B.Q5_K.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q5_K.gguf) | Q5_K | 15.96GB |
35
+ | [NeuralMona_MoE-4x7B.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q5_K_M.gguf) | Q5_K_M | 15.96GB |
36
+ | [NeuralMona_MoE-4x7B.Q5_1.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q5_1.gguf) | Q5_1 | 16.88GB |
37
+ | [NeuralMona_MoE-4x7B.Q6_K.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q6_K.gguf) | Q6_K | 18.46GB |
38
+ | [NeuralMona_MoE-4x7B.Q8_0.gguf](https://huggingface.co/RichardErkhov/CultriX_-_NeuralMona_MoE-4x7B-gguf/blob/main/NeuralMona_MoE-4x7B.Q8_0.gguf) | Q8_0 | 23.9GB |
39
+
40
+
41
+
42
+
43
+ Original model description:
44
+ ---
45
+ license: apache-2.0
46
+ tags:
47
+ - moe
48
+ - frankenmoe
49
+ - merge
50
+ - mergekit
51
+ - lazymergekit
52
+ - CultriX/MonaTrix-v4
53
+ - mlabonne/OmniTruthyBeagle-7B-v0
54
+ - CultriX/MoNeuTrix-7B-v1
55
+ - paulml/OmniBeagleSquaredMBX-v3-7B
56
+ base_model:
57
+ - CultriX/MonaTrix-v4
58
+ - mlabonne/OmniTruthyBeagle-7B-v0
59
+ - CultriX/MoNeuTrix-7B-v1
60
+ - paulml/OmniBeagleSquaredMBX-v3-7B
61
+ ---
62
+
63
+ # NeuralMona_MoE-4x7B
64
+
65
+ NeuralMona_MoE-4x7B is a Mixture of Experts (MoE) made with the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
66
+ * [CultriX/MonaTrix-v4](https://huggingface.co/CultriX/MonaTrix-v4)
67
+ * [mlabonne/OmniTruthyBeagle-7B-v0](https://huggingface.co/mlabonne/OmniTruthyBeagle-7B-v0)
68
+ * [CultriX/MoNeuTrix-7B-v1](https://huggingface.co/CultriX/MoNeuTrix-7B-v1)
69
+ * [paulml/OmniBeagleSquaredMBX-v3-7B](https://huggingface.co/paulml/OmniBeagleSquaredMBX-v3-7B)
70
+
71
+ ## 🧩 Configuration
72
+
73
+ ```yaml
74
+ base_model: CultriX/MonaTrix-v4
75
+ dtype: bfloat16
76
+ experts:
77
+ - source_model: "CultriX/MonaTrix-v4" # Historical Analysis, Geopolitics, and Economic Evaluation
78
+ positive_prompts:
79
+ - "Historic analysis"
80
+ - "Geopolitical impacts"
81
+ - "Evaluate significance"
82
+ - "Predict impact"
83
+ - "Assess consequences"
84
+ - "Discuss implications"
85
+ - "Explain geopolitical"
86
+ - "Analyze historical"
87
+ - "Examine economic"
88
+ - "Evaluate role"
89
+ - "Analyze importance"
90
+ - "Discuss cultural impact"
91
+ - "Discuss historical"
92
+ negative_prompts:
93
+ - "Compose"
94
+ - "Translate"
95
+ - "Debate"
96
+ - "Solve math"
97
+ - "Analyze data"
98
+ - "Forecast"
99
+ - "Predict"
100
+ - "Process"
101
+ - "Coding"
102
+ - "Programming"
103
+ - "Code"
104
+ - "Datascience"
105
+ - "Cryptography"
106
+
107
+ - source_model: "mlabonne/OmniTruthyBeagle-7B-v0" # Multilingual Communication and Cultural Insights
108
+ positive_prompts:
109
+ - "Describe cultural"
110
+ - "Explain in language"
111
+ - "Translate"
112
+ - "Compare cultural differences"
113
+ - "Discuss cultural impact"
114
+ - "Narrate in language"
115
+ - "Explain impact on culture"
116
+ - "Discuss national identity"
117
+ - "Describe cultural significance"
118
+ - "Narrate cultural"
119
+ - "Discuss folklore"
120
+ negative_prompts:
121
+ - "Compose"
122
+ - "Debate"
123
+ - "Solve math"
124
+ - "Analyze data"
125
+ - "Forecast"
126
+ - "Predict"
127
+ - "Coding"
128
+ - "Programming"
129
+ - "Code"
130
+ - "Datascience"
131
+ - "Cryptography"
132
+
133
+ - source_model: "CultriX/MoNeuTrix-7B-v1" # Problem Solving, Innovation, and Creative Thinking
134
+ positive_prompts:
135
+ - "Devise strategy"
136
+ - "Imagine society"
137
+ - "Invent device"
138
+ - "Design concept"
139
+ - "Propose theory"
140
+ - "Reason math"
141
+ - "Develop strategy"
142
+ - "Invent"
143
+ negative_prompts:
144
+ - "Translate"
145
+ - "Discuss"
146
+ - "Debate"
147
+ - "Summarize"
148
+ - "Explain"
149
+ - "Detail"
150
+ - "Compose"
151
+
152
+ - source_model: "paulml/OmniBeagleSquaredMBX-v3-7B" # Explaining Scientific Phenomena and Principles
153
+ positive_prompts:
154
+ - "Explain scientific"
155
+ - "Discuss impact"
156
+ - "Analyze potential"
157
+ - "Elucidate significance"
158
+ - "Summarize findings"
159
+ - "Detail explanation"
160
+ negative_prompts:
161
+ - "Cultural significance"
162
+ - "Engage in creative writing"
163
+ - "Perform subjective judgment tasks"
164
+ - "Discuss cultural traditions"
165
+ - "Write review"
166
+ - "Design"
167
+ - "Create"
168
+ - "Narrate"
169
+ - "Discuss"
170
+ ```
171
+
172
+ ## 💻 Usage
173
+
174
+ ```python
175
+ !pip install -qU transformers bitsandbytes accelerate
176
+
177
+ from transformers import AutoTokenizer
178
+ import transformers
179
+ import torch
180
+
181
+ model = "CultriX/NeuralMona_MoE-4x7B"
182
+
183
+ tokenizer = AutoTokenizer.from_pretrained(model)
184
+ pipeline = transformers.pipeline(
185
+ "text-generation",
186
+ model=model,
187
+ model_kwargs={"torch_dtype": torch.float16, "load_in_4bit": True},
188
+ )
189
+
190
+ messages = [{"role": "user", "content": "Explain what a Mixture of Experts is in less than 100 words."}]
191
+ prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
192
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
193
+ print(outputs[0]["generated_text"])
194
+ ```
195
+