InferenceIllusionist
commited on
Commit
•
59a7800
1
Parent(s):
87461f4
Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,21 @@
|
|
1 |
---
|
|
|
|
|
|
|
|
|
2 |
license: cc-by-4.0
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
tags:
|
3 |
+
- gguf
|
4 |
+
- mistral
|
5 |
+
- conversational
|
6 |
license: cc-by-4.0
|
7 |
---
|
8 |
+
<img src="https://i.imgur.com/P68dXux.png" width="400"/>
|
9 |
+
|
10 |
+
|
11 |
+
# Open_Gpt4_8x7B_v0.2 iMat GGUF
|
12 |
+
|
13 |
+
|
14 |
+
Open_Gpt4_8x7B_v0.2 iMat GGUF quantized from fp16 with love.
|
15 |
+
* Quantizations made possible using mixtral-8x7b-instruct-v0.1.imatrix file from [this](https://huggingface.co/datasets/ikawrakow/imatrix-from-wiki-train) repo (special thanks to [ikawrakow](https://huggingface.co/ikawrakow) again)
|
16 |
+
|
17 |
+
<b>Legacy quants (i.e. Q8, Q5_K_M) in this repo have all been enhanced with importance matrix calculation. These quants show improved KL-Divergence over their static counterparts.</b>
|
18 |
+
|
19 |
+
All files have been tested for your safety and convenience. No need to clone the entire repo, just pick the quant that's right for you.
|
20 |
+
|
21 |
+
For more information on latest iMatrix quants see this PR - https://github.com/ggerganov/llama.cpp/pull/5747
|