InferenceIllusionist
commited on
Commit
•
acc1f2a
1
Parent(s):
081948f
Update README.md
Browse files
README.md
CHANGED
@@ -30,7 +30,7 @@ PROUDLY PRESENTS
|
|
30 |
|
31 |
Quantized from fp16 with love.
|
32 |
* Quantizations made possible using mixtral-8x7b-instruct-v0.1.imatrix file from [this](https://huggingface.co/datasets/ikawrakow/imatrix-from-wiki-train) repo (special thanks to [ikawrakow](https://huggingface.co/ikawrakow) again)
|
33 |
-
* An analysis was run on mixtral-8x7b.imatrix that showed worse KL-Divergence than mixtral-8x7b-instruct-v0.1, hence the latter was used for the
|
34 |
* For a brief rundown of iMatrix quant performance please see this [PR](https://github.com/ggerganov/llama.cpp/pull/5747)
|
35 |
|
36 |
<i>All quants are verified working prior to uploading to repo for your safety and convenience. </i>
|
|
|
30 |
|
31 |
Quantized from fp16 with love.
|
32 |
* Quantizations made possible using mixtral-8x7b-instruct-v0.1.imatrix file from [this](https://huggingface.co/datasets/ikawrakow/imatrix-from-wiki-train) repo (special thanks to [ikawrakow](https://huggingface.co/ikawrakow) again)
|
33 |
+
* An analysis was run on mixtral-8x7b.imatrix that showed worse KL-Divergence than mixtral-8x7b-instruct-v0.1, hence the latter was used for the final quantization.
|
34 |
* For a brief rundown of iMatrix quant performance please see this [PR](https://github.com/ggerganov/llama.cpp/pull/5747)
|
35 |
|
36 |
<i>All quants are verified working prior to uploading to repo for your safety and convenience. </i>
|