InferenceIllusionist
commited on
Commit
•
6ab41c0
1
Parent(s):
c359203
Update README.md
Browse files
README.md
CHANGED
@@ -26,7 +26,7 @@ license: apache-2.0
|
|
26 |
|
27 |
Other front-ends like the main branch of llama.cpp, kobold.cpp, and text-generation-web-ui may not work as intended</b>
|
28 |
|
29 |
-
Quantized from Mistral-Nemo-Instruct-2407 fp16
|
30 |
* Weighted quantizations were creating using fp16 GGUF and groups_merged.txt in 92 chunks and n_ctx=512
|
31 |
* Static fp16 will also be included in repo
|
32 |
|
|
|
26 |
|
27 |
Other front-ends like the main branch of llama.cpp, kobold.cpp, and text-generation-web-ui may not work as intended</b>
|
28 |
|
29 |
+
Quantized from Mistral-Nemo-Instruct-2407 fp16
|
30 |
* Weighted quantizations were creating using fp16 GGUF and groups_merged.txt in 92 chunks and n_ctx=512
|
31 |
* Static fp16 will also be included in repo
|
32 |
|