NeoChen1024
commited on
Commit
•
09716bf
1
Parent(s):
18b6153
Update README.md
Browse files
README.md
CHANGED
@@ -4,5 +4,5 @@ base_model:
|
|
4 |
- cognitivecomputations/dolphin-2.7-mixtral-8x7b
|
5 |
---
|
6 |
|
7 |
-
GGUF IQ3_M quant of cognitivecomputations/dolphin-2.7-mixtral-8x7b (non-imatrix)
|
8 |
It fits into 24GiB VRAM with 32768 context (@ 8bit KV cache quantization).
|
|
|
4 |
- cognitivecomputations/dolphin-2.7-mixtral-8x7b
|
5 |
---
|
6 |
|
7 |
+
GGUF IQ3_M quant of cognitivecomputations/dolphin-2.7-mixtral-8x7b (both non-imatrix and imatrix)
|
8 |
It fits into 24GiB VRAM with 32768 context (@ 8bit KV cache quantization).
|