Update README.md
Browse files
README.md
CHANGED
@@ -72,9 +72,9 @@ Please see https://github.com/cmp-nct/ggllm.cpp for further details and instruct
|
|
72 |
| Name | Quant method | Bits | Size | Max RAM required | Use case |
|
73 |
| ---- | ---- | ---- | ---- | ---- | ----- |
|
74 |
| falcon40b-instruct.ggccv1.q2_K.bin | q2_K | 2 | 13.74 GB | 16.24 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.vw and feed_forward.w2 tensors, GGML_TYPE_Q2_K for the other tensors. |
|
75 |
-
| falcon40b-instruct.ggccv1.
|
76 |
-
| falcon40b-instruct.ggccv1.
|
77 |
-
| falcon40b-instruct.ggccv1.
|
78 |
| falcon40b-instruct.ggccv1.q6_K.bin | q6_K | 6 | 34.33 GB | 36.83 GB | New k-quant method. Uses GGML_TYPE_Q8_K - 6-bit quantization - for all tensors |
|
79 |
| falcon40b-instruct.ggccv1.q8_0.bin | q8_0 | 8 | 44.46 GB | 46.96 GB | Original llama.cpp quant method, 8-bit. Almost indistinguishable from float16. High resource use and slow. Not recommended for most users. |
|
80 |
|
|
|
72 |
| Name | Quant method | Bits | Size | Max RAM required | Use case |
|
73 |
| ---- | ---- | ---- | ---- | ---- | ----- |
|
74 |
| falcon40b-instruct.ggccv1.q2_K.bin | q2_K | 2 | 13.74 GB | 16.24 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.vw and feed_forward.w2 tensors, GGML_TYPE_Q2_K for the other tensors. |
|
75 |
+
| falcon40b-instruct.ggccv1.q3_K.bin | q3_K_S | 3 | 17.98 GB | 20.48 GB | New k-quant method. Uses GGML_TYPE_Q3_K for all tensors |
|
76 |
+
| falcon40b-instruct.ggccv1.q4_K.bin | q4_K_S | 4 | 23.54 GB | 26.04 GB | New k-quant method. Uses GGML_TYPE_Q4_K for all tensors |
|
77 |
+
| falcon40b-instruct.ggccv1.q5_K.bin | q5_K_S | 5 | 28.77 GB | 31.27 GB | New k-quant method. Uses GGML_TYPE_Q5_K for all tensors |
|
78 |
| falcon40b-instruct.ggccv1.q6_K.bin | q6_K | 6 | 34.33 GB | 36.83 GB | New k-quant method. Uses GGML_TYPE_Q8_K - 6-bit quantization - for all tensors |
|
79 |
| falcon40b-instruct.ggccv1.q8_0.bin | q8_0 | 8 | 44.46 GB | 46.96 GB | Original llama.cpp quant method, 8-bit. Almost indistinguishable from float16. High resource use and slow. Not recommended for most users. |
|
80 |
|