New GGMLv3 format for breaking llama.cpp change May 19th commit 2d5db48
Browse files
README.md
CHANGED
@@ -32,7 +32,7 @@ This repo is the result of quantising to 4-bit, 5-bit and 8-bit GGML for CPU (+C
|
|
32 |
|
33 |
llama.cpp recently made another breaking change to its quantisation methods - https://github.com/ggerganov/llama.cpp/pull/1508
|
34 |
|
35 |
-
I have quantised the GGML files in this repo with the latest version. Therefore you will require llama.cpp compiled on May
|
36 |
|
37 |
For files compatible with the previous version of llama.cpp, please see branch `previous_llama_ggmlv2`.
|
38 |
|
|
|
32 |
|
33 |
llama.cpp recently made another breaking change to its quantisation methods - https://github.com/ggerganov/llama.cpp/pull/1508
|
34 |
|
35 |
+
I have quantised the GGML files in this repo with the latest version. Therefore you will require llama.cpp compiled on May 19th or later (commit `2d5db48` or later) to use them.
|
36 |
|
37 |
For files compatible with the previous version of llama.cpp, please see branch `previous_llama_ggmlv2`.
|
38 |
|