Update README.md
Browse files
README.md
CHANGED
@@ -39,11 +39,7 @@ Please note that these GGMLs are **not compatible with llama.cpp, or currently w
|
|
39 |
|
40 |
[KoboldCpp](https://github.com/LostRuins/koboldcpp) just added GPU accelerated (OpenCL) support for MPT models, so that is the client I recommend using for these models.
|
41 |
|
42 |
-
**Note**:
|
43 |
-
|
44 |
-
To resolve this, add argument `--forceversion 500`
|
45 |
-
|
46 |
-
This should be fixed in the next release of KoboldCpp, so if you are running a version later than 1.32 it should not be necessary.
|
47 |
|
48 |
## Repositories available
|
49 |
|
|
|
39 |
|
40 |
[KoboldCpp](https://github.com/LostRuins/koboldcpp) just added GPU accelerated (OpenCL) support for MPT models, so that is the client I recommend using for these models.
|
41 |
|
42 |
+
**Note**: Please use version 1.32.1 or later of KoboldCpp, as there was a bug in 1.32 that affected loading of MPT models.
|
|
|
|
|
|
|
|
|
43 |
|
44 |
## Repositories available
|
45 |
|