Alpacino-13b-ggml / README.md
verymuchawful
Update README.md
3bb4765
|
raw
history blame
345 Bytes
metadata
inference: false

GGML conversion of https://huggingface.co/digitous/Alpacino13b using https://github.com/ggerganov/llama.cpp/pull/896. (Edited to write the model with ftype 2 so it won't be incorrectly identified as 4 - mostly q4_1 some f16.)

GPTQ(cuda) quantization available here: https://huggingface.co/gozfarb/alpacino-13b-4bit-128g