llama 65B ggml model weight running alpaca.cpp
make 65B ggml story
1. clone 65B model data
git clone https://huggingface.co/datasets/nyanko7/LLaMA-65B/
2. clone alpaca.cpp
git clone https://github.com/antimatter15/alpaca.cpp
3. weight quantize.sh
mv LLaMA-65B/tokenizer.model ./
python convert-pth-to-ggml.py ../LLaMA-65B/ 1
cd alpaca.cpp
mkdir -p models/65B
mv ../LLaMA-65B/ggml-model-f16.bin models/65B/
mv ../LLaMA-65B/ggml-model-f16.bin.* models/65B/
bash quantize.sh 65B
4. upload weight file
Upload is slower. The upload is taking almost 2 days, I decided to curve the upload
I using colab and huggingface api upload
run
git clone https://github.com/antimatter15/
./chat -m alpaca.cpp_65b_ggml/ggml-model-q4_0.bin