--- language: - en - es --- # Model Card for Carpincho-13b This is Carpincho-13B an Instruction-tuned LLM based on LLama-13B. It is trained to answer in colloquial spanish Argentine language. It's based on LLama-13b (https://huggingface.co/decapoda-research/llama-13b-hf). ## Model Details The model is provided in ggml format, for use with the llama.cpp CPU-only LLM inference (https://github.com/ggerganov/llama.cpp) ## Usage Clone the llama.cpp repository: git clone https://github.com/ggerganov/llama.cpp Compile the tool: make Download the file carpincho-13b-ggml-model-q4_0.bin into the llama.cpp directory and run this command: ./main -m ./carpincho-13b-ggml-model-q4_0.bin -i -ins -t 4 Change -t 4 to the number of physical CPU cores you have. This model requires at least 8GB of free RAM. No GPU is needed to run llama.cpp. ### Model Description - **Developed by:** Alfredo Ortega (@ortegaalfredo) - **Model type:** 13B LLM - **Language(s):** (NLP): English and colloquial Argentine Spanish - **License:** Free for non-commercial use, but I'm not the police. - **Finetuned from model:** https://huggingface.co/decapoda-research/llama-13b-hf ### Model Sources [optional] - **Repository:** https://huggingface.co/decapoda-research/llama-13b-hf - **Paper [optional]:** https://arxiv.org/abs/2302.13971 ## Uses This is a generic LLM chatbot that can be used to interact directly with humans. ## Bias, Risks, and Limitations This bot is uncensored and may provide shocking answers. Also it contains bias present in the training material. ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. ## Model Card Contact Contact the creator at @ortegaalfredo on twitter/github