Edit model card

Model Card for Carpincho-13b

This is Carpincho-13B an Instruction-tuned LLM based on LLama-13B. It is trained to answer in colloquial spanish Argentine language. It's based on LLama-13b (https://huggingface.co/decapoda-research/llama-13b-hf).

Model Details

The model is provided in ggml format, for use with the llama.cpp CPU-only LLM inference (https://github.com/ggerganov/llama.cpp)

Usage

Clone the llama.cpp repository:

git clone https://github.com/ggerganov/llama.cpp

Compile the tool:

make

Download the file carpincho-13b-ggml-model-q4_0.bin into the llama.cpp directory and run this command:

./main -m ./carpincho-13b-ggml-model-q4_0.bin -i -ins -t 4

Change -t 4 to the number of physical CPU cores you have.

This model requires at least 8GB of free RAM. No GPU is needed to run llama.cpp.

Model Description

  • Developed by: Alfredo Ortega (@ortegaalfredo)
  • Model type: 13B LLM
  • Language(s): (NLP): English and colloquial Argentine Spanish
  • License: Free for non-commercial use, but I'm not the police.
  • Finetuned from model: https://huggingface.co/decapoda-research/llama-13b-hf

Model Sources [optional]

Uses

This is a generic LLM chatbot that can be used to interact directly with humans.

Bias, Risks, and Limitations

This bot is uncensored and may provide shocking answers. Also it contains bias present in the training material.

Recommendations

Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model.

Model Card Contact

Contact the creator at @ortegaalfredo on twitter/github

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference API
Unable to determine this model's library. Check the docs .