--- base_model: "TinyLlama/TinyLlama-1.1B-Chat-v1.0" language: - en tags: - transformers - safetensors - llama - text-generation - conversational - en - dataset:cerebras/SlimPajama-627B - dataset:bigcode/starcoderdata - dataset:HuggingFaceH4/ultrachat_200k - dataset:HuggingFaceH4/ultrafeedback_binarized - license:apache-2.0 - autotrain_compatible - endpoints_compatible - text-generation-inference - region:us license: "apache-2.0" inference: false datasets: - cerebras/SlimPajama-627B - bigcode/starcoderdata - HuggingFaceH4/ultrachat_200k - HuggingFaceH4/ultrafeedback_binarized quantized_by: pbatra --- # TinyLlama-1.1B-Chat-v1.0 This repository contains quantized versions of the model from the original repository: [TinyLlama/TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0). | Name | Quantization Method | Size (GB) | |------|---------------------|-----------| | tinyllama-1.1b-chat-v1.0.fp16.gguf | FP16 | 2.05 | | tinyllama-1.1b-chat-v1.0.fp32.gguf | FP32 | 4.10 | | tinyllama-1.1b-chat-v1.0.Q2_K.gguf | q2_k | 0.40 | | tinyllama-1.1b-chat-v1.0.Q3_K_S.gguf | q3_k_s | 0.47 | | tinyllama-1.1b-chat-v1.0.Q3_K_M.gguf | q3_k_m | 0.51 | | tinyllama-1.1b-chat-v1.0.Q3_K_L.gguf | q3_k_l | 0.55 | | tinyllama-1.1b-chat-v1.0.Q4_0.gguf | q4_0 | 0.59 | | tinyllama-1.1b-chat-v1.0.Q4_K_S.gguf | q4_k_s | 0.60 | | tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf | q4_k_m | 0.62 | | tinyllama-1.1b-chat-v1.0.Q5_0.gguf | q5_0 | 0.71 | | tinyllama-1.1b-chat-v1.0.Q5_K_S.gguf | q5_k_s | 0.71 | | tinyllama-1.1b-chat-v1.0.Q5_K_M.gguf | q5_k_m | 0.73 | | tinyllama-1.1b-chat-v1.0.Q6_K.gguf | q6_k | 0.84 | | tinyllama-1.1b-chat-v1.0.Q8_0.gguf | q8_0 | 1.09 |