--- license: apache-2.0 base_model: v2ray/Mixtral-8x22B-v0.1 inference: false model_creator: MaziyarPanahi model_name: Mixtral-8x22B-v0.1-GGUF pipeline_tag: text-generation quantized_by: MaziyarPanahi tags: - quantized - 2-bit - 3-bit - 4-bit - 5-bit - 6-bit - 8-bit - 16-bit - GGUF - mixtral - moe --- # Mixtral-8x22B-v0.1-GGUF in progress ... ## Load sharded model `llama_load_model_from_file` will detect the number of files and will load additional tensors from the rest of files. ``` main --model Mixtral-8x22B-v0.1.fp16-00001-of-00005.gguf -ngl 64 ```