Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
tags:
|
3 |
+
- uqff
|
4 |
+
- mistral.rs
|
5 |
+
base_model: meta-llama/Llama-3.2-11B-Vision-Instruct
|
6 |
+
base_model_relation: quantized
|
7 |
+
---
|
8 |
+
|
9 |
+
<!-- Autogenerated from user input. -->
|
10 |
+
|
11 |
+
# `meta-llama/Llama-3.2-11B-Vision-Instruct`, UQFF quantization
|
12 |
+
|
13 |
+
|
14 |
+
Run with [mistral.rs](https://github.com/EricLBuehler/mistral.rs). Documentation: [UQFF docs](https://github.com/EricLBuehler/mistral.rs/blob/master/docs/UQFF.md).
|
15 |
+
|
16 |
+
1) **Flexible** ๐: Multiple quantization formats in *one* file format with *one* framework to run them all.
|
17 |
+
2) **Reliable** ๐: Compatibility ensured with *embedded* and *checked* semantic versioning information from day 1.
|
18 |
+
3) **Easy** ๐ค: Download UQFF models *easily* and *quickly* from Hugging Face, or use a local file.
|
19 |
+
3) **Customizable** ๐ ๏ธ: Make and publish your own UQFF files in minutes.
|
20 |
+
## Files
|
21 |
+
|
22 |
+
|Name|Quantization type(s)|Example|
|
23 |
+
|--|--|--|
|
24 |
+
|llama-3.2-11b-vision-q4k.uqff|Q4K|`./mistralrs-server -i plain -m meta-llama/Llama-3.2-11B-Vision-Instruct --from-uqff EricB/Llama-3.2-11B-Vision-Instruct-UQFF/llama-3.2-11b-vision-q4k.uqff`|
|
25 |
+
|llama-3.2-11b-vision-q8_0.uqff|Q8_0|`./mistralrs-server -i plain -m meta-llama/Llama-3.2-11B-Vision-Instruct --from-uqff EricB/Llama-3.2-11B-Vision-Instruct-UQFF/llama-3.2-11b-vision-q8_0.uqff`|
|