Upload FLAMA-0.1-3B_q4_k_m.gguf with huggingface_hub
Browse files- .gitattributes +1 -0
- FLAMA-0.1-3B_q4_k_m.gguf +3 -0
.gitattributes
CHANGED
@@ -40,3 +40,4 @@ FLAMA-0.1-3B_q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
|
40 |
FLAMA-0.1-3B_q5_0.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
FLAMA-0.1-3B_q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
flama.png filter=lfs diff=lfs merge=lfs -text
|
|
|
|
40 |
FLAMA-0.1-3B_q5_0.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
FLAMA-0.1-3B_q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
flama.png filter=lfs diff=lfs merge=lfs -text
|
43 |
+
FLAMA-0.1-3B_q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
FLAMA-0.1-3B_q4_k_m.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:91416b203a57cbd5cf43e46a2a5a7aa2e593137ba662b67c8b35c76b72f4a699
|
3 |
+
size 2580116128
|