Upload FLAMA-0.1-3B_q6_k.gguf with huggingface_hub
Browse files- .gitattributes +1 -0
- FLAMA-0.1-3B_q6_k.gguf +3 -0
.gitattributes
CHANGED
@@ -41,3 +41,4 @@ FLAMA-0.1-3B_q5_0.gguf filter=lfs diff=lfs merge=lfs -text
|
|
41 |
FLAMA-0.1-3B_q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
flama.png filter=lfs diff=lfs merge=lfs -text
|
43 |
FLAMA-0.1-3B_q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
41 |
FLAMA-0.1-3B_q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
flama.png filter=lfs diff=lfs merge=lfs -text
|
43 |
FLAMA-0.1-3B_q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
+
FLAMA-0.1-3B_q6_k.gguf filter=lfs diff=lfs merge=lfs -text
|
FLAMA-0.1-3B_q6_k.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:53cce30d1ef0dd804c5a9659b2fa218c2b80b4eafe6e502c1ca85e77a95a3016
|
3 |
+
size 3641876128
|