Initial GPTQ model commit
Browse files
README.md
CHANGED
@@ -17,9 +17,9 @@ license: other
|
|
17 |
</div>
|
18 |
<!-- header end -->
|
19 |
|
20 |
-
# LmSys' Vicuna 13B 1.3.0
|
21 |
|
22 |
-
These files are GPTQ 4bit model files for [LmSys' Vicuna 13B 1.3.0
|
23 |
|
24 |
It is the result of quantising to 4bit using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa).
|
25 |
|
@@ -206,7 +206,7 @@ I trained the LoRA with the following configuration:
|
|
206 |
- AdamW beta1 of 0.9 and beta2 0.99, epsilon of 1e-5
|
207 |
- Trained on 4-bit base model
|
208 |
|
209 |
-
# Original model card: LmSys' Vicuna 13B 1.3.0
|
210 |
|
211 |
|
212 |
# Vicuna Model Card
|
|
|
17 |
</div>
|
18 |
<!-- header end -->
|
19 |
|
20 |
+
# LmSys' Vicuna 13B 1.3.0 GPTQ
|
21 |
|
22 |
+
These files are GPTQ 4bit model files for [LmSys' Vicuna 13B 1.3.0](https://huggingface.co/lmsys/vicuna-13b-v1.3) merged with [Kaio Ken's SuperHOT 8K](https://huggingface.co/kaiokendev/superhot-30b-8k-no-rlhf-test).
|
23 |
|
24 |
It is the result of quantising to 4bit using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa).
|
25 |
|
|
|
206 |
- AdamW beta1 of 0.9 and beta2 0.99, epsilon of 1e-5
|
207 |
- Trained on 4-bit base model
|
208 |
|
209 |
+
# Original model card: LmSys' Vicuna 13B 1.3.0
|
210 |
|
211 |
|
212 |
# Vicuna Model Card
|