TheBloke commited on
Commit
dcb44d0
1 Parent(s): 599184c

Initial GPTQ model commit

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -17,9 +17,9 @@ license: other
17
  </div>
18
  <!-- header end -->
19
 
20
- # LmSys' Vicuna 13B 1.3.0 merged with Kaio Ken's SuperHOT 8K GPTQ
21
 
22
- These files are GPTQ 4bit model files for [LmSys' Vicuna 13B 1.3.0 merged with Kaio Ken's SuperHOT 8K](https://huggingface.co/lmsys/vicuna-13b-v1.3) merged with [Kaio Ken's SuperHOT 8K](https://huggingface.co/kaiokendev/superhot-30b-8k-no-rlhf-test).
23
 
24
  It is the result of quantising to 4bit using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa).
25
 
@@ -206,7 +206,7 @@ I trained the LoRA with the following configuration:
206
  - AdamW beta1 of 0.9 and beta2 0.99, epsilon of 1e-5
207
  - Trained on 4-bit base model
208
 
209
- # Original model card: LmSys' Vicuna 13B 1.3.0 merged with Kaio Ken's SuperHOT 8K
210
 
211
 
212
  # Vicuna Model Card
 
17
  </div>
18
  <!-- header end -->
19
 
20
+ # LmSys' Vicuna 13B 1.3.0 GPTQ
21
 
22
+ These files are GPTQ 4bit model files for [LmSys' Vicuna 13B 1.3.0](https://huggingface.co/lmsys/vicuna-13b-v1.3) merged with [Kaio Ken's SuperHOT 8K](https://huggingface.co/kaiokendev/superhot-30b-8k-no-rlhf-test).
23
 
24
  It is the result of quantising to 4bit using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa).
25
 
 
206
  - AdamW beta1 of 0.9 and beta2 0.99, epsilon of 1e-5
207
  - Trained on 4-bit base model
208
 
209
+ # Original model card: LmSys' Vicuna 13B 1.3.0
210
 
211
 
212
  # Vicuna Model Card