vivicai commited on
Commit
7ca4f1a
1 Parent(s): b3de569

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -4
README.md CHANGED
@@ -13,9 +13,9 @@ license: apache-2.0
13
 
14
 
15
 
16
- This is a 4-bit GPTQ version of the [Tigerbot 13b chat](https://huggingface.co/TigerResearch/tigerbot-13b-chat).
17
 
18
- It was quantized to 8bit using: https://github.com/qwopqwop200/GPTQ-for-LLaMa
19
 
20
  ## How to download and use this model in github: https://github.com/TigerResearch/TigerBot
21
 
@@ -34,7 +34,10 @@ pip install -r requirements.txt
34
  Inference with command line interface
35
 
36
  ```
37
- cd TigerBot/gptq
38
- CUDA_VISIBLE_DEVICES=0 python tigerbot_infer.py TigerResearch/tigerbot-13b-chat-8bit --wbits 4 --groupsize 128 --load TigerResearch/tigerbot-13b-chat-8bit/tigerbot-13b-8bit-128g.pt
 
 
 
39
  ```
40
 
 
13
 
14
 
15
 
16
+ This is a 8-bit GPTQ version of the [Tigerbot 13b chat](https://huggingface.co/TigerResearch/tigerbot-13b-chat).
17
 
18
+ It was quantized to 8bit using: https://github.com/PanQiWei/AutoGPTQ
19
 
20
  ## How to download and use this model in github: https://github.com/TigerResearch/TigerBot
21
 
 
34
  Inference with command line interface
35
 
36
  ```
37
+ # 安装auto-gptq
38
+ pip install auto-gptq
39
+
40
+ # 启动推理
41
+ CUDA_VISIBLE_DEVICES=0 python other_infer/gptq_infer.py --model_path ${MODEL_PATH}
42
  ```
43