Griffin-3B-GPTQ / README.md
daedalus314's picture
Update README.md
c2f36b7
|
raw
history blame
1.89 kB
---
license: apache-2.0
datasets:
- LDJnr/Puffin
language:
- en
library_name: transformers
tags:
- NLP
- GPTQ
---
# Overview
This model is a quantized version of [Griffin-3B](https://huggingface.co/acrastt/Griffin-3B), using [GPTQ](https://arxiv.org/abs/2210.17323). The quantization has been done following the sample [notebook](https://colab.research.google.com/drive/1_TIrmuKOFhuRRiTWN94iLKUFu6ZX4ceb) provided by Hugging Face.
# Usage
The model has been quantized as part of the project [GPTStonks](https://github.com/GPTStonks). It works with `transformers>=4.33.0` and it can run on a consumer GPU, with less than 3GB of GPU RAM. The libraries `optimum`, `auto-gptq`, `peft` and `accelerate` should also be installed.
Here is a sample code to load the model and run inference with it using sampling as decoding strategy:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model_id = "daedalus314/Griffin-3B-GPTQ"
tokenizer = AutoTokenizer.from_pretrained(model_id)
quant_model = AutoModelForCausalLM.from_pretrained(model_id, device_map='auto')
text = """### HUMAN:
What is artifical intelligence?
### RESPONSE:
"""
inputs = tokenizer(text, return_tensors="pt").to(0)
out = quant_model.generate(
**inputs,
do_sample=True,
top_p=0.9,
temperature=0.9,
max_length=1024,
)
print(tokenizer.decode(out[0], skip_special_tokens=True))
```
And a sample output:
```
### HUMAN:
What is artifical intelligence?
### RESPONSE:
Artificial intelligence, or AI, refers to the ability of computers to perform tasks that typically require human intelligence, such as decision making, problem solving, and language understanding. AI has been used in various fields, including healthcare, manufacturing, and finance, among others.
```
# Further details
Please refer to the original model [Griffin-3B](https://huggingface.co/acrastt/Griffin-3B).