Edit model card

Gemma 2 9B 8-bit

This is an 8-bit quantized version of Gemma 2 9B. The models belong to Google and are licensed under the Gemma Terms of Use and are only stored in quantized form here for convenience.

How to use

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
dtype = torch.float16
model = AutoModelForCausalLM.from_pretrained("nev/gemma-2-9b-8bit", torch_dtype=dtype, device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("nev/gemma-2-9b-8bit")
Downloads last month
20
Safetensors
Model size
9.24B params
Tensor type
F32
FP16
I8
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.