colab t4

#4
by sdyy - opened

It never works in Colab T4
If you have a code that works in Colab T4
Please provide it to us

Unsloth AI org

It never works in Colab T4
If you have a code that works in Colab T4
Please provide it to us

It doesnt work on a T4 because Llama 3 70B needs at least 48GB of VRAM. T4 only has 16GB

Can I use device map auto

With devicemapatuo
Not run

How to run it with VGA 16G

Can you provide the python code for
Inference with model

Load model directly

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("unsloth/Llama-3.1-Nemotron-70B-Instruct-bnb-4bit")
model = AutoModelForCausalLM.from_pretrained("unsloth/Llama-3.1-Nemotron-70B-Instruct-bnb-4bit", device map=auto)

Errors

Can the 4-bit model be run with the acclrate library to distribute the model to the graphics unit and processor? The problem is that the code I used generated errors and did not work.

Is there a free cloud service other than Colab T4 that has a graphics module that runs the model?

Sign up or log in to comment