colab t4
It never works in Colab T4
If you have a code that works in Colab T4
Please provide it to us
It never works in Colab T4
If you have a code that works in Colab T4
Please provide it to us
It doesnt work on a T4 because Llama 3 70B needs at least 48GB of VRAM. T4 only has 16GB
Can I use device map auto
With devicemapatuo
Not run
How to run it with VGA 16G
Can you provide the python code for
Inference with model
Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("unsloth/Llama-3.1-Nemotron-70B-Instruct-bnb-4bit")
model = AutoModelForCausalLM.from_pretrained("unsloth/Llama-3.1-Nemotron-70B-Instruct-bnb-4bit", device map=auto)
Errors
Can the 4-bit model be run with the acclrate library to distribute the model to the graphics unit and processor? The problem is that the code I used generated errors and did not work.
Is there a free cloud service other than Colab T4 that has a graphics module that runs the model?