Asking for params

#1
by Manel-Hik - opened

Hi
Thanks for this great effort
Could you share with us parameters used in training?
Thanks in advance

Hi, this is a lora model must be merged with the base model to be used.
If you need a model with all params merged in 16 bits , check here please : https://huggingface.co/Omartificial-Intelligence-Space/Arabic-llama3.1-16bit-FT

Hope I answer your query

Hi
thanks for sharing
But I meant params of the finetuning aka: lr, warmup, batch size, quantization (4bit or 8bit)...
Thanks in advance

The based model was loaded in 4 bits and then fine-tuned with the following params:

lora_alpha = 16,
lora_dropout = 0,
bias = "none",
learning_rate = 2e-4,
per_device_train_batch_size = 2,
gradient_accumulation_steps = 4,
warmup_steps = 5,

what was the rank for this model?

nvmsaw int eh config file "r": 16

Omartificial-Intelligence-Space changed discussion status to closed

Sign up or log in to comment