Finetuning Scripts
Does anyone happen to have the fine-tuning scripts for this model? What is the minimum GPU requirement for fine-tuning?
@abrakaa Thanks for your interest in our model. We may open-source the finetuning script within the near future.
@donniems Need it now! Really looking forward to it!
I wrote a finetuning script for Phi3-V: https://github.com/GaiZhenbiao/Phi3V-Finetuning/tree/main , enjoy π
https://github.com/2U1/Phi3-Vision-ft
I've made a code that has option to fine-tune full module (including vision model) like llava-1.6 !
Thank you all your interest in Phi-3 Vision model.
This is the finetuning recipe https://github.com/microsoft/Phi-3CookBook/blob/main/md/04.Fine-tuning/FineTuning_Vision.md
@eddtsoi
there's no explicit code for handling OOM. Note that the 2k average #tokens include the image tokens. In some cases that doesn't require high resolution, it is possible to finetune with a lower --num_crops
to reduce the sequence length. For full finetuning, we have tested on 4x 48GB and 8x 32GB GPUs. If using (q)lora, I believe a single consumer level 24GB GPU works.