-
Notifications
You must be signed in to change notification settings - Fork 182
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Florence-2 QLoRA support #108
Comments
Hi @0xD4rky 👋🏻 Thank you for your interest in @0xD4rky I saw that you checked the |
Yes @SkalskiP, it would be absolutely great if I am able to contribute towards maestro. Will go through the code base, develop the QLoRA approach for Florence-2 and then raise the PR. This works? |
@0xD4rky that would be awesome! 🔥 I'm excited. We are building something cool here. |
@0xD4rky I'll assign this issue to you! BTW I have a question: I'm thinking about launching a Discord server dedicated to VLM fine-tuning with Maestro) and while talking about current issues I'm trying to understand if people would like such a server to be created. |
@SkalskiP, people would love to hop on a Roboflow's discord server in general. It would be great to discuss the fine-tunings and approaches using Maestro, people would be definitely interested! |
@0xD4rky sounds great; I'll try to set something up this week! By the way, I updated the contributors guidelines today. Take a look, considering you want to add support for Florence-2 QLoRA. |
@SkalskiP sure, I will go through it before starting to contribute. Will keep you in touch with all the changes that I come up with |
Search before asking
Description
I was going through the maestro repo and found out that both paligemma and florence models didn't support the implementation of 4-bit quantization (i.e. using QLoRA config).
Use case
Using QLoRA, we could easily fine-tune vision language models on even low end devices without losing on precision a lot. As the models grow, we would eventually need to implement QLoRA to make finetuning fast and possible on memory constraints.
Additional
I would like to learn your take on implementing quantization.
Are you willing to submit a PR?
The text was updated successfully, but these errors were encountered: