Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fine-tuning via quantization #108

Open
2 tasks done
0xD4rky opened this issue Jan 11, 2025 · 0 comments
Open
2 tasks done

fine-tuning via quantization #108

0xD4rky opened this issue Jan 11, 2025 · 0 comments
Labels
enhancement New feature or request

Comments

@0xD4rky
Copy link

0xD4rky commented Jan 11, 2025

Search before asking

  • I have searched the Multimodal Maestro issues and found no similar feature requests.

Description

I was going through the maestro repo and found out that both paligemma and florence models didn't support the implementation of 4-bit quantization (i.e. using QLoRA config).

Use case

Using QLoRA, we could easily fine-tune vision language models on even low end devices without losing on precision a lot. As the models grow, we would eventually need to implement QLoRA to make finetuning fast and possible on memory constraints.

Additional

I would like to learn your take on implementing quantization.

Are you willing to submit a PR?

  • Yes I'd like to help by submitting a PR!
@0xD4rky 0xD4rky added the enhancement New feature or request label Jan 11, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant