Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Llava 1.6 Instruction Finetuning #25

Open
gdsaikrishna opened this issue Nov 28, 2024 · 0 comments
Open

Llava 1.6 Instruction Finetuning #25

gdsaikrishna opened this issue Nov 28, 2024 · 0 comments

Comments

@gdsaikrishna
Copy link

gdsaikrishna commented Nov 28, 2024

Hi @ChenMnZ,

I am seeking clarification and guidance on the process of quantizing the Llava 1.6 model using the Efficient QAT repository. Specifically, I would like to confirm the steps involved and understand the details regarding what components are finetuned at each stage of the process.

Queries

  1. Applying BlockAP on LLM : Is the initial step to apply BlockAP quantization on the LLM? If so, are there any specific datasets, considerations or configurations required during this step?
  2. Freezing LLM and Vision Transformer (ViT), and Training the Projector: After obtaining the BlockAP-quantized LLM, the next step appears to involve freezing both the LLM and ViT while training the projector. Could you provide details on where to perform the projector training ? Any relevant scripts or functions would be helpful for implementing this step effectively.
  3. End-to-End Fine-Tuning of LLM and Projector: During the end-to-end fine-tuning stage, do we:
    a. Finetune only the scales for the LLM?
    b. Finetune the weights for the projector?
    Are there any additional parameters or components involved in this finetuning process that I might be missing?

Could you please provide clarity on the above queries and confirm if the outlined process aligns with the intended approach for quantizing Llava 1.6? Additionally, I would appreciate any guidance on specific code references or best practices for implementing the training and fine-tuning stages.

Looking forward to your insights! Thank you in advance for your support.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant