Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

training code #46

Open
ehartford opened this issue Jun 2, 2024 · 6 comments
Open

training code #46

ehartford opened this issue Jun 2, 2024 · 6 comments

Comments

@ehartford
Copy link

Hello, I am trying to find the training code, but it seems like there is just inference code.

Can you please point to the training code?

@NicoZenith
Copy link

That would be great to get the training scripts, as it was done in the original LLaVA repo :)

@carlos-havier
Copy link

I'd also love to use them for fine-tuning with several images, for few-shot image classification.

@NicoZenith
Copy link

what do you guys think of this ?
https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LLaVa/Fine_tune_LLaVa_on_a_custom_dataset_(with_PyTorch_Lightning).ipynb

By replacing llava by lava-next (processor and model)

@chuangchuangtan
Copy link

I implement a LLava-llama3 Lora finetuning https://github.com/chuangchuangtan/LLaVA-NeXT-Image-Llama3-Lora

@NicoZenith
Copy link

I implement a LLava-llama3 Lora finetuning https://github.com/chuangchuangtan/LLaVA-NeXT-Image-Llama3-Lora

Great thank you! Does it also work with Llama3 70b?
Btw, does it train only the bridger and language model, or does it also train the vision encoder (that we want to avoid)? Can we train without LoRA ?

@chuangchuangtan
Copy link

I implement a LLava-llama3 Lora finetuning https://github.com/chuangchuangtan/LLaVA-NeXT-Image-Llama3-Lora

Great thank you! Does it also work with Llama3 70b? Btw, does it train only the bridger and language model, or does it also train the vision encoder (that we want to avoid)? Can we train without LoRA ?

It only trains the bridge and language model. We have set up to print the names of trainable parameters in the code, you can check them. We haven't tested it on 70b, but it should be work. You can set training commands to train without LoRA.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants