-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Question] Have you finished LoRA/QLoRA training? #206
Comments
Does anyone complete the lora/qlora training on llava? |
HI, thank you for your interest in our work. We are implementing the LoRA/QLoRA training and we are now verifying the correctness with training some model checkpoints. Will update soon if the performance is verified. Thanks! |
@haotian-liu Looking forward to the implementation of LoRA/QLoRA training, could you put the code in the develop branch, and we can try to develop and verify it together. |
LoRA support (preview) and an initial checkpoint I trained is released here. Please let me know if there is something unclear in the instruction. Also, if any of you are interested in contributing to the hyperparameter search, please let me know. Note that QLoRA support is partially finished, but distributed training is not supported yet. |
Hi... I find that LoRA Fine-tuning is slower (almost half - 5.72 s/it) as compared to full training (2.77 s/it). Same LR - 2e-5 and Batch Size -4, Grad-accum- 1 was considered for both the setups. Script was run using torchrun with deepspeed config(zero3.json) as an argument. Are there any similar observations? Intuitively, LoRA should be faster! btw, wonderful work! |
Hi @devrituraj Thank you for trying this out and for providing the feedback. Which GPUs (and how many) are you using? And can you provide your commands? And do you notice a reduction in the GPU memory consumption? For smaller models I noticed a smaller performance benefit. It would be better to know more about the specific configurations on your side. Thanks! |
@haotian-liu How many hours do you spend when finetune 13B with lora? |
@haotian-liu ,It's great work support deepspeed and lora, significantly reduces GPU memory, but I changed tf32 and bf16 to False, it also fails to train, do you have the same problem? |
I think lora works now, but quantization does not work for me, does anyone complete the QLora training? quantization reports this error:
|
@aprilehannibal I do not see too much of a difference between LoRA and full-finetuning, both around 1.5-2 hours on finetune lightning. |
@haotian-liu |
Question
Nice work!
I just wonder when could you finish LoRA/QLoRA training? It may help a lot to finetune.
The text was updated successfully, but these errors were encountered: