Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

full-parameter or lora? #3

Open
Nastu-Ho opened this issue Jun 17, 2024 · 3 comments
Open

full-parameter or lora? #3

Nastu-Ho opened this issue Jun 17, 2024 · 3 comments

Comments

@Nastu-Ho
Copy link

Will full-parameter fine-tuning be better?

@mmaaz60
Copy link
Member

mmaaz60 commented Jun 17, 2024

Hi @Nastu-Ho,

As per our experiments, we did not notice any accuracy gains when using Phi3-mini-4K LLM or Vicuna. However, full fine-tuning was better in the case of LLaMA-3 as compared to LoRA.

@Nastu-Ho
Copy link
Author

Nastu-Ho commented Jun 18, 2024

Hi @Nastu-Ho,

As per our experiments, we did not notice any accuracy gains when using Phi3-mini-4K LLM or Vicuna. However, full fine-tuning was better in the case of LLaMA-3 as compared to LoRA.

Thank you for your reply.
One thing I'm curious about is whether a stronger LLM can continue to bring improvements in mvbench.
The problem I am currently encountering is that my model has been replaced with a stronger LLM (qwen2, mistral) and has not been significantly improved like videochat2.
So I'm curious about the performance of your method using different LLMs (such as LLaMA3, mistral) on mvbench?

@mmaaz60
Copy link
Member

mmaaz60 commented Jun 18, 2024

Hi @Nastu-Ho,

We do not have the MVBench results using LLaMA-3 LLM, however, I can share the numbers with Vicuna 7B and 13B that we observe during experiments and this may give us some clues about the trend.

With Vicuna 7B and 13B, we obtain 53.10 and 58.67 average scores on MVBench. These experiments show that using a better/stronger LLM improves the MVBench performance. However, we do not have any ablations with LLaMA-3 and Mistral.

If you have any findings, please do share. Thank You.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants