-
Notifications
You must be signed in to change notification settings - Fork 2.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Usage] Some Weights not used, when loaded in eval mmbench. #672
Comments
Is the checkpoint trained by yourself? If so, this is expected, as DeepSpeed saves the frozen vision encoder weights as well. If your results are normal, than you can safely ignore this warning. |
@haotian-liu ,Yes, the checkpoint is trained myself. Thanks your reply. But I found that when use your released checkpoint , there are not such logs. |
If you want to remove the vision tower as the checkpoint we released, you can do this:
The model prediction would be the same regardless of you do anything like that. |
Hi, why is vision_tower not initialized through model.from_pretrain(model_name_or_path) since getattr('vision_tower') is true and there is state_dict in checkpoint? |
@haotian-liu as for the mmbench dataset, gt answers are provided in |
@haotian-liu I have the same question, hhh |
Hi @CrossLee1, I also found the gt answers, but the accuracy I calculated is much higher than reported. could you please give me some suggestions? thanks |
@annopackage Same question here... have you figured it out? |
When I was loading the LORA fine-tuned model, I encountered this issue as well. The message said: 'Some weights of the model checkpoint at [my lora model] were not used when initializing LlavaLlamaForCausalLM.' I also tried fine-tuned models with different numbers of iterations, but the result was the same. It seems like the weights from the LORA fine-tuning were not loaded? |
Describe the issue
Issue:
When I use my second stage trained model , there are some logs
Command:
Log:
Screenshots:
You may attach screenshots if it better explains the issue.
The text was updated successfully, but these errors were encountered: