-
Notifications
You must be signed in to change notification settings - Fork 19
Issues: Q-Future/Q-Align
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
when try to use flash_attention_v2, not find
_prepare_4d_causal_attention_mask_for_sdpa
#40
opened Nov 6, 2024 by
dongdk
When I try to use koniq-10 for training, I encounter the problem that the loss is always 0.
#37
opened Sep 27, 2024 by
style09131111
RuntimeError: Error(s) in loading state_dict for MPLUGOwl2LlamaForCausalLM: size mismatch for norm.weight: copying a param with shape torch.Size([768]) from checkpoint, the shape in current model is torch.Size([4096]). You may consider adding
ignore_mismatched_sizes=True
in the model from_pretrained
method.
#33
opened Sep 6, 2024 by
style09131111
ProTip!
Exclude everything labeled
bug
with -label:bug.