Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix Q3_K_XS for MoE models #5113

Merged
merged 1 commit into from
Jan 25, 2024
Merged

Fix Q3_K_XS for MoE models #5113

merged 1 commit into from
Jan 25, 2024

Conversation

ikawrakow
Copy link
Contributor

I made the very same mistake as I when I was restoring k-quants quantization mixture for MoE models.
This PR fixes it and MoE models should now work with Q3_K_XS.

@ikawrakow ikawrakow mentioned this pull request Jan 24, 2024
@ikawrakow ikawrakow merged commit faa3526 into master Jan 25, 2024
48 checks passed
@ikawrakow ikawrakow deleted the ik/fix_q3k_xs branch January 25, 2024 15:58
jordankanter pushed a commit to jordankanter/llama.cpp that referenced this pull request Feb 3, 2024
hodlen pushed a commit to hodlen/llama.cpp that referenced this pull request Apr 1, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants