-
Notifications
You must be signed in to change notification settings - Fork 327
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Convert to Gguf format to work with Llama.cpp? #32
Comments
Hi, thank you for your suggestion. I will add compatibility with community tools to my to-do list. |
gguf format is good for ollama users, Any Update? |
It will be nice to have this model in gguf format in ollama. |
any updates on this? the 4b intern model is killer for its size! would love to see it supported with llama.cpp |
Would love internvl-chat-v1-5 in a gguf format! |
I second this |
@ErfeiCui why did you close this as completed? |
Llava has various quantized models in gguf format, so it can be used with Llama.cpp.
ggerganov/llama.cpp#3436
Is this possible?
The text was updated successfully, but these errors were encountered: