Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question] What is the status of Vulkan backend? #542

Open
DanielMazurkiewicz opened this issue Sep 27, 2023 · 12 comments
Open

[Question] What is the status of Vulkan backend? #542

DanielMazurkiewicz opened this issue Sep 27, 2023 · 12 comments

Comments

@DanielMazurkiewicz
Copy link

DanielMazurkiewicz commented Sep 27, 2023

Vulkan may be not the the best/fastest/easiest or so solution for inference, but is probably most portable GPU acceleration approach.

Is anyone working actively to add support for it? And if so what is status/progress? If not, is it planned?

@Green-Sky
Copy link
Contributor

there are ggerganov/llama.cpp#2059 and ggerganov/llama.cpp#2039

@0cc4m
Copy link
Contributor

0cc4m commented Sep 29, 2023

Yeah, I'm working on it. Let me know if you have any questions. It's a big project, but I'm making progress.

@Calandiel
Copy link

Yeah, I'm working on it. Let me know if you have any questions. It's a big project, but I'm making progress.

Is there any low hanging fruit a newcomer to the project could help with?

@0cc4m
Copy link
Contributor

0cc4m commented Nov 16, 2023

Yeah, I'm working on it. Let me know if you have any questions. It's a big project, but I'm making progress.

Is there any low hanging fruit a newcomer to the project could help with?

@Calandiel If you have experience with Vulkan, maybe. Otherwise probably not.

@Calandiel
Copy link

Yeah, I'm working on it. Let me know if you have any questions. It's a big project, but I'm making progress.

Is there any low hanging fruit a newcomer to the project could help with?

@Calandiel If you have experience with Vulkan, maybe. Otherwise probably not.

I have. I've written vulkan based render pipelines professionally and made toy neural networks in vulkan trained with sgd. Been working with it at least in some capacity for the last 4 years or so.

@0cc4m
Copy link
Contributor

0cc4m commented Nov 16, 2023

Yeah, I'm working on it. Let me know if you have any questions. It's a big project, but I'm making progress.

Is there any low hanging fruit a newcomer to the project could help with?

@Calandiel If you have experience with Vulkan, maybe. Otherwise probably not.

I have. I've written vulkan based render pipelines professionally and made toy neural networks in vulkan trained with sgd. Been working with it at least in some capacity for the last 4 years or so.

Oh cool, I'd be glad to work something out. If you have Discord, send me a message (_occam), otherwise send me an Email and we'll find another way.

@Calandiel
Copy link

Yeah, I'm working on it. Let me know if you have any questions. It's a big project, but I'm making progress.

Is there any low hanging fruit a newcomer to the project could help with?

@Calandiel If you have experience with Vulkan, maybe. Otherwise probably not.

I have. I've written vulkan based render pipelines professionally and made toy neural networks in vulkan trained with sgd. Been working with it at least in some capacity for the last 4 years or so.

Oh cool, I'd be glad to work something out. If you have Discord, send me a message (_occam), otherwise send me an Email and we'll find another way.

Will do, see you on Discord!

@sorasoras
Copy link

I think nomic-ai have functional kompute of llama.cpp right now
https://github.com/nomic-ai/llama.cpp
and, GPT4ALL is plenty fast on my 7900XTX via vulkan.
but I am not sure how to integrate this on ggml as i am not a programmer.

@Green-Sky
Copy link
Contributor

@sorasoras ggerganov/llama.cpp#4456

CCLDArjun pushed a commit to CCLDArjun/ggml that referenced this issue Dec 18, 2023
* Revert 7e53955 (ggerganov#542)

Still needs to be fixed properly

* Fix linking on mingw32
@slaren
Copy link
Collaborator

slaren commented Jan 29, 2024

The vulkan and kompute backends have been merged in llama.cpp, all that is left is update the cmake build files to be able to use them in other ggml projects.

@Kreijstal
Copy link

The vulkan and kompute backends have been merged in llama.cpp, all that is left is update the cmake build files to be able to use them in other ggml projects.

does this affect anything using ggml? whisper.cpp stabledifussion.cpp? etc?

@Green-Sky
Copy link
Contributor

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

7 participants