Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for MPS device (Apple M1/M2) #61

Open
louis030195 opened this issue Apr 28, 2023 · 2 comments
Open

Support for MPS device (Apple M1/M2) #61

louis030195 opened this issue Apr 28, 2023 · 2 comments

Comments

@louis030195
Copy link

Would be nice to support MPS to use this model on consumer hardware, it would be super useful, for example with Apple Shortcuts + raycast etc. i already have a bunch of gpt4 shortcuts which i would be happy to try with a non-privacy leaking/faster model

RuntimeError: MPS does not support cumsum op with int64 input

Since there is no code available, I cannot point out where is the fix for this

@mcmonkey4eva
Copy link

mcmonkey4eva commented Apr 29, 2023

I believe using external projects like GGML aka llama.cpp, apple silicon is supported for StableLM models.

@parkerlamb
Copy link

The pytorch nightly releases have support for int64 cumsum ops (on macOS 13.3+), I've managed to get the sample code working by installing:

pip3 install --pre torch --index-url https://download.pytorch.org/whl/nightly/cpu

and changing the device to mps instead of cuda. The outputs are mostly nonsensical though, I assume that's due to the an issue in the MPS backend seeing as the CUDA implementation works ok (or at least the sample app on huggingface does).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants