-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Officially Support AMD GPUs #954
Comments
Will take a look at this! |
Do we still need to detect flash attention 1 at this point since AMD has already ported their version 2? (https://github.com/ROCmSoftwarePlatform/flash-attention) Branch: https://github.com/R0n12/gpt-neox-fork/tree/lang/amd |
No I think we can drop that. |
Status Update: Been busy with life things recently, will get a new clean branch pushed out by next week. |
Done! Great work @R0n12 |
Currently, AMD GPU support lives experimentally in main...AMD
We should port the kernel guards in microsoft/Megatron-DeepSpeed@b4d4a0e#diff-059209398b62b21e2524b387fc6fd23ded28ec725798322c266ee4246d253670 to gpt-neox and bring this into
main
main
on AMD GPUs with fused kernels set tofalse
The text was updated successfully, but these errors were encountered: