-
Notifications
You must be signed in to change notification settings - Fork 948
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support for RedPajama #134
Conversation
Co-authored-by: Tom Bailey <[email protected]>
just tested, it works. But I feel like adding more and more small variations of the same code is kinda bad. we should merge stableml and redpajama (and dolly?) and call it gptneox :) |
I agree. Dolly and stableLM were pretty similar too.
…On Sun, May 7, 2023, 2:45 PM Erik Scholz ***@***.***> wrote:
just tested, it works. But I feel like adding more and more small
variations of the same code is kinda bad. we should merge stableml and
redpajama (and dolly?) and call it gptneox :)
—
Reply to this email directly, view it on GitHub
<#134 (comment)>, or
unsubscribe
<https://github.com/notifications/unsubscribe-auth/ABPM37YZJHAIFPAZXMLWNOTXE6DNLANCNFSM6AAAAAAXYE2JEM>
.
You are receiving this because you authored the thread.Message ID:
***@***.***>
|
Yup, will combine them soon. Just need some time to test the Dolly model and make sure the inference is correct - currently can't convert the model on Mac OS since no bfloat16 python support. |
@amirza1 awesome stuff! uploading ggml models on hugging face https://huggingface.co/huggersbro/RedPajama-INCITE-Chat-3B-v1-GGML (should be good to link since it's open source right?) |
Hi @ggerganov
I liked @keldenl's idea so I posted the ggml bins here if that helps you test: Let me know if I can assist. |
Here's the 3B instruct model: https://huggingface.co/keldenl/RedPajama-INCITE-Instruct-3B-v1-GGML/ @amirza1 @ggerganov should we link these ggml models in the readme like gpt-2 (since this has apache 2 license) as an alternative option (i.e. "or you can get the ggml directly") Update: Here's the 7B instruct model https://huggingface.co/keldenl/RedPajama-INCITE-Instruct-7B-v0.1-GGML only chat 7b left, i'll upload it later tonight |
I did give a shot and tried Q5 locally, no luck so far
|
@mudler This works fine for me. This is the mverrilli/dolly-v2-12b-ggml/ggml-model-q5_0.bin model, correct? Maybe check your hash? SHA256: 79280421cc792330eaa56621060b8e2fb48ef570ace4572a91a1cf0e18ce7f38 There isn't a lot of error handling on the examples. Do you have enough ram to load the model? |
I've downloaded the 7b model (https://huggingface.co/mverrilli/dolly-v2-7b-ggml/blob/main/ggml-model-q5_0.bin):
Note I have 64GB of RAM, so that shouldn't be the issue |
@mudler Hash matches mine and I repulled master and rebuilt and it is working. I don't want to clutter up this PR any further, if you want to create a new issue I can work through it with you. |
I've been a bit busy these days - will start looking soon into the newly proposed models here. Please check if #139 works with RedPajama, and if so - I think we should merge it instead of adding a new example in order to reduce code duplication |
I've decided to merge #139 |
reconverted and works 👍
|
That commit has a bug. It calls |
This supports RedPajama, which is gptneo-x with use_parallel_residual=False.