-
Notifications
You must be signed in to change notification settings - Fork 154
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Convert the huggingface checkpoint to GPT-Neox checkpoint #116
Comments
I have another question: do python-1b, python-12b, and python-20b share the same tokenizer? |
A script of GPT-Neox has made it (https://github.com/EleutherAI/gpt-neox/blob/main/tools/convert_hf_to_sequential.py). |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I want to continue the pretraining based on the pretrained pythia checkpoints. But I don't know whether I can use the checkpoints from huggingface directly. How to convert the huggingface checkpoint to the checkpoints that can be loaded by GPT-Neox.
Can you can provide the raw checkpoints from GPT-Neox?
The text was updated successfully, but these errors were encountered: