Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Pythia 12b flash config #162

Open
jvendrow opened this issue May 26, 2024 · 0 comments
Open

Pythia 12b flash config #162

jvendrow opened this issue May 26, 2024 · 0 comments

Comments

@jvendrow
Copy link

jvendrow commented May 26, 2024

The pythia 12b config has:

"attention-config": [[["flash"], 40]],

However, in the gpt-neox repo the 40 is replaced by 36, and in the file:

https://huggingface.co/EleutherAI/neox-ckpt-pythia-12b-v1/blob/main/12B.yml

This value of 36. Is this a mistake? Also, the attention config line at:

"attention-config": [[["flash"], 40]]

seems to be missing a comma?

Edit: The num-layers value also seems off, 36 v. 40.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant