-
Notifications
You must be signed in to change notification settings - Fork 978
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Do pythia untied embedding and unembedding matrics? #913
Comments
Which config files are you looking at? the ones in the Pythia repo and the ones here https://github.com/EleutherAI/gpt-neox/blob/main/configs/pythia/1-4B.yml use the value “no-weight-tying”: true as expected. |
Sorry, I Made a mistake. I searched config files and found no "-" and "_" are different from readme and config files. |
@Life-0-1 what readme are you talking about? The one in |
|
Opened an issue to fix the inconsistency #920 |
In the Pythia paper, it is stated that embedding and unembedding matrics are untied.
But I find that config files of all kinds of pythia models are with the
no_weight_tying
remaining a default value ofFalse
, which results a tying between embed and unembed layers.Is this a bug? What setting did you actually use while training pythia?
The text was updated successfully, but these errors were encountered: