Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Validation Perplexities #38

Closed
shoaibahmed opened this issue Dec 21, 2022 · 5 comments
Closed

Validation Perplexities #38

shoaibahmed opened this issue Dec 21, 2022 · 5 comments

Comments

@shoaibahmed
Copy link

Thanks for sharing this amazing work. This will hopefully help in developing a better understanding of how LLMs work.

I had one question. Are the validation perplexities for each of the models available (ideally with every model snapshot) so that we can compare models on equal footing?

@haileyschoelkopf
Copy link
Collaborator

Hi! We're actively working on gathering all evals and getting them posted in this repo, including LAMBADA perplexity. If you're wondering about validation perplexity on the Pile, unfortunately in order to save compute we did not run on the validation set during training very frequently at all.

We plan to have LAMBADA perplexity up for all models on this repo for 15 evenly spaced checkpoints (steps 3000, 13000,..., 133000, 143000) asap! Would you want or need more granular results for any experiments?

@shoaibahmed
Copy link
Author

Hi! Thank you for your kind response.

We plan to have LAMBADA perplexity up for all models on this repo for 15 evenly spaced checkpoints (steps 3000, 13000,..., 133000, 143000) asap! Would you want or need more granular results for any experiments?

That's great! Thank you for the awesome work. I was particularly using the aforementioned checkpoints, so having the Lambada perplexity (which should be proportional to the validation perplexity on PILE) should help.

@haileyschoelkopf
Copy link
Collaborator

fantastic, we should have those up and all corrected within a week ideally!

@haileyschoelkopf
Copy link
Collaborator

Hi @shoaibahmed ! I believe all models' evaluations should now be up-to-date.

Let me know if any evals look suspect or anything's missing! Hopefully all is well though, barring some cleanup of filenames (also, all lambada tasks are the lambada_openai task in the Eleuther lm-evaluation-harness.)

Also let me know if having more granular evals/PPL would be helpful to your research for any reason :)

@shoaibahmed
Copy link
Author

That's really awesome. Thank you for your prompt action on this. I will go ahead and close this issue now.

I will open a new issue in case something else comes up which might be beneficial. Thanks once again for your time and effort.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants