-
Notifications
You must be signed in to change notification settings - Fork 1.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Implement the SAT evaluation #27
Closed
2 tasks done
Labels
Comments
Should be available on request from Peter Turney. |
StellaAthena
added
Eval Set
and removed
feature request
A feature that isn't implemented yet.
labels
Oct 23, 2020
Will post here if/when we get a response. |
Got a response. PM me on the Discord if you need access. |
Merged
Merged
StellaAthena
added
feature request
A feature that isn't implemented yet.
good first issue
Good for newcomers
labels
Jan 5, 2021
I could do the eval here |
Merged
Merged
StellaAthena
added a commit
that referenced
this issue
Apr 29, 2022
qmdnls
pushed a commit
to qmdnls/lm-evaluation-harness
that referenced
this issue
Aug 17, 2023
LZY-the-boys
pushed a commit
to LZY-the-boys/lm-evaluation-harness-fast
that referenced
this issue
Sep 12, 2023
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Labels
From the GPT-3 paper:
The evaluation code should be modeled after the interface in
lm_eval/base.py
and the example of theBoolQ
task inlm_eval/tasks/suerglue.py
The text was updated successfully, but these errors were encountered: