Skip to content

Issues: EleutherAI/lm-evaluation-harness

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

Using Language Models as Evaluators feature request A feature that isn't implemented yet.
#1831 opened May 13, 2024 by lintangsutawika
Add More Tests feature request A feature that isn't implemented yet.
#1827 opened May 12, 2024 by haileyschoelkopf
Multi Label Classification
#1814 opened May 9, 2024 by IsraelAbebe
Gemini 1.5/Ultra support
#1808 opened May 8, 2024 by notrichardren
Support OpenAI's Batch API
#1770 opened May 2, 2024 by djstrong
Cannot have both a group list and task list asking questions For asking for clarification / support on library usage. bug Something isn't working.
#1767 opened Apr 29, 2024 by steven-basart
Bug in yaml parsing
#1762 opened Apr 28, 2024 by jordane95
Output constrained support
#1759 opened Apr 27, 2024 by Mihaiii
New Task Request: LegalBench feature request A feature that isn't implemented yet. good first issue Good for newcomers help wanted Contributors and extra help welcome.
#1754 opened Apr 26, 2024 by haileyschoelkopf
ProTip! Follow long discussions with comments:>50.