-
Notifications
You must be signed in to change notification settings - Fork 1.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add better test coverage for models #1613
Comments
Hi Hailey, I am interested in contributing to this great project and helping. Let me take spend some time on the docs and library to learn how it is setup and how the tests are designed, so I can help improve the test coverage. I will ask you more questions as I get going and run into any knowledge blocks :) |
Hi @haileyschoelkopf , actually would you or one of your colleagues point me to a video or design doc that explains how the code base is setup and how to navigate it? from my quick look, it is complex even for an experienced coder like me, so a little kickstart wtih an existing developer would make it easier. I am happy to do a video chat also if that's easier. |
Hey @raamana , I don't have the bandwidth to do a video chat unfortunately, but we'll be putting out a preprint soon that has in part a description of the high-level moving pieces / design! If you try to understand the codebase but hit some snags, feel free to ping on discord in the #lm-thunderdome channel of EleutherAI's discord (link should be in readme) and I or someone else will probably be able to respond! It would be helpful to know what components are most confusing or should be better documented somewhere. |
We should improve our test coverage of the various model classes supported.
Ideally we'd have at least https://github.com/EleutherAI/lm-evaluation-harness/blob/main/tests/models/test_huggingface.py tests for every model class. This would require, for API models, mocking out the API call with a dummy one, and probably something similar for VLLM, which currently requires CUDA to be run so we don't have the tests run on the CI.
Help or contributions on this welcome! Else, I'll look into this hopefully soonish.
The text was updated successfully, but these errors were encountered: