-
Notifications
You must be signed in to change notification settings - Fork 32
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Parallelize probe training across layers #64
Comments
To be clear: a fix for this issue should remove all the vestigial calls to |
I'd be interested in picking this up! Is there a preferred timeframe in which this gets done? My guess is that I could probably finish this over the coming weekend, but if that's not soon enough probably someone else should tackle it. |
@anshradh Yeah if it's done by the end of the weekend that should be fine! Feel free to ping me on the eliciting-latent-knowledge channel of the Eleuther Discord with questions |
Instead of using
torch.distributed
to paralleize probe training, we can simply create amultiprocessing.Queue
with one worker per GPU, and train the probes for each layer of a transformer in parallel.The text was updated successfully, but these errors were encountered: