Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implement DeepSpeed Main autotuning for NeoX #739

Merged
merged 101 commits into from
Mar 9, 2023
Merged

Implement DeepSpeed Main autotuning for NeoX #739

merged 101 commits into from
Mar 9, 2023

Conversation

dashstander
Copy link
Contributor

@dashstander dashstander commented Dec 9, 2022

Implements autotuning feature from DeepSpeed, which programmatically explores different combinations of microbatch size, gradient accumulation, and ZeRO sharding options given your model and compute set up.

Dashiell Stander and others added 30 commits September 21, 2022 20:33
Signed-off-by: Dashiell Stander <[email protected]>
Signed-off-by: Dashiell Stander <[email protected]>
Signed-off-by: Dashiell Stander <[email protected]>
Signed-off-by: Dashiell Stander <[email protected]>
Signed-off-by: Dashiell Stander <[email protected]>
Signed-off-by: Dashiell Stander <[email protected]>
Signed-off-by: Dashiell Stander <[email protected]>
Signed-off-by: Dashiell Stander <[email protected]>
Signed-off-by: Dashiell Stander <[email protected]>
@dashstander
Copy link
Contributor Author

Ok, should be ready for review @Quentin-Anthony

@Quentin-Anthony Quentin-Anthony changed the base branch from main to deepspeed_main February 14, 2023 22:02
Base automatically changed from deepspeed_main to main March 9, 2023 16:55
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

4 participants