-
Notifications
You must be signed in to change notification settings - Fork 4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support GPUs in scale from 0 #401
Conversation
Fixes #392 |
/lgtm |
cc: @vishh |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Keep in mind starting from v1.8, the resource name for GPUs is nvidia.com/gpu and it's not included in core k8s API. So this PR IIUC won't autoscaling clusters>1.8.1 since it's using the old resource name alpha.kubernetes.io/nvidia-gpu
(apiv1.ResourceNvidiaGPU)
@@ -182,6 +190,7 @@ func TestBuildAllocatableFromKubeEnv(t *testing.T) { | |||
capacityMemory string | |||
expectedCpu string | |||
expectedMemory string | |||
gcuCount int64 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Typo: This should be gpuCount
.
Add enhancements template and script for table of contents
When scaling up from 0, read information about GPUs from instance template. This allows us to support scaling up a relevant NodeGroup from 0 when we have a pending pod that needs GPUs.