-
Notifications
You must be signed in to change notification settings - Fork 67
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Move each hub to its own nodegroup on the openscapes cluster #4482
Comments
I think there's a language problem here (and in #4486) of tags vs. labels, both of which exist. As I understand it, tags operate at the cloud vendor level, but labels can be used as selectors at the kubernetes level. If we want pods to be spun up in specific node pools, we definitely want to be using labels. But I don't know if the cost-tracking system we are going to use will be looking at cloud tags or kubernetes labels. In the end, neither of which are costly to apply so I will probably just do both. |
I'm attempting this, but I have no idea where to put the |
If they are dictionaries (rather than lists), they should merge (since jupyterhub/kubespawner#650). So your instinct to put it in
Can you provide more detail? |
This is using my first instinct to add |
After the outcome of the spike in #4465, we are going to give each hub its own nodepool that is properly tagged to track cost on a per-hub basis.
staging
,prod
andworkshop
2i2c:hub-name
to match the name of the hub2i2c:node-purpose
set touser
Definition of done
tag
2i2c:hub-name
and2i2c:node-purpose
on all the nodes spawned when users log on to the hub. You can verify this by looking at EC2 instances on the AWS console.Trade-offs
Since our health check triggers a user spawn, this means that instead of spawning 1 node when we trigger deploys on all of the hubs, we will trigger 3 separate nodes. This is fine - the autoscaler reclaims them after ~10min, and even with the largest nodes that doesn't cost enough to be a problem.
Out of scope
dask-gateway is out of scope here, and handled by #4485
The text was updated successfully, but these errors were encountered: