Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question: How do we deploy on Kubernetes for a specific node #15

Open
AMIYAMAITY opened this issue Jan 11, 2024 · 1 comment
Open

Question: How do we deploy on Kubernetes for a specific node #15

AMIYAMAITY opened this issue Jan 11, 2024 · 1 comment

Comments

@AMIYAMAITY
Copy link

AMIYAMAITY commented Jan 11, 2024

Hi,

Thank you for your hard work to build a GPU memory sharing package and to make it open for everyone.
I have tested locally with multiple processes with different kinds of scenarios. It's working fine as expected.
I went through the README file.

--
The Device Plugin runs on every GPU-enabled node in your Kubernetes cluster (currently it will fail on non-GPU nodes but that is OK) and manages a single GPU on every node. It consumes a single nvidia.com/gpu device and advertizes it as multiple (by default 10) nvshare.com/gpu devices. This means that up to 10 containers can concurrently run on the same physical GPU.

So, If we deploy the deployment files it will deploy on every GPU-enabled node, but I want to deploy in a specific node only for now. I was thinking of changing the matchLabels under the selector section to nodeSelector but I see matchLabels' name is using other places as well.

Could you please help me out to deploy the node specifically?

@grgalex
Copy link
Owner

grgalex commented Jan 11, 2024

I'm glad you found this useful!

The label/nodeselector concept that K8s offers looks like a good starting point for what you want: https://kubernetes.io/docs/tasks/manage-daemon/pods-some-nodes/

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants