Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Time-slicing with multiple GPUs - asking for ability to block single GPU #465

Open
Alexbay218 opened this issue Jan 4, 2023 · 3 comments

Comments

@Alexbay218
Copy link

Alexbay218 commented Jan 4, 2023

The template below is mostly useful for bug reports and support questions. Feel free to remove anything which doesn't apply to you and add more information where it makes sense.

1. Issue or feature description

I'm looking to have the ability to configure the scheduler to perform the exact opposite behavior as specified here: #386

Instead of grabbing gpu resources evenly from all gpus on the node, I'd like a config option to grab from one GPU at a time. This allows some applications to get exclusive access to a single GPU as needed while allowing the rest to time share.

2. Steps to reproduce the issue

  1. Perform a new fresh installation of the GPU operator in a cluster where nodes have more than one GPU.
  2. Enable time-slicing and configure it to allow 2 replicas per GPU
  3. Start a pod to consume 2 gpu extended resource
  4. Pod should have exclusive access to a single GPU, but instead it has access to two GPUs (as intended by Time-slicing with multiple GPUs - asking for two GPUs puts both slots on a single GPU #386)
@shivamerla
Copy link
Contributor

@klueska does it make sense to introduce knobs(env/args) to control allocation logic during GetPreferredAllocation within the device plugin?

@anencore94
Copy link

@shivamerla I believe there are cases where we would want distributed GPU scheduling, but there might also be opposite scenarios. It would be great if this setting could be easily changed by configmap or so.

@zr-idol
Copy link

zr-idol commented Aug 4, 2024

Any update regarding this issue?
We'd also like to allocate 2 or more slices on the same GPU

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants