I have the following environment:
Pods: Pod0, Pod1 (launched as a k8s Job)
GPUs: GPU0, GPU1
GPU0 is dedicated to Pod0, and GPU1 is dedicated to Pod1.
There can be multiple Pod0s and Pod1s at the same time. If there are two Pod0s, only a single Pod0 can use GPU0, and another Pod0 should be in Pending
state until the first Pod0 finishes.
Is this workload possible?
Currently I use nvidia.com/gpu
resource setting but the only thing I can do is to set the number of GPU allocated to each Pod. I had tried setting NVIDIA_VISIBLE_DEVICES
but could not make it as I desired.