-
Notifications
You must be signed in to change notification settings - Fork 317
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can't get vGPU licensed against my License Server #663
Comments
This is what i found out during the drivers creation:
But I do have the configmap with the token and the following helm values:
|
Ok, solved maunally editing the driver deployment after creation and rollout restarting it, adding the needed volumeMount and volume: volumeMounth:
volume:
|
Hi @urbaman please set As a matter of interest, what GPU Operator helm chart version are you using? NLS should be enabled by default in the helm chart. |
Ah, ok. You're right, I must have worked out the values from a previous version. I usually work with values files, so I modified an old one instead of creating a new one.
It could be helpful to have a complete helm values reference somewhere (threre's a partial one on the docs) Thank you very much! |
1. Quick Debug Information
2. Issue or feature description
Installed the helm chart to enable vGPU on my 9 workers, can't get it to be licensed
3. Steps to reproduce the issue
Install the helm chart, run a test worload:
4. Information to attach (optional if deemed irrelevant)
kubectl get pods -n OPERATOR_NAMESPACE
kubectl get ds -n OPERATOR_NAMESPACE
kubectl describe pod -n OPERATOR_NAMESPACE POD_NAME
kubectl logs -n OPERATOR_NAMESPACE POD_NAME --all-containers
nvidia-smi
from the driver container:kubectl exec DRIVER_POD_NAME -n OPERATOR_NAMESPACE -c nvidia-driver-ctr -- nvidia-smi
journalctl -u containerd > containerd.log
containerd.log
Collecting full debug bundle (optional):
NOTE: please refer to the must-gather script for debug data collected.
This bundle can be submitted to us via email: [email protected]
The text was updated successfully, but these errors were encountered: