You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
On the 1.6 containerd release series this behavior started appearing with the v1.6.9 (the v1.6.8 release seems to be working fine). The offending PR over at containerd seems to be containerd/containerd@a91dd67 and has been ported to the 1.5.x releases to so the 1.5 containerd track should be also affected. The offending PR seems to be changing how the pod sandboxes are handled.
When deploying the gpu-operator the nvidia-container-toolkit is crashlooping. The container logs
Full logs: https://paste.ubuntu.com/p/f3byGQ4kpJ/
Here are the pods currently running:
On the 1.6 containerd release series this behavior started appearing with the v1.6.9 (the v1.6.8 release seems to be working fine). The offending PR over at containerd seems to be containerd/containerd@a91dd67 and has been ported to the 1.5.x releases to so the 1.5 containerd track should be also affected. The offending PR seems to be changing how the pod sandboxes are handled.
I am attaching the debug logs of an older containerd where the deployment is successful and a newer that is failing:
working-containerd.log
not-working-containerd.log
The text was updated successfully, but these errors were encountered: