-
Notifications
You must be signed in to change notification settings - Fork 567
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG][v1.6.2-rc1] Workload pod got stuck in ContainerStatusUnknown after node shutdown and reboot #8550
Comments
cc @c3y1huang |
|
This issue doesn't look like a Longhorn's bug, as the deployment pod eviction was initiated by the Kubelet and it's expected that the Kubelet would handle the cleanup as well. There's an ongoing upstream discussion that seems to be related: kubernetes/kubernetes#122160 We could consider enhancing the test case, filtering out pods terminated by the Kubelet. |
Pre Ready-For-Testing Checklist
|
Closing as this has been tested. longhorn/longhorn-tests#1902 (review) |
Describe the bug
When running negative test case
Power Off Node One By Once For More Than Pod Eviction Timeout While Workload Heavy Writing
, accidentally encountered a deployment pod got stuck inContainerStatusUnknown
forever.The test case steps are like:
In https://ci.longhorn.io/job/private/job/longhorn-e2e-test/557/, accidentally encountered after nodes rebooted in step 7 ~ 9, the pod of deployment with rwo strict-local volume got stuck in
ContainerStatusUnknown
forever:To Reproduce
Expected behavior
Running negative test case
Power Off Node One By Once For More Than Pod Eviction Timeout While Workload Heavy Writing
repeatedly.Support bundle for troubleshooting
supportbundle_26ac5094-4709-4ab5-bb99-867e4b13cb8f_2024-05-12T02-28-45Z.zip
Environment
Additional context
The text was updated successfully, but these errors were encountered: