Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Pods stuck with status ContainerCreating #30177

Closed
montanaflynn opened this issue Aug 6, 2016 · 3 comments
Closed

Pods stuck with status ContainerCreating #30177

montanaflynn opened this issue Aug 6, 2016 · 3 comments
Labels
area/kubectl kind/bug Categorizes issue or PR as related to a bug. sig/storage Categorizes an issue or PR as relevant to SIG Storage.

Comments

@montanaflynn
Copy link

For some reason today when I rolled out a new version to one of our deployments the pod got stuck in ContainerCreating with this error events:

1h          1m         37        some-api-2275263275-01pq7              Pod                                             Warning   FailedMount               {kubelet gke-cluster-1-default-pool-4399eaa3-os4v}      Unable to mount volumes for pod "some-api-2275263275-01pq7_default(afc5ae68-5b5e-11e6-afbb-42010a800105)": timeout expired waiting for volumes to attach/mount for pod "some-api-2275263275-01pq7"/"default". list of unattached/unmounted volumes=[default-token-880jy]
1h          1m         37        some-api-2275263275-01pq7              Pod                                             Warning   FailedSync                {kubelet gke-cluster-1-default-pool-4399eaa3-os4v}      Error syncing pod, skipping: timeout expired waiting for volumes to attach/mount for pod "some-api-2275263275-01pq7"/"default". list of unattached/unmounted volumes=[default-token-880jy]

I then attempted to scale the cluster and more than 75% of the previously running pods switched to ContainerCreating and also got stuck there. This caused widespread failure in our system and I had to quickly create a new cluster.

We're using google cloud platform's container engine and the cluster version is 1.3.2.

@saad-ali
Copy link
Member

saad-ali commented Aug 6, 2016

@montanaflynn There were a number of storage related issues with v1.3.2 that were fixed with v1.3.4. You probably hit one of those.

If you share the complete /var/log/kubelet log from a node with a stuck deployment I can take a look and confirm if it's a known issue or not. I'd need your GKE project name/cluster name/zone as well to grab your master logs. Feel free to email me if you don't want to share publicly.

@abawany
Copy link

abawany commented Aug 10, 2016

I saw a similar issue with v1.3.3 but in my case, the root cause was a lot more pedestrian. My deployment requires a secrets volume and I had forgotten to create the associated secret for the cluster to which I was trying to perform the new deployment. I saw no errors when using kubectl describe or kubectl logs but eventually realized that the deployment stayed stuck in ContainerCreating state (without logs afaict) if a volume that it depended on is missing.

@timothysc timothysc added kind/bug Categorizes issue or PR as related to a bug. sig/storage Categorizes an issue or PR as relevant to SIG Storage. labels Aug 12, 2016
@calebamiles calebamiles modified the milestone: v1.6 Mar 8, 2017
@grodrigues3 grodrigues3 removed this from the v1.6 milestone Mar 10, 2017
@grodrigues3
Copy link
Contributor

This issue is stale. Closing.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area/kubectl kind/bug Categorizes issue or PR as related to a bug. sig/storage Categorizes an issue or PR as relevant to SIG Storage.
Projects
None yet
Development

No branches or pull requests

7 participants