Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

contiv-netmaster always Pending #363

Open
amwork2010 opened this issue Dec 27, 2018 · 0 comments
Open

contiv-netmaster always Pending #363

amwork2010 opened this issue Dec 27, 2018 · 0 comments

Comments

@amwork2010
Copy link

CentOS Linux release 7.6.1810
docker 18.09
kubeadm 1.13

kubeadm init --kubernetes-version=v1.13.0 --pod-network-cidr=10.10.0.0/16 --apiserver-advertise-address=192.168.55.31
wget https://github.com/contiv/install/releases/download/1.2.0/contiv-1.2.0.tgz
tar zxvf contiv-1.2.0.tgz
cd contiv-1.2.0/
./install/k8s/install.sh -n 192.168.55.31
......
contiv netmaster is not ready !!

kubectl get po --all-namespaces

NAMESPACE NAME READY STATUS RESTARTS AGE
kube-system contiv-netmaster-tqjnh 0/3 Pending 0 43s
kube-system coredns-86c58d9df4-hzt6d 0/1 Pending 0 71m
kube-system coredns-86c58d9df4-zwn9d 0/1 Pending 0 71m
kube-system etcd-kubecontiv1 1/1 Running 0 70m
kube-system kube-apiserver-kubecontiv1 1/1 Running 0 70m
kube-system kube-controller-manager-kubecontiv1 1/1 Running 0 70m
kube-system kube-proxy-f79dv 1/1 Running 0 71m
kube-system kube-scheduler-kubecontiv1 1/1 Running 0 70m

kubectl describe pod contiv-netmaster-tqjnh -n kube-system

Name: contiv-netmaster-tqjnh
Namespace: kube-system
Priority: 0
PriorityClassName:
Node:
Labels: k8s-app=contiv-netmaster
Annotations: prometheus.io/port: 9005
prometheus.io/scrape: true
scheduler.alpha.kubernetes.io/critical-pod:
Status: Pending
IP:
Controlled By: ReplicaSet/contiv-netmaster
Containers:
netmaster-exporter:
Image: contiv/stats
Port:
Host Port:
Environment:
CONTIV_ETCD: <set to the key 'contiv_etcd' of config map 'contiv-config'> Optional: false
EXPORTER_MODE: netmaster
Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from contiv-netmaster-token-l5dxj (ro)
contiv-netmaster:
Image: contiv/netplugin:1.2.0
Port:
Host Port:
Environment:
CONTIV_ROLE: netmaster
CONTIV_NETMASTER_MODE: <set to the key 'contiv_mode' of config map 'contiv-config'> Optional: false
CONTIV_NETMASTER_ETCD_ENDPOINTS: <set to the key 'contiv_etcd' of config map 'contiv-config'> Optional: false
CONTIV_K8S_CONFIG: <set to the key 'contiv_k8s_config' of config map 'contiv-config'> Optional: false
CONTIV_NETMASTER_FORWARD_MODE: <set to the key 'contiv_fwdmode' of config map 'contiv-config'> Optional: false
CONTIV_NETMASTER_NET_MODE: <set to the key 'contiv_netmode' of config map 'contiv-config'> Optional: false
Mounts:
/var/contiv from var-contiv (rw)
/var/log/contiv from var-log-contiv (rw)
/var/run/secrets/kubernetes.io/serviceaccount from contiv-netmaster-token-l5dxj (ro)
contiv-api-proxy:
Image: contiv/auth_proxy:1.2.0
Port:
Host Port:
Args:
--tls-key-file=/var/contiv/auth_proxy_key.pem
--tls-certificate=/var/contiv/auth_proxy_cert.pem
--data-store-address=$(STORE_URL)
--data-store-driver=$(STORE_DRIVER)
--netmaster-address=localhost:9999
Environment:
NO_NETMASTER_STARTUP_CHECK: 0
STORE_URL: <set to the key 'contiv_etcd' of config map 'contiv-config'> Optional: false
STORE_DRIVER: etcd
Mounts:
/var/contiv from var-contiv (rw)
/var/run/secrets/kubernetes.io/serviceaccount from contiv-netmaster-token-l5dxj (ro)
Conditions:
Type Status
PodScheduled False
Volumes:
var-contiv:
Type: HostPath (bare host directory volume)
Path: /var/contiv
HostPathType:
var-log-contiv:
Type: HostPath (bare host directory volume)
Path: /var/log/contiv
HostPathType:
contiv-netmaster-token-l5dxj:
Type: Secret (a volume populated by a Secret)
SecretName: contiv-netmaster-token-l5dxj
Optional: false
QoS Class: BestEffort
Node-Selectors: node-role.kubernetes.io/master=
Tolerations: node-role.kubernetes.io/master:NoSchedule
node.kubernetes.io/not-ready:NoExecute for 300s
node.kubernetes.io/unreachable:NoExecute for 300s
Events:
Type Reason Age From Message


Warning FailedScheduling 7s (x11 over 94s) default-scheduler 0/1 nodes are available: 1 node(s) had taints that the pod didn't tolerate.

How to solve this problem?
Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant