Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Issue with k8s.io/docs/getting-started-guides/fedora/flannel_multi_node_cluster/ #5359

Closed
1 of 2 tasks
ghost opened this issue Sep 8, 2017 · 7 comments
Closed
1 of 2 tasks
Labels
lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed.

Comments

@ghost
Copy link

ghost commented Sep 8, 2017

This is a...

  • Feature Request
  • Bug Report

Problem:
based on this documentation it should be possible to ping pod 2 running on node 2 from pod 1 running on node 1 but this does not work

Proposed Solution:
unfortunately still investigating
possibly Docker option --ip-masq
Docker currently use
"com.docker.network.bridge.enable_ip_masquerade": "true",

Page to Update:
http://kubernetes.io/...

Additional Information:
plain Fedora 26 Server installation
node 1
cat /var/run/flannel/subnet.env
FLANNEL_NETWORK=10.244.0.0/16
FLANNEL_SUBNET=10.244.10.1/24
FLANNEL_MTU=1450
FLANNEL_IPMASQ=false
ip -4 a|grep inet
inet 127.0.0.1/8 scope host lo
inet 172.16.0.31/24 brd 172.16.0.255 scope global ens32
inet 10.244.10.0/32 scope global flannel.1
inet 10.244.10.1/24 scope global docker0
pod 1
ip a l eth0 | grep inet
inet 10.244.10.2/24 scope global eth0
node 2
cat /var/run/flannel/subnet.env
FLANNEL_NETWORK=10.244.0.0/16
FLANNEL_SUBNET=10.244.6.1/24
FLANNEL_MTU=1450
FLANNEL_IPMASQ=false
ip -4 a|grep inet
inet 127.0.0.1/8 scope host lo
inet 172.16.0.32/24 brd 172.16.0.255 scope global ens32
inet 10.244.6.0/32 scope global flannel.1
inet 10.244.6.1/24 scope global docker0
ip a l eth0 | grep inet
inet 10.244.6.2/24 scope global eth0

ping 10.244.6.2 from pod 1 fails
ping 10.244.6.1 from pod 1 success
ping 10.244.6.2 from node 1 fails
ping 10.244.10.2 from pod 2 fails
ping 10.244.10.1 from pod 2 success
ping 10.244.10.2 from node 1 fails

@ghost
Copy link
Author

ghost commented Sep 8, 2017

i successfully disabled dockers ip masquerading

docker network inspect bridge
"com.docker.network.bridge.enable_ip_masquerade": "true"
systemctl stop docker
systemctl stop flanneld
ip link delete docker0
vi /etc/sysconfig/flanneld
FLANNEL_OPTIONS="-ip-masq"
systemctl start flanneld
systemctl start docker
docker network inspect bridge
"com.docker.network.bridge.enable_ip_masquerade": "false"

now at least i can ping Pod 2 from Node 1 but pings between Pods are still not possible

@kairen
Copy link

kairen commented Sep 12, 2017

Hi @cyberh0me Which Docker version running on your node?

@kairen
Copy link

kairen commented Sep 12, 2017

This may be the effect of the previous Docker change, refer from moby/moby#14041.

@ghost
Copy link
Author

ghost commented Sep 12, 2017

docker-1.13.1-22.gitb5e3294.fc26.x86_64.rpm

@fejta-bot
Copy link

Issues go stale after 90d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale.
Stale issues rot after an additional 30d of inactivity and eventually close.

Prevent issues from auto-closing with an /lifecycle frozen comment.

If this issue is safe to close now please do so with /close.

Send feedback to sig-testing, kubernetes/test-infra and/or @fejta.
/lifecycle stale

@k8s-ci-robot k8s-ci-robot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Jan 5, 2018
@fejta-bot
Copy link

Stale issues rot after 30d of inactivity.
Mark the issue as fresh with /remove-lifecycle rotten.
Rotten issues close after an additional 30d of inactivity.

If this issue is safe to close now please do so with /close.

Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/lifecycle rotten
/remove-lifecycle stale

@k8s-ci-robot k8s-ci-robot added lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. and removed lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. labels Feb 9, 2018
@fejta-bot
Copy link

Rotten issues close after 30d of inactivity.
Reopen the issue with /reopen.
Mark the issue as fresh with /remove-lifecycle rotten.

Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/close

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed.
Projects
None yet
Development

No branches or pull requests

4 participants