You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We were experiencing constant segfaults in keepalived in our deployment, which was causing the MASTER to frequently bounce between nodes and services to be inconsistently reachable (frequent dropped connections).
Since there was no apparent cause, we decided to switch to layer2 mode, I switched the cluster to IPVS mode, and re-deployed openelb via helm chart. However I have a new problem;
On startup, it still deploys keepalived and when it does, all services go unreachable and even openelb-manager now has trouble reaching the k8s apiservice. There is a log entry that it is "cleaning ipvs configuration" which doesn't sound very good in a k8s IPVS environment.
A quick glance at the code looks to me like all the Speakers start up no matter the deployed config, but I'm only quickly reading the code.
I chose to kill the daemonset as I don't want it to segfault and cause problems with the LB services.
My question is why does keepalived daemonset get started if there are no eips with protocol "vip" ?
Other than that, services seem more reliable in Layer2 with the simple Speaker setup. the services work fine, the ARPs are being sent out, etc.
** Keepalived Startup Log **
I0305 21:20:04.413733 7 main.go:196] Creating API server client for https://10.66.0.1:443
I0305 21:20:04.414209 7 main.go:136] starting LVS configuration
I0305 21:20:05.519985 7 main.go:388] No interface was provided, proceeding with the node's default: eth0
I0305 21:20:05.522272 7 main.go:242] cleaning ipvs configuration
E0305 21:20:05.524371 7 reflector.go:126] github.com/aledbf/kube-keepalived-vip/pkg/controller/main.go:293: Failed to list
*v1.Service: Get https://10.66.0.1:443/api/v1/services?limit=500&resourceVersion=0: write tcp 10.66.0.1:55816->100.66.0.1:443
: write: broken pipe
As I dig a little deeper, I realize that the dependent container https://github.com/aledbf/kube-keepalived-vip is an archived project that hasn't been updated in 5 years. That's a lifetime in k8s ecosystem years.
v2.0.19 of keepalived (circa 2022; 2.2.8 is May 2023).
What are the plans around this? There should definitely be an option to not be forced to run keepalived-vip containers if they are not supported/maintained.
fwiw, for anyone who desires not to have the keepalived pods running if you are not using VIP mode, I simply applied an empty nodeSelector to the openelb-keepalived-vip daemonset:
Describe the bug
We were experiencing constant segfaults in keepalived in our deployment, which was causing the MASTER to frequently bounce between nodes and services to be inconsistently reachable (frequent dropped connections).
Since there was no apparent cause, we decided to switch to layer2 mode, I switched the cluster to IPVS mode, and re-deployed openelb via helm chart. However I have a new problem;
On startup, it still deploys keepalived and when it does, all services go unreachable and even openelb-manager now has trouble reaching the k8s apiservice. There is a log entry that it is "cleaning ipvs configuration" which doesn't sound very good in a k8s IPVS environment.
A quick glance at the code looks to me like all the Speakers start up no matter the deployed config, but I'm only quickly reading the code.
I chose to kill the daemonset as I don't want it to segfault and cause problems with the LB services.
My question is why does keepalived daemonset get started if there are no
eips
with protocol "vip" ?Other than that, services seem more reliable in Layer2 with the simple Speaker setup. the services work fine, the ARPs are being sent out, etc.
** Keepalived Startup Log **
Output
Version Info
The text was updated successfully, but these errors were encountered: