Istio telemetry with als extension provider seems to have some sort of session stickiness #49831
Unanswered
joelmathew003
asked this question in
Q&A
Replies: 1 comment 14 replies
-
This is among replicas of the ALS pods, not your app pod, right? ALS is a gRPC streaming protocol (https://github.com/envoyproxy/envoy/blob/a0cc3001b1cf42396e401f0527bd56246dd2fba9/api/envoy/service/accesslog/v3/als.proto#L28) so I suspect once envoy opens a connection to one pod it will 'stick' to that replica. Istio has a similar issue in the connection to the control plane (Istiod); there we set a 30min max connection age to rebalance. Not perfect solution, though. |
Beta Was this translation helpful? Give feedback.
14 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I have a telemetry resource in my namespace which looks like:
and have described these providers in the istio configmap mesh
My
collector
want the access logs of the traffic in my namespace and does some processing. I have an hpa resource created forcollector
which scales on 60% utilization. On some stress testing(~200 tps) and new replicas are created by hpa, I see that the older replicas receive most of the traffic, even after the new replicas are created and ready. Sometimes the older replicas cross the limit and restarts, on which the newer replicas start receiving the load. Is there some sort of session stickiness that causes this? Ideally I want the load to get distributed evenly among all ready replicas.Beta Was this translation helpful? Give feedback.
All reactions