-
Notifications
You must be signed in to change notification settings - Fork 349
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
K8s ingress: 502s from ALB when using multiple hosts #964
Comments
Also experiencing this issue but we only have one host in our ingress manifest. Talking with AWS they pulled logs from the ALB associated with the issue: "upstream prematurely closed connection while reading response header from upstream". skipper version v0.10.150 |
@seancurran157 thanks for commenting. From my understanding the message AWS wrote, means that ALB timedout in the backend call to skipper. |
We currently have skipper timeout lower than the ALB. Looks like skipper may not be responding with a 504 when the back end timeouts and the ALB responds with a 502. |
TODO:
|
Also experiencing similar issue, @herojan @seancurran157 could you please confirm #998 has fixed the issue or you still see 502 appearing from time to time ? |
|
@olevchyk I asked the team who reported it and they said they haven't seen the issue in the past 15 days worth of logs, so it seems to be gone |
For the records: skipper |
Hi, I've had a look and I don't see similar issues so I'll just provide the information here.
Using the K8s ingress controller with three different hosts, there are bursts of 502 errors happening regularly each day, often several times per day.
These 502 errors are happening at the ALB level, they never reach skipper. We know this because skipper does not print any logs about it and because the 502 returns html in the formatAWS uses, instead of the way skipper does.
Sample log statement from the application:
Sample of the ELB 02 errors over the past week:
Sample ingress file with hostnames and paths etc stripped out:
It reminds me of another issue where skipper was not receiving information about new pod ips quickly enough when nodes were rotated, and so redirected people to the old pod ips. I wonder if this is similar but a level up, with the ELB not getting information about skipper pod rotations quickly enough and directing to old skipper ips.
Let me know if more information is needed.
Edit: Updated the ELB image since I had linked the wrong one originally.
The text was updated successfully, but these errors were encountered: