-
Notifications
You must be signed in to change notification settings - Fork 24
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bug: wasm log aws.appmesh.egress_http_stats: panicked at 'called Option::unwrap()
on a None
value', amzn-appmesh-aggregate-stats/src/http/egress.rs:172:9
#485
Comments
Hi @tnsardesai, I see the stack trace is showing different line of code for the failure.
For this log, do you know which specific version of Envoy it is associated with? I agree this may be a separate issue that we need to address apart from the original issue #484. Another question, are you seeing this error in older version of Envoy as well? |
Oh I didn't notice the difference in line number. Looking back at our logs In v1.25.1.0 we see
In v1.27.2.0 we see
I looked back at logs from November (when our applications were all on v1.24.0.0) and I actually see a similar log from that version too
|
Thanks for getting back with the information so quickly. Yeah this looks like an issue that exists for a long time. We have found the corresponding code. We are working on a fix for this along with #484. One question - was this issue affecting your fleet in anyway? Since it's been there for a while. |
The issue in #484 resulted in metrics getting dropped. Our setup is similar to https://github.com/aws/aws-app-mesh-examples/tree/main/walkthroughs/howto-metrics-extension-ecs#optional-filtering-metrics-with-the-aws-distro-for-opentelemetry. If I had to guess this issue also caused something similar but we never noticed it because the impact was very low. Looking back at our logs I see this panic log a maximum of 3 times for a single service in one day but between scaling events and deployments we never noticed any impact. |
This issue is fixed by Envoy release v1.27.3.0 #486 |
Summary
What are you observing that doesn't seem right?
Seeing a panic during normal operations which is not expected! This panic leads to metrics getting dropped
Steps to Reproduce
What are the steps you can take to reproduce this issue?
No idea. It seems to happen when a request to bad request is made by some bot or attacker.
Are you currently working around this issue?
How are you currently solving this problem?
We are not
Additional context
Anything else we should know?
I saw this log in both v1.25.1.0 and v1.27.2.0
Attachments
If you think you might have additional information that you'd like to include via an attachment, please do - we'll take a look. (Remember to remove any personally-identifiable information.)
I see requests like
[2024-01-13T17:32:43.610Z] "GET /oauth/lookup-ip HTTP/1.1" 400 - 0 0 22 21 "1 ����%2527%2522, 188.241.177.125, 130.176.91.135,127.0.0.1" "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/117.0.0.0 Safari/537.36" "6c222c47-5c31-9b08-820c-edca79a3f90b" "production--oauth--833ed5a3.prod.mesh" "127.0.0.1:80"
with badX-Forwarded-For
around the same time as this panic which might be relevantThe text was updated successfully, but these errors were encountered: