Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Investigate why e2e tests did not catch OOMKilled errors #3484

Open
jcreixell opened this issue Nov 21, 2024 · 4 comments
Open

Investigate why e2e tests did not catch OOMKilled errors #3484

jcreixell opened this issue Nov 21, 2024 · 4 comments

Comments

@jcreixell
Copy link
Contributor

Component(s)

auto-instrumentation

Describe the issue you're reporting

This is a follow up to #3479

Auto-instrumentation for some languages could lead to OOMKilled errors with default settings. This was fixed in #3473 but we need to understand why e2e tests did not catch this problem and modify them to ensure that they do.

@iblancasa
Copy link
Contributor

@IshwarKanse, can you take a look into this?

@IshwarKanse
Copy link
Contributor

I'll check on this next week.

@atoulme
Copy link
Contributor

atoulme commented Mar 12, 2025

@IshwarKanse any update?

@IshwarKanse
Copy link
Contributor

@atoulme In the e2e tests, we have asserts for container status to check that the containers in the pod are running. https://github.com/open-telemetry/opentelemetry-operator/blob/main/tests/e2e-instrumentation/instrumentation-python/01-assert.yaml#L68 https://github.com/open-telemetry/opentelemetry-operator/blob/main/tests/e2e-instrumentation/instrumentation-go/02-assert.yaml#L60 Also according to comment #3479 (comment) this could be caused due to disk speeds and specific to some envs. The issue is still under investigation and once we have a root cause, we can try to update our tests to detect it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants