Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CI fails when other jobs are running because of time out. #832

Open
n-eiling opened this issue Oct 31, 2024 · 14 comments
Open

CI fails when other jobs are running because of time out. #832

n-eiling opened this issue Oct 31, 2024 · 14 comments
Labels
bug Something isn't working ci

Comments

@n-eiling
Copy link
Member

See https://git.rwth-aachen.de/acs/public/villas/node/-/jobs/5510111 for details.

Relevant output:

[TIME]  pipe-loopback-zeromq (ran for more then 1m)
[SKIP]  pipe-python-protobuf: Test is broken
[SKIP]  relay: Test is broken
[PASS]  signal (ran for .100993659s)
[PASS]  test-config (ran for 7.858488891s)
Summary: 1 of 61 tests failed.
   Timedout: 1
   Skipped: 18

Maybe we can remove the time out?

@n-eiling n-eiling added bug Something isn't working ci labels Oct 31, 2024
@stv0g
Copy link
Contributor

stv0g commented Nov 29, 2024

I guess the timeout is probably more of an indicator for a broken test?

Shouldn't we attempt to fix the test, or disable it?

@n-eiling
Copy link
Member Author

n-eiling commented Dec 2, 2024

The test does work if it I rerun it. So it seems to work, but is just too slow if run in parallel with everything else.
I haven't investigated thoroughly though and I don't know if a similar functionality is tested in another test.

@stv0g
Copy link
Contributor

stv0g commented Feb 10, 2025

Is this still happening? I've seen it not myself yet..

@n-eiling
Copy link
Member Author

Haven't seen it in a while - but that doesn't mean it fixed itself. Don't you think it's bad practice to rely on timing in the CI? The runner might be really slow sometimes.

@stv0g
Copy link
Contributor

stv0g commented Feb 13, 2025

Yeah, its a valid consern.

I have added those timeout checks to avoid clogging the CI with jobs which are dead-locked.
Maybe we should raise the timeout to something like 1h?

@n-eiling
Copy link
Member Author

sounds good to me

@IgnoreWarnings
Copy link
Collaborator

Haven't seen it in a while - but that doesn't mean it fixed itself. Don't you think it's bad practice to rely on timing in the CI? The runner might be really slow sometimes.

Is this still happening? I've seen it not myself yet..

I got the timeout yesterday, it is still there.

@IgnoreWarnings
Copy link
Collaborator

I get it consistently now on every single commit.

IgnoreWarnings added a commit to IgnoreWarnings/node that referenced this issue Feb 24, 2025
@IgnoreWarnings
Copy link
Collaborator

Increasing the timeout didnt help, the integrationtest seems to get stuck somehow.

Image

@IgnoreWarnings
Copy link
Collaborator

I guess the timeout is probably more of an indicator for a broken test?

Shouldn't we attempt to fix the test, or disable it?

The test does work if it I rerun it. So it seems to work, but is just too slow if run in parallel with everything else. I haven't investigated thoroughly though and I don't know if a similar functionality is tested in another test.

It looks like @stv0g is right that the test is broken. I increased the timeout and the CI hangs on integration test.

@IgnoreWarnings
Copy link
Collaborator

"[TIME] pipe-loopback-nanomsg (ran for more then 15m)"

@IgnoreWarnings
Copy link
Collaborator

But it looks like before it was on another test: [TIME] pipe-loopback-zeromq (ran for more then 1m)

@stv0g
Copy link
Contributor

stv0g commented Feb 25, 2025

Okay, if the tests are broken, we need to figure out what broke them as they used to work in previous commits.

Maybe git bisect could help..

@n-eiling
Copy link
Member Author

I suspect this is more due to docker / strange interactions between the tests.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working ci
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants