-
-
Notifications
You must be signed in to change notification settings - Fork 2.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[🐛 Bug]: Each test-execution starts multiple jobs #1904
Comments
@maxnitze, thank you for creating this issue. We will troubleshoot it as soon as we can. Info for maintainersTriage this issue by using labels.
If information is missing, add a helpful comment and then
If the issue is a question, add the
If the issue is valid but there is no time to troubleshoot it, consider adding the
If the issue requires changes or fixes from an external project (e.g., ChromeDriver, GeckoDriver, MSEdgeDriver, W3C),
add the applicable
After troubleshooting the issue, please add the Thank you! |
Can you share the test script you are using to see this behavior? |
Hey @diemol , I asked the KEDA project as well. And it seems the issue is with the See here: kedacore/keda#4833 Is there any specific reason the default is set to
|
To answer your question: I have Geb tests for some of my applications. To connect to the Grid I use the |
@msvticket do you know? |
For reference: It was set to I could not find any discussion about the strategy in the PR. |
I am seeing similar behavior with scalingType: deployment . Kubernetes version is 1.23 for me as well. I will try following and share results:
|
We currently experience a problem with the Maybe that's what was tried to be fixed by using the |
Updates with scalingType: deployment . We have seen improvements after increasing the timeouts in the ingress. The pending sessions are not there anymore. |
Your mileage may vary apparently. For me it worked much better with |
That might be another issue (we did not have issues with too slow startup though). A bigger problem is the calculation of the scaling itself. I dug deeper into the KEDA code and found out, that the This issue is exactly what the
|
The issue was not only that we started too many pods, but rather, that additional jobs were started which never finished. I had this in a test setup with only a single session though. I'm not sure, if not later on another session might be taken over by the additional job. Do you have any experience there? |
Which is the same thing.
Yes it would. |
Exactly. Which is why I choose |
I have been experimenting with both type of scaling strategies (job/deployment) and seeing multiple jobs getting triggered . In one occurrence it started 16 Jobs for just two test cases. |
Any update on this one? we are also started having this issue after upgrading KEDA to 2.12.0 from 2.11.1 |
Fortunately (or unfortunately for you) we don't have the problem anymore. This was happening when we had a test setup, that only one application used at the time. When we scaled this up it just went away. We are running 100s of jobs daily now and no issues with "extra spawned jobs" so far. Sorry, that I cannot be of more help :/ |
See also my comment here: kedacore/keda#4833 (comment) |
You can follow https://github.com/SeleniumHQ/docker-selenium/tree/trunk/.keda |
@VietND96 Using the Selenium Grid Helm chart v0.36.1 and therefore
Because I expect my sessions to take several minutes each (CI tests), and we run into test timeouts for about half of the sessions when using |
@quiqueg, thanks for your feedback, I am checking on that. I also set up a test with strategy |
What happened?
When I start selenium tests using the grid there are always two jobs started.
There is one started immediately. Once it is up-and-running a second is scheduled. The second job is then used for the test. After the test is done only the second is finished. The other keeps on running (doing nothing). Today I stopped one, that was in running state the whole weekend.
The second is only started up a soon as the first is ready. I saw this when it was scheduled on a node, that did not have the image yet. It took about 2:30m to pull it. Only after that was done the second job got scheduled. First I thought this might have something to do with a timeout, because it took to long to pull the image. But it also happens if the image is available and the first job only takes seconds to get ready.
Command used to start Selenium Grid with Docker
I installed the Grid from the Helm chart using an existing KEDA installation.
My Kubernetes cluster is in version 1.23.
Relevant log output
I only put the KEDA log in the form, as I could not see any interesting output in the Grid logs.
Operating System
Kubernetes 1.23 on Flatcar Linux
Docker Selenium version (tag)
4.10.0-20230607
The text was updated successfully, but these errors were encountered: