-
Notifications
You must be signed in to change notification settings - Fork 119
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merge rerun results when --flaky-test-attempts is used #456
Comments
Will it also fix I went through the discussion in Slack. I have an idea of how it can works with xml report out:
I have used similar flaky-test approach with another ui testing framework quite successful. |
Yep! I think the idea is if the test passes at least once then it's considered passed. |
Current situation. If
Right now, flank will aggregate all the result producing a JUnitReport.xml with 8 total tests and 4 failures. Instead of 8 result, we'd really like 2 results (one for each unique test executed). Proposed solution:
An alternative to 2. could be to dump the flaky info to a separate file inside each shard, but to me, it seems like keeping it in one location would be desirable. A couple of questions I haven't looked into yet:
|
Sounds like a great idea. But I'd suggest valuing the JUnit XSD specifications, meaning I think the (2) approach would be more suitable, even though (1) would be more userfriendly. Maybe we could also dump the |
I think we'd want to repeat them normally. It does seem like a weird use case.
I think modifying the report to include the data we need makes sense. The iOS JUnit is already custom from FTL so following JUnit XSD exactly isn't possible. |
I thought the whole idea of repeatTests was to measure flakiness of a test. It would seem redundant to use that in conjunction with flakyTestsAttempts, since you'd want the raw data anyway. Or am I missing another use case? |
A couple more small questions
|
That's my use case for it. The idea is simply repeating tests though. If I wanted to check for something else, like performance measurement to average the metrics, we may want both repeat and flaky test attempts.
We should probably leave them since the folder structure is intended to mirror what's on GCS. We're not deleting them on GCS so deleting them locally would be surprising I think.
Yeah if we are generating the file by merging XML then naming it appropriately makes sense. |
how it will consider the last rerun then? |
I think we'll download the results from all runs |
flakyTestAttempts runs tests multiple times. Currently any failures will be reported as a failure in Flank, even if the test passes on retry.
Fetch the test rerun results, and consider the full history of a test to determine if it's passed or not.
The text was updated successfully, but these errors were encountered: