-
Notifications
You must be signed in to change notification settings - Fork 634
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Slurm oom-kill due to memory is ignored. #5332
Labels
Comments
Some questions:
I'm wondering if the failure happened with the process script or with the copying of task outputs. Possibly related to #3711 |
Hi @bentsherman
Do you mean this output?
I'd doubt that this failure is linked to the output step, because it happens way before. As far as i can tell the job is killed by Slurm while running and the output is generated ignoring the kill. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Bug report
Expected behavior and actual behavior
Slurm jobs that run out of memory get oom-killed. In nearly all cases this works. In an awk process i run there is excessive RAM usage that gets only logged in .command.log but is ignored in the nextflow process. This results in premature end of the awk processes leading to corrupted output.
Steps to reproduce the problem
The following code produces issues with fastq.gz files with 20 million reads or more.
Program output
In the nextflow.log the jobs look as if they are successful.
Environment
Additional context
(Add any other context about the problem here)
The text was updated successfully, but these errors were encountered: