-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Process gets killed if several large files are input #68
Comments
I wonder if it's due to the system's OOM killer, were you running out of memory? (though that would imply there's a memory leak if all individual files do work). |
I was trying to search in logs to track down the cause, but did not find a way to identify what happened. Grepping for |
As far as I can see, there is no significant memory leak in FoLiA-txt I assume there is NO problem when that file is processed on its own? |
The files process fine, if I call the converter one by one. I experienced the same thing on other files too, when calling the converter on directories of large files -- there can be near 1 mln tokens per file.... Typically, after having converted the first file, the process is killed. |
Well, I just ran tests on some fairly small files, and here seems to be some random effect which makes the run to fail, but not always. |
OK, it is some multithreading problem I guess. A deadlock occurs. FoLiA-txt seems to 'stall' when running om multiple threads Best is to upgrade to the newest GIT version, which tells on how many threads you actually run. |
@pirolen the git master has a fix now, which hopefully fixes the deadlock |
Closing, considering it to be fixed |
Hi, on large files, the FoLiA-txt tool in the containerized foliautils gets killed.
I get:
It is not a big problem since one can call the tool separately per file, but thought to let you know.
Maybe it is better to call the tool per file in a shell script in the container, I did not try that.
The text was updated successfully, but these errors were encountered: