You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
a user reports via e-mail that they have a custom database of about 300k genomes, and their sourmash index job is being killed - presumably due to being out of memory. I responded with the following suggestions:
what version of sourmash are you using?
this is likely due to running out of memory; a few thoughts -
first, try increasing the --scaled with which you are indexing the database. For example, you can do sourmash index --scaled=10000 … and it should reduce the memory usage.
second, try decreasing the size of the bloom filters you are using for sourmash index - maybe try sourmash index -x 1e4, I think. On the downside this will decrease speed of database search ;(.
third, you can split the signatures up. It’s not that much slower to search 6 databases of 50k signatures each, and the build process should be much faster/smaller for those smaller databases. Just put them all on the command line for sourmash search, gather, etc.
a user reports via e-mail that they have a custom database of about 300k genomes, and their
sourmash index
job is being killed - presumably due to being out of memory. I responded with the following suggestions:The text was updated successfully, but these errors were encountered: