New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Assembling big data #14
Comments
Hi Livia, Could you please post the log of the run? Plass should split up the work so it always fits into the available memory. Best regards, |
Sure... here is the log file I got the following message: |
Thanks a lot! How much memory does your machine have? Normally Plass try to split the database if it does not fit in memory. |
CentOS server, I can use up to 900Gb ram. |
So it seems that the extractorfs step is hanging, which mostly requires IO. Is it possible that the tmp folder is on some slow network share? One trick to reduce the amount of sequences extracted is to increase the minimum orf length with --min-length (default: 20). |
Hey,
I have a big dataset (>600M paired-end reads) and I am trying to generate a protein catalog using Plass. I am using the version 2.c7e35 in a server with 900Gb ram. The processing is ending without completion due to exceeding the resources requested. I am wondering if it is possible to tweak the parameters to allocate less memory.
Any input will be greatly appreciated.
Thanks,
Livia
The text was updated successfully, but these errors were encountered: