-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
BUG: predector exit status (137) #98
Comments
I got above message on bash terminal |
Hi @bioinfoanalyst, Thanks for reaching out. Regarding SignalP exiting 137. Exit code 137 indicates that the process has run out of memory while running. The automatic retry should have worked, and if so you can safely ignore the message. Regarding your attached error log So this is a rare occurrence and it should be fine to just re-run the code and ignore it. Regarding runtime I can't answer this without more details. Can you please share some details of the command and configuration that you're running it with? Are you running all of the proteomes separately or all at once? |
Thanks for responding.
I am solely using server these days. It is not shared with others and not other commands are running I run only predector command at a time without running other commands thinking of same what you suggested.
RAM is free to use. I think i need to do settings in config file but i do
not know how to do I am using following nextflow command:
**nextflow run -resume -r 1.2.7 -profile docker_sudo ccdmb/predector --proteome "/home/fungus/Data/Proteomes/*.faa" **
Note that if I use c32,r250 in the command like below:
**nextflow run -resume -r 1.2.7 -profile c32,r250,docker_sudo ccdmb/predector --proteome "/home/fungus/Data/Proteomes/*.faa"**
it gives following on console:
**Unknown configuration profile: 'c32'
Did you mean one of these?
r32**
I can only use given profiles in the usage like maximum of c16,r64
**nextflow run -resume -r 1.2.7 -profile c16,r64,docker_sudo ccdmb/predector --proteome "/home/fungus/Data/Proteomes/*.faa"**
How to use customized more cpus and memory?
…On Mon, Dec 2, 2024 at 2:44 PM Darcy Jones ***@***.***> wrote:
Hi @bioinfoanalyst <https://github.com/bioinfoanalyst>,
Thanks for reaching out.
*Regarding SignalP exiting 137.*
Exit code 137 indicates that the process has run out of memory while
running.
Is the server you're running shared with other people?
It might just be that someone else was using a lot of RAM that day.
The automatic retry should have worked, and if so you can safely ignore
the message.
If not, try running again and check back in if it fails.
*Regarding your attached error log*
So this is a rare occurrence and it should be fine to just re-run the code
and ignore it.
For more explanation, i create a temporary directory with a name based on
the process it's running in (tmpdir$$). I'm not sure of the details but
in rare cases you can have different processes spawn with the same number
as a race condition.
It's a known issue but so uncommon that normally you shouldn't need to
worry about it. I'll add part of a random number to reduce the possibility
of this happening further.
*Regarding runtime*
I can't answer this without more details.
I can run predector for a single proteome in 3 hours on a regular laptop,
and I've run predector in 2 hours for >200 proteomes on a compute cluster.
I'd expect something with 96 cpus to get through your proteomes in under a
day, but it depends heavily on how you've set it up to run.
Can you please share some details of the command and configuration that
you're running it with?
Is the computer shared with other people that could be competing for
resources?
Is top (or htop) showing that it's using CPUs as you've configured it to?
When you say proteome size range is 7-9 MB do you mean the fasta file size
is about 7 megabytes? So around 15000 sequences i'm guessing?
Are you running all of the proteomes separately or all at once?
Predector internally removes duplicate sequences and reduplicates them at
the end, so if your proteomes are fairly similar it can speed things up
quite a lot to run them all in one go.
—
Reply to this email directly, view it on GitHub
<#98 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/BFSGYQIVLIVOH7RPHEL2J4L2DQTZNAVCNFSM6AAAAABSUZCBWSVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDKMJRGAZTSNZQGA>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
About configuration I am using default as given in the usage Using docker:
*nextflow run -resume -r 1.2.7 -profile docker_sudo ccdmb/predector
--proteome "/home/fungus/Data/Proteomes/*.faa" *
…On Mon, Dec 2, 2024 at 2:44 PM Darcy Jones ***@***.***> wrote:
Hi @bioinfoanalyst <https://github.com/bioinfoanalyst>,
Thanks for reaching out.
*Regarding SignalP exiting 137.*
Exit code 137 indicates that the process has run out of memory while
running.
Is the server you're running shared with other people?
It might just be that someone else was using a lot of RAM that day.
The automatic retry should have worked, and if so you can safely ignore
the message.
If not, try running again and check back in if it fails.
*Regarding your attached error log*
So this is a rare occurrence and it should be fine to just re-run the code
and ignore it.
For more explanation, i create a temporary directory with a name based on
the process it's running in (tmpdir$$). I'm not sure of the details but
in rare cases you can have different processes spawn with the same number
as a race condition.
It's a known issue but so uncommon that normally you shouldn't need to
worry about it. I'll add part of a random number to reduce the possibility
of this happening further.
*Regarding runtime*
I can't answer this without more details.
I can run predector for a single proteome in 3 hours on a regular laptop,
and I've run predector in 2 hours for >200 proteomes on a compute cluster.
I'd expect something with 96 cpus to get through your proteomes in under a
day, but it depends heavily on how you've set it up to run.
Can you please share some details of the command and configuration that
you're running it with?
Is the computer shared with other people that could be competing for
resources?
Is top (or htop) showing that it's using CPUs as you've configured it to?
When you say proteome size range is 7-9 MB do you mean the fasta file size
is about 7 megabytes? So around 15000 sequences i'm guessing?
Are you running all of the proteomes separately or all at once?
Predector internally removes duplicate sequences and reduplicates them at
the end, so if your proteomes are fairly similar it can speed things up
quite a lot to run them all in one go.
—
Reply to this email directly, view it on GitHub
<#98 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/BFSGYQIVLIVOH7RPHEL2J4L2DQTZNAVCNFSM6AAAAABSUZCBWSVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDKMJRGAZTSNZQGA>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
Thanks for responding @darcyabjones! I am solely using server these days. It is not shared with others and not other commands are running I run only predector command at a time without running other commands thinking of same what you suggested. RAM is free to use. I think i need to do settings in config file but i do nextflow run -resume -r 1.2.7 -profile docker_sudo ccdmb/predector --proteome "/home/fungus/Data/Proteomes/*.faa" Note that if I use c32,r250 in the command like below: nextflow run -resume -r 1.2.7 -profile c32,r250,docker_sudo ccdmb/predector --proteome "/home/fungus/Data/Proteomes/*.faa" it gives following on console: **Unknown configuration profile: 'c32' Did you mean one of these? I can only use given profiles in the usage like maximum of c16,r64 nextflow run -resume -r 1.2.7 -profile c16,r64,docker_sudo ccdmb/predector --proteome "/home/fungus/Data/Proteomes/*.faa" How to use customized more cpus and memory? |
There's a section in the documentation on customising the configuration. In your case you'll just need to specify 96 cpus for the large jobs. Have a look at the documentation, and try something below. Save this as
try running like this:
Let me know how you go :) Cheers, |
This time I got following error on console:
The '.nextflow.log' file is attached below |
By rerunning, no error this time but processing is stuck at: Plus 4 more processes waiting for tasks… I will let you know about the progress by next 12 hours. And yes each fast file contains about 15000 sequences. I have added screenshots of top, htop and predector commands below: |
Hi there, Your server should work fine with all CPUs at 100% use, thread swapping will allow normal OS processes to run as some processes are stopped or spawned. Unless you have critical processes that need to use CPUs (like a network filesystem?) you shouldn't need to reduce this. However, to answer your question, just reduce the number of CPUs from 96 in the configuration file. You could also try to set the nextflow |
Dear @darcyabjones! I encountered following error after 15 hours of running command: Thanks
|
Dear all, I got following error while running predector command.
How to avoid this or should i ignore it?
Also predector command is taking days to for 145 fungus proteomes (proteome size range from 7-9mb ). What can be the estimated completion time? Note that server contains 96 cores, 440.5 GiB memory, 10tb space
The text was updated successfully, but these errors were encountered: