-
Notifications
You must be signed in to change notification settings - Fork 118
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Running BUSCO with your own database #697
Comments
Without really knowing how BUSCO works (sorry), I have a suspicion... could you try again but with I've tried adding Lines 30 to 38 in d4e6cac
|
Hi @jfy133 I have started a run now with your modification, I let you know when I have a result. |
Hi, my run has finished but stopped at the BUSCO step and with an error. The error message:
My command to run the job was this:
When I checked the error in the work directory I find this in the
and the log file from BUSCO gave this output:
But that does not give a lot of info, so I checked the folder called BUSCO which contains a log directory.
The last line indicates that the archaea directory is not found. BUSCO indicates this directory location
but on my system the location is this:
which is without the "lineages" subfolder. So I just tried running the workflow againafter making my BUSCO database look the same as in the output from BUSCO
With all the taxa directories dropped in lineages. And that seems to work, the output from one of the jobs
|
So it looks like the error with BUSCO running offline, had to do with the folder structure of the BUSCO database directory. Modifying the directory by adding a subfolder called lineages resolved the issue. But it does not solve it completely. I ran into this issue as well: #545 and I checked this issue https://gitlab.com/ezlab/busco/-/issues/324 Which shows that I should set-up my database correctly. I therefor used the Busco version 5.5.0 that is installed on our cluster to set-up the database correctly with this command:
Should have done that earlier :-) Will try again, if my database has this folder structure:
|
Okay. I now have set-up a correct busco database, and I ran the pipeline with the commands
And the Busco step finished without any problems. Now I went back to the original job which was the reason for this issue and check the logs from BUSCO. There I also now observed that the error was due to the database not being correctly set-up. So I tried to run my original command for the pipeline with the now correctly set-up database.
And that works without a problem So the whole issue was due to me not setting up the BUSCO database correctly. So @jfy133 I will close this issue as it was not a problem of the MAG workflow, but of the database needed for BUSCO. |
Ok thanks for follow up and clarifications, we appreciate it! |
Description of the bug
Hi I tried to use BUSCO for the bin checking. Since our cluster does not allow to download things when inside a slurm job, I download the database and installed it. I did not use BUSCO for it as suggested in issue: #545 . But my BUSCO job fails in two ways. It seems to not find the option: --auto-lineage-prok and it does not want to us the database I downloaded.
My params file looks like this:
When I check the log file from BUSCO I get this :
I saw in the BUSCO manual that there is an flag called:
--offline
that you can use when you provide it with your own database. I see in the .command.sh file that it is there. But that is not used here when I run BUSCO, with my own databasethe .command.sh file looks like this:
I am not understanding what is the error here, it looks like the db location is only using the last bit of the db location,
Command used and terminal output
My nextflow command was:
nextflow run nf-core/mag -r 3.0.3 -profile apptainer -work-dir $USERWORK/nf_mag -resume -c saga_mag.simple.config -params-file params_test_2.json
Relevant files
No response
System information
Nextflow version: 24.04.3
Hardware: HPC
executor: Slurm
Container engine: Apptainer
OS : CentOS linux
The text was updated successfully, but these errors were encountered: