/bacannot

Generic but comprehensive pipeline for prokaryotic genome annotation and interrogation with interactive reports and shiny app.

Primary LanguageNextflowGNU General Public License v3.0GPL-3.0

Cite with Zenodo GitHub release (latest by date including pre-releases) Documentation Nextflow run with docker run with singularity License Follow on Twitter

Open in Gitpod

bacannot pipeline

A generic but comprehensive bacterial annotation pipeline


See the documentation ยป

Report Bug ยท Request Feature

About

Bacannot is an easy to use nextflow docker-based pipeline that adopts state-of-the-art software for prokaryotic genome annotation. It is a wrapper around several tools that enables a better understanding of prokaryotic genomes.

Its main steps are:

Analysis steps Used software or databases
Genome assembly (if raw reads are given) Flye and Unicycler
Identification of closest 10 NCBI Refseq genomes RefSeq Masher
Generic annotation and gene prediction Prokka or Bakta
rRNA prediction barrnap
Classification within multi-locus sequence types (STs) mlst
KEGG KO annotation and visualization KofamScan and KEGGDecoder
Annotation of secondary metabolites antiSMASH
Methylation annotation Nanopolish
Annotation of antimicrobial (AMR) genes AMRFinderPlus, ARGminer, Resfinder and RGI
Annotation of virulence genes Victors and VFDB
Prophage sequences and genes annotation PHASTER, Phigaro and PhySpy
Annotation of integrative and conjugative elements ICEberg
Focused detection of insertion sequences digIS
In silico detection of plasmids Plasmidfinder and Platon
Prediction and visualization of genomic islands IslandPath-DIMOB and gff-toolbox
Custom annotation from formatted FASTA or NCBI protein IDs BLAST
Merge of annotation results bedtools
Genome Browser renderization JBrowse
Circos plot generation easy_circos
Renderization of automatic reports and shiny app for results interrogation R Markdown, Shiny and SequenceServer

๐ŸŽฏ In order to increase the accuracy of prokka annotation, this pipeline includes an additional HMM database to prokka's defaults. It can be either TIGRFAM (smaller but curated) or PGAP (bigger comprehensive NCBI database that contains TIGRFAM).

Release notes

Are you curious about changes between releases? See the changelog.

  • I strongly, vividly, mightily recommend the usage of the latest versions hosted in master branch, which is nextflow's default.
    • The latest will always have support, bug fixes and generally maitain the same processes (I mainly add things instead of removing) that also were in previous versions.
    • But, if you really want to execute an earlier release, please see the instructions for that.
  • Versions below 2.0 are no longer supported.

Further reading and complementary analyses

Moreover, this pipeline has two complementary pipelines (also written in nextflow) for NGS preprocessing and Genome assembly that can give the user a more thorough and robust workflow for bacterial genomics analyses.

Requirements

These images have been kept separate to not create massive Docker image and to avoid dependencies conflicts.

Installation

  1. If you don't have it already install Docker in your computer.

    • After installed, you need to download the required Docker images
    docker pull fmalmeida/bacannot:v3.2_misc    ;
    docker pull fmalmeida/bacannot:v3.2_perlenv ;
    docker pull fmalmeida/bacannot:v3.2_pyenv   ;
    docker pull fmalmeida/bacannot:v3.2_renv    ;
    docker pull fmalmeida/bacannot:jbrowse      ;

๐Ÿ”ฅ Nextflow can also automatically handle images download on the fly when executed. If docker has exceeded its download limit rates, please try again in a few hours.

  1. Install Nextflow (version 20.10 or higher):

    curl -s https://get.nextflow.io | bash
    
  2. Give it a try:

    nextflow run fmalmeida/bacannot -profile docker --help
    

๐Ÿ”ฅ To run the pipeline now users need to pass the -profile docker or -profile singularity parameter explicitly. The pipeline does not load it automatically anymore.

๐Ÿ”ฅ Users can get let the pipeline always updated with: nextflow pull fmalmeida/bacannot

Downloading and updating databases

Bacannot databases are not inside the docker images anymore to avoid huge images and problems with connections and limit rates with dockerhub.

Pre-formatted

Users can directly download pre-formatted databases from Zenodo: https://doi.org/10.5281/zenodo.7615811

Useful for standardization and also overcoming known issues that may arise when formatting databases with singularity profile.

I want to generate a new formatted database

To download and format a copy of required bacannot databases users can execute the following:

# Download pipeline databases
nextflow run fmalmeida/bacannot --get_dbs --output bacannot_dbs -profile <docker/singularity>

This will produce a directory like this:

bacannot_dbs
โ”œโ”€โ”€ amrfinder_db
โ”œโ”€โ”€ antismash_db
โ”œโ”€โ”€ argminer_db
โ”œโ”€โ”€ card_db
โ”œโ”€โ”€ iceberg_db
โ”œโ”€โ”€ kofamscan_db
โ”œโ”€โ”€ mlst_db
โ”œโ”€โ”€ phast_db
โ”œโ”€โ”€ phigaro_db
โ”œโ”€โ”€ pipeline_info
โ”œโ”€โ”€ plasmidfinder_db
โ”œโ”€โ”€ platon_db
โ”œโ”€โ”€ prokka_db
โ”œโ”€โ”€ resfinder_db
โ”œโ”€โ”€ vfdb_db
โ””โ”€โ”€ victors_db

To update databases you can either download a new one to a new directory. Remove the database you want to get a new one from the root bacannot dir and use the same command above to save in the same directory (the pipeline will only try to download missing databases). Or, you can use the parameter --force_update to download everything again.

Quickstart

Please refer to the quickstart page ยป

Overview of outputs

A nice overview of the output directory structure and the main tools/features produced by the pipeline is provided at https://bacannot.readthedocs.io/en/latest/outputs.

Documentation

Usage

Users are advised to read the complete documentation ยป

  • Complete command line explanation of parameters:
    • nextflow run fmalmeida/bacannot --help

Command line usage examples

Command line executions are exemplified in the manual.

Using the configuration file

All the parameters showed above can be, and are advised to be, set through the configuration file. When a configuration file is set the pipeline is run by simply executing nextflow run fmalmeida/bacannot -c ./configuration-file

Your configuration file is what will tell to the pipeline the type of data you have, and which processes to execute. Therefore, it needs to be correctly set up.

Create a configuration file in your working directory:

  nextflow run fmalmeida/bacannot --get_config

Interactive graphical configuration and execution

Via NF tower launchpad (good for cloud env execution)

Nextflow has an awesome feature called NF tower. It allows that users quickly customise and set-up the execution and configuration of cloud enviroments to execute any nextflow pipeline from nf-core, github (this one included), bitbucket, etc. By having a compliant JSON schema for pipeline configuration it means that the configuration of parameters in NF tower will be easier because the system will render an input form.

Checkout more about this feature at: https://seqera.io/blog/orgs-and-launchpad/

Via nf-core launch (good for local execution)

Users can trigger a graphical and interactive pipeline configuration and execution by using nf-core launch utility. nf-core launch will start an interactive form in your web browser or command line so you can configure the pipeline step by step and start the execution of the pipeline in the end.

# Install nf-core
pip install nf-core

# Launch the pipeline
nf-core launch fmalmeida/bacannot

It will result in the following:

Known issues

  1. Sometimes when navigating through the shiny parser the reports and JBrowse tabs may still be pointing to old, or just different, samples that have been analysed before and not the actual sample in question. For example, you open the shiny server for the Sample 2, but the reports and JBrowse are showing results of Sample 1. This is caused by the browser's data storages and cookies.
    • To solve this problem user's can just clear the cookies and data cache from the browser.
  2. The JBrowse wrapper in the shiny server is not capable of displaying the GC content and methylation plots when available. It can only display the simpler tracks. If the user wants to visualise and interrogate the GC or methylation tracks it must open the JBrowse outside from the shiny server. For that, two options are available:
    • You can navigate to the jbrowse directory under your sample's output folder and simply execute http-server. This command can be found at: https://www.npmjs.com/package/http-server
    • Or, you can download the JBrowse Desktop app and, from inside the app, select the folder jbrowse/data that is available in your sample's output directory.
  3. If you face some weird error using v3.1 or v3.2, please, before opening a ticket, try updating your docker images, we had some inconsistencies lately and this may be the source of the issue.
  4. If facing an issue with the BACANNOT:SUMMARY module, identical or similar to the one reported in issue [#96], please, before opening a ticket, try updating the python env docker image: docker pull fmalmeida/bacannot:v3.2_pyenv. The image has been recently updated to have the latest version of my python scripts, and that may solve the issue. If not, please open another.
  5. Sometimes, the BACANNOT:UNICYCLER may fail with different, random issues, that does not seem correct, or seem really very random. For example, saying that a read is not available, even though it is there. After some tracing, we realised that the unicycler 0.4.8 installation from conda, and the biocontainer form quay.io is causing this random problem. To solve this issue, please run with a newer version of the tool. This solves the issue in most cases: --unicycler_version 0.5.0--py310h6cc9453_3.
    • Because v3.2 is already tagged and frozen with Zenodo, we will not update it, thus, for this version, using the parameter to overwrite the tool version should be used.
    • In v3.3, unicycler version will be defaulted to 0.5.0--py310h6cc9453_3

Citation

To cite this tool please refer to our Zenodo tag.

This pipeline uses code and infrastructure developed and maintained by the nf-core community, reused here under the GPLv3.

The nf-core framework for community-curated bioinformatics pipelines.

Philip Ewels, Alexander Peltzer, Sven Fillinger, Harshil Patel, Johannes Alneberg, Andreas Wilm, Maxime Ulysse Garcia, Paolo Di Tommaso & Sven Nahnsen.

Nat Biotechnol. 2020 Feb 13. doi: 10.1038/s41587-020-0439-x.

In addition, users are encouraged to cite the programs used in this pipeline whenever they are used. Links to resources of tools and data used in this pipeline are in the list of tools.