/Raw-to-Rawr-server

Primary LanguageRCreative Commons Zero v1.0 UniversalCC0-1.0

Raw-to-Rawr-server

This pipeline can be used for processing of pair-end Illumina raw amplicon data on Linux servers. Can be used for 16S and ITS data and potentially other amplicons. A version for computing clusters using SLURM (ALICE, PICASSO etc.) will be available soon.

The pipeline can be downloaded to the server using command:

git clone "https://github.com/AdamOss88/Raw-to-Rawr-server.git"

*you have to have git installed

The user has to install and activate the conda envronment specified in AmpliconENV.yml file.

conda env create -f AmpliconENV.yml
conda activate AmpliconENV

It is important to mainatain the dedicated folder structure and file naming convention for the raw data. All the raw data has to be in "/raw_data" folder, named: XXX_raw_1.fq.gz AND XXX_raw_2.fq.gz where "XXX" is an identifier the same in both paired end reads and unique between samples.

The pipeline also needs the primer sequences to trim them from the reads and they have to be in the file names primers.fasta in fasta format. An example file is provided but remember to change it according to what primers were used otherwise your results will be unrelaiable (but the pipeline will go through).

For taxonomic classification you need to provide a path to dedicated database for formated for dada2. More info here: https://benjjneb.github.io/dada2/training.html At this moment you have to provide the path manually. By default it is set for IBL blis server. Make sure to use the database dedicated for your type of data, for example SILVA for 16S and UNITE for ITS.

Running the pipeline

  1. Clone the pipeline
  2. Copy or link the data to the /raw_data folder
  3. Make sure the right primers are in the primers.fasta file
  4. Activate the conda environment
  5. run module 1
#for 16S
./module1-16S.sh
#for ITS
./module1-ITS.sh
  1. check the quality of the data and if youre satisfied continue
  2. make sure the path/s to taxonomy reference databases are correct
  3. run module 2
#for 16S
./module2-16S.sh
#for ITS
./module2-ITS.sh
  1. say "rawr!" (only if the results were generated)
  2. Download the results

output

The output of the pipeline is saved in the folder "/results" and includes:

  • otu_table.csv
  • tax_table.csv - taxonomy table
  • refseq.fasta - ASV sequences in fasta file all the .csv tables are also saved as .RDS equivalents to be directly loaded to R

The pipeline also generates a series of reports, all in "/reports" folder:

  • cutadapt_report.txt - report from trimming primers
  • dada2_error_plots.pdf - error model from dada2
  • folder "/quality" with .pdf quality reports from all the reads from fastp

Other output:

  • /processing/moduleX.out - command line output to remember what was done and track potential errors
  • /processing/Renvironment.RData - save R environment to get to intermediate states of analysis of you need