Gene Myers produced a nice set of tools for scrubbing long reads: trimming them, removing chimeras and patching up low quality regions. While raw long reads can contain a fair bit of junk, scrubbed reads should all be contiguous pieces of the underlying sequence, and this makes assembly much easier. Read all about it on his blog: Scrubbing Reads for Better Assembly
I had two issues with DASCRUBBER. First, it only works for PacBio reads, because PacBio FASTA headers are needed to build a Dazzler database. Second, it is not simple to run, involving more than 10 separate tools and commands. I wrote this wrapper script to solve these issues: it carries out the entire DASCRUBBER process with a single, easy to run command, and it works on any set of long reads (including Oxford Nanopore reads) by faking PacBio read names.
Disclaimer: this wrapper was designed for small genomes where the read sets aren't too large. If you have a large genome or read set, you may run into system resource problems. Read more here.
You'll need a number of Dazzler tools available in your PATH: DAZZ_DB, DALIGNER, DAMASKER and DASCRUBBER
This bash loop should clone and build them. Just replace ~/.local/bin
with wherever you want to put the executables:
for repo in DAZZ_DB DALIGNER DAMASKER DASCRUBBER; do
git clone https://github.com/thegenemyers/"$repo"
cd "$repo" && make -j && cd ..
find "$repo" -maxdepth 1 -type f -executable -exec cp {} ~/.local/bin \;
done
This tool is a single Python 3 script with no third-party dependencies. It will run without any installation:
git clone https://github.com/rrwick/DASCRUBBER-wrapper
DASCRUBBER-wrapper/dascrubber_wrapper.py --help
If you want, you can copy the it to somewhere in your PATH for easy usage:
cp DASCRUBBER-wrapper/dascrubber_wrapper.py ~/.local/bin
dascrubber_wrapper.py --help
This script has two required parameters: input reads and genome size. It outputs scrubbed reads to stdout, so its output should be directed to a file.
Default parameters for a 5.5 Mbp bacterial genome:
dascrubber_wrapper.py -i reads.fastq.gz -g 5.5M | gzip > scrubbed.fasta.gz
Limit daligner memory usage to 50 GB:
dascrubber_wrapper.py -i reads.fastq.gz -g 5.5M --daligner_options="-M50" | gzip > scrubbed.fasta.gz
Keep Dazzler files after completion:
dascrubber_wrapper.py -i reads.fastq.gz -g 5.5M -d working_files --keep | gzip > scrubbed.fasta.gz
- Convert reads to a FASTA file with PacBio-style headers.
- It is assumed that the input reads are either FASTQ or FASTA (one line per sequence). Gzipped reads are okay. Multi-line-per-read FASTA files are not okay (I might get around to adding that later).
- Build a Dazzler database of the reads with fasta2DB.
- Split the database with DBsplit.
- Splitting seems to be necessary or else the DASedit command will fail – I'm not sure why.
- Find all read-read overlaps with daligner.
- This is the slowest and most memory-hungry step of the process.
- Mask repeats with REPmask.
- See the Parameters section for details on the repeat depth threshold.
- Find tandem repeats with datander.
- Mask tandem repeats with TANmask.
- Find all read-read overlaps with daligner again, this time masking repeats.
- Find estimated genome coverage with DAScover.
- Find intrinsic quality values with DASqv.
- Trim reads and split chimeras with DAStrim.
- See the Parameters section for details on the good and bad quality thresholds.
- Patch low-quality regions of reads with DASpatch.
- Produce a new database of scrubbed reads with DASedit.
- Extract a FASTA of scrubbed reads with DB2fasta.
- Restore original read names and output to stdout.
- A range is appended to the end of the new read names. For example, if the original read was named
read1975
, the scrubbed read might be namedread1975/400_9198
. - It is possible for one original read to result in more than one scrubbed read. For example, a chimeric read named
read2392
might result in two scrubbed reads:read2392/0_12600
andread2392/12700_25300
.
- A range is appended to the end of the new read names. For example, if the original read was named
Parameters can be passed to any of the subcommands using the 'options' arguments. E.g. --daligner_options
. Read Gene Myers' command guides for details about the tools.
daligner can use a lot of memory – by default it will use all available memory on the system! I therefore find it useful to limit its memory usage, e.g. --daligner_options="-M50"
.
The REPmask command replies on a threshold depth, above which a sequence is considered to be a repeat (read more here). I found that 3x the base depth works well for my datasets (high depth bacterial genomes). E.g. if the base depth (as determined using the genome size) is 50x then regions with 150x or greater depth are considered repeats. You can adjust this ratio from the default of 3 using the --repeat_depth
option – a higher value may be appropriate for lower coverage datasets or highly repetitive genomes. Alternatively, you can manually set the threshold: e.g. --repmask_options="-c40"
.
DAStrim takes two important parameters: -g
and -b
for good and bad quality thresholds. Its default behaviour is to use the 80th and 93rd percentiles of the QV scores made by DASqv. If you want to use different values, set them manually: e.g. --dastrim_options="-g20 -b30"
.
usage: DASCRUBBER_wrapper.py -i INPUT_READS -g GENOME_SIZE [-d TEMPDIR] [-k]
[-r REPEAT_DEPTH]
[--dbsplit_options DBSPLIT_OPTIONS]
[--daligner_options DALIGNER_OPTIONS]
[--repmask_options REPMASK_OPTIONS]
[--datander_options DATANDER_OPTIONS]
[--tanmask_options TANMASK_OPTIONS]
[--dasqv_options DASQV_OPTIONS]
[--dastrim_options DASTRIM_OPTIONS]
[--daspatch_options DASPATCH_OPTIONS]
[--dasedit_options DASEDIT_OPTIONS] [-h]
A wrapper tool for the DASCRUBBER pipeline for scrubbing (trimming and chimera
removal) of long read sets (PacBio or ONT reads)
Required arguments:
-i INPUT_READS, --input_reads INPUT_READS
input set of long reads to be scrubbed
-g GENOME_SIZE, --genome_size GENOME_SIZE
approximate genome size (examples: 3G, 5.5M or 800k),
used to determine depth of coverage
Optional arguments:
-d TEMPDIR, --tempdir TEMPDIR
path of directory for temporary files (default: use a
directory in the current location named
dascrubber_temp_PID where PID is the process ID)
-k, --keep keep the temporary directory (default: delete the
temporary directory after scrubbing is complete)
-r REPEAT_DEPTH, --repeat_depth REPEAT_DEPTH
REPmask will be given a repeat threshold of this
depth, relative to the overall depth (e.g. if 3, then
regions with 3x the base depth are considered
repeats) (default: 3)
Command options:
You can specify additional options for each of the Dazzler commands if you
do not want to use the defaults (example: --daligner_options="-M80")
--dbsplit_options DBSPLIT_OPTIONS
--daligner_options DALIGNER_OPTIONS
--repmask_options REPMASK_OPTIONS
--datander_options DATANDER_OPTIONS
--tanmask_options TANMASK_OPTIONS
--dascover_options DASCOVER_OPTIONS
--dasqv_options DASQV_OPTIONS
--dastrim_options DASTRIM_OPTIONS
--daspatch_options DASPATCH_OPTIONS
--dasedit_options DASEDIT_OPTIONS
Help:
-h, --help show this help message and exit
This wrapper isn't really suited to large datasets, because it does a simple all-vs-all read alignment with DALIGNER. For a very large read set, this will be too slow or use too much memory, and you'll instead want to use some of DALIGNER's HPC functionality. Read more here.
The only benefit this wrapper might still have is its first step: faking PacBio read names, enabling the Dazzler tools to work for Nanopore reads. If you are in this position (wanting to scrub a large Nanopore dataset), then here's what I recommend:
- Start this wrapper:
dascrubber_wrapper.py -i reads.fastq.gz -g 100M --tempdir renamed_reads
(it doesn't matter what you use for the genome size option) - When the wrapper finishes the 'Processing and renaming reads' step, kill it with Ctrl-C.
- You should now have a directory named
renamed_reads
with a file inside namedrenamed_reads.fasta
. You can delete anything else in that directory like thereads.db
file or thealign_temp
directory, if they exist. - Take the renamed reads and run the database/aligning/scrubbing commands yourself. Plenty of information here: dazzlerblog.wordpress.com
Good luck!