The DTR phage pipeline is a Snakemake pipeline for obtaining polished, full-length dsDNA bacteriophage genomes with direct terminal repeats (DTRs) from enviromental samples.
The pipeline relies heavily on conda
to supply Snakemake and the software packages required for various steps in the analysis. If conda
is not currently installed on your system (highly encouraged!), please first follow the installation instructions for installing miniconda.
Snakemake is a workflow management system that uses Python-based language that is ideally-suited for providing reproducible and scalable bioinformatic pipelines. If you are unfamiliar with Snakemake, this tutorial is a good place to start.
First clone the repositorty to the desired location and enter the cloned directory:
git clone https://github.com/nanoporetech/DTR-phage-pipeline.git
cd DTR-phage-pipeline
Next, create a conda
environment where you will run Snakemake
:
conda env create -f environment.yml
That's all. Ideally, conda
shoud take care of all the remaining dependencies when each specific Snakemake step (described below) is executed.
For speed tips related to conda, see mamba section below.
To verify that everything is set up correctly, simply run the workflow without modifying config.yaml:
snakemake --use-conda -p -r -j 10
The pipeline determines the input files, output path, and the various software parameters from the file config.yml
, which contains key:value pairs that control how the pipeline runs.
-
Attention must be paid to the first section of
config.yml
('Editing mandatory'), as these might change from run to run.sample
: name of the sample (e.g. 25m)stype
: sample type (e.g. 1D or filtered_02um, etc)version
: run version (for changing run parameters)input_fasta
: FASTA file containing nanopore reads to be analyzedinput_summary
: Sequencing summary file associated with the FASTA fileoutput_root
: Base directory where pipeline results will be writtenmax_threads
: Maximum number of threads to allocate for each ruleMEDAKA:model
: Medaka model to use for polishing the draft genome (e.g.r941_min_high_g303
). See Medaka documentation for details on selecting the proper model for you basecalled reads.MEDAKA:threads
: Number of threads to use for each Medaka genome polishing. Should be <<max_threads
so that multiple genomes can be polished in parallel.KAIJU:db
: Path to Kaiju database to use for taxonomic annotation. Folder should contain the*.fmi
,names.dmp
, andnodes.dmp
files. These be downloaded from the Kaiju web server. nr+euk is recommended.KAIJU:switch
: Parameters to use for Kaiju annotation. These parameters should suffice in most cases.SKIP_BINS:<sample>:<stype>:<version>
: List of k-mer bins to skip for analysis. HDBSCAN assigns all unbinned reads to bin_id = -1, which should always be skipped. Occasionally downstream errors in the pipeline indicate that additional bad k-mer bins should also be skipped.
-
The second section of
config.yml
('Editing optional') contains software parameters that can be adjusted, but should perform sufficiently well in most cases using the default values.- 'BIN_AVA_ALGN:CLUST:max_recursion': If you get an error in this step from reaching Python's max recursion depth, you can try increasing the limit here if you have the system resources available.
The full pipeline, starting from raw reads and ending with nanopore-polished phage genomes, can be executed in one step.
snakemake --use-conda -p -j <nproc> -r
Where <nproc>
is the maximum number of cores for all tasks in the pipeline. The output files for this workflow are placed in a path according to the variables set in the config.yml
file. In this description, <run_output_dir>
will refer to the path consisting of <output_root>/<sample>/<stype>/<version>
as defined in the config.yml
file.
The first step provides summary plots for the input reads, identifies reads containing DTR sequences, creates k-mer count vectors, embeds these vectors into 2-d via UMAP, and calls bins in the embedding via HDBSCAN.
The outputs from this step will fall into three directories:
<run_output_dir>/reads_summary
reads.summary.stats.png
: Read length and qscore distributions for all input reads
<run_output_dir>/dtr_reads
output.dtr.fasta
: FASTA file of all DTR-containing readsoutput.dtr.hist.png
: Read length distribution of all DTR-containing readsoutput.dtr.stats.tsv
: Statistics for each DTR-containing read
<run_output_dir>/kmer_binning
bin_membership.tsv
: bin assignments for each DTR-containing readkmer_comp.tsv
: 5-mer count vectors for each DTR-containing readkmer_comp.umap.tsv
: x- and y-coordinates for each read in the 2-d embeddingkmer_comp.umap.*.png
: Variety of scatter plots of UMAP embedding of k-mer count vectors, annoted by features including GC-content, read length, and bin assigned by HDBSCANkmer_comp.umap.bins.tsv
: Mean x- and y-coordinates for each bin assigned by HDBSCAN (for finding bins in 2-d embedding)
The next (optional) step annotates reads using Kaiju and is not strictly required for producing polished genomes. However, it can be informative for verifying the integrity of the k-mer bins and for other downstream analyses.
Some of the output files for this step include:
<run_output_dir>/kaiju
results.html
: Krona dynamic plot of annotated taxonomic composition of DTR-containing readsresults.taxa.tsv
: Per-read annotation results
<run_output_dir>/kmer_binning
kmer_comp.umap.nr_euk.[0-6].png
: Scatter plots of UMAP embedding of k-mer count vectors, annotated by various levels of taxonomic annotation
The next step simply populates subdirectories with the binned reads as assigned by HDBSCAN.
These subdirectories are located in the kmer_binning
directory:
<run_output_dir>/kmer_binning/bins/<bin_id>
- Each bin subdirectory contains a list of binned read names (
read_list.txt
) and an associated FASTA file of reads (<bin_id>.reads.fa
)
Next, each k-mer bin is refined by all-vs-all aligning all reads within a bin. The resulting alignment scores are clustered hierarchically and refined alignment clusters are called from the clustering.
The bin refinement results for each k-mer bin are also placed in kmer_binning
directory:
<run_output_dir>/kmer_binning/refine_bins/align_clusters/<bin_id>
- Each bin refinement procedure generates an alignment (
<bin_id>.ava.paf
), clustering heatmap (<bin_id>.clust.heatmap.png
), and information on alignment cluster assignments (<bin_id>.clust.info.tsv
)
Next, a single read is selected from each valid alignment cluster and is polished by the remaining reads in the alignment cluster. Polishing is first done using multiple rounds of Racon (3x by default), then is finished using a single round of Medaka polishing.
This step produces polished output in the following directories:
<run_output_dir>/kmer_binning/refine_bins/align_cluster_reads/<clust_id>
simply contains the reads corresponding to each alignment cluster<run_output_dir>/kmer_binning/refine_bins/align_cluster_polishing/racon/<clust_id>
contains the Racon polishing output for each alignment cluster<run_output_dir>/kmer_binning/refine_bins/align_cluster_polishing/medaka/<clust_id>
contains the Medaka polishing output for each alignment cluster- The critical output file in each of these
medaka/<clust_id>
folders is the<clust_id>.ref_read.medaka.fasta
file containing the Medaka-polished genome produced from this alignment cluster. Subsequent Snakemake rules analyze, aggregate, and deduplicate these polished genomes. <clust_id>.ref_read.medaka.prodigal.cds.*
files describe the coding sequence annotations from Prodigal for this Medaka-polished genome.<clust_id>.ref_read.strands.*
files describe the strand abundance for reads in each alignment cluster. Clusters containing >80% reads from a single strand should be treated with suspicion.<clust_id>.ref_read.dtr.aligns.*
files describe the results of aligning the DTR sequence from each corresponding k-mer bin to the polished genome from each alignment cluster. If the DTR sequences all align to the same reference positions, the DTR is fixed. However, if they align all over the reference genome, this suggests that a headful DNA packaging mechanism was used.
- The critical output file in each of these
Next, we finish up the genome discovery portion of the pipeline by running a series of aggregations and evaluations of the final polished genome sequences.
All output from this step is written to a single directory:
<run_output_dir>/kmer_binning/refine_bins/align_cluster_polishing
polished.seqs.fasta
: The combined set of genomes from each alignment clusterpolished.seqs.unique.fasta
: Same as above but only containing unique sequences after the deduplication steppolished.stats.tsv
: Various statistics for each polished genome, including length, GC content, DTR details, cluster strand abundance, CDS annotation statistics, circular permutation status, and many otherspolished.stats.unique.tsv
: Same as above but only containing unique sequences after the deduplication steppolished.unique.cds.summary.all.png
: Summary plots of summary statistics for the coding sequences (CDS) annotated by Prodigal for each unique polished genomepolished.unique.cds.summary.dtr_npol10.png
: Same as above but only including polished genomes with a confirmed DTR and at least 10 reads used for polishing
Finally, we run one final step to query the sequencing dataset for linear concatemer reads that could represent interesting mobile elements in the environmental sample.
All output from this step is written to a single directory:
<run_output_dir>/concatemers
concats.fasta
: All identified concatemeric reads found in the input readsconcats.tsv
: Statistic for each concatemeric read found, including readname, length, repeat size, and repeat copy countconcats.contours.png
: Scatter plot with density contours showing the relationship between the observed repeat length and copy count in all identified concatemeric reads.
(c) 2019 Oxford Nanopore Technologies Ltd.
This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/.
The conda package manager is a powerful tool for replicating bioinformatics, but it can be slow for large envirnonments. If you find installing any of the environments is taking too long, try using mamba.
Mamba is a fast drop-in replacement for conda. It is still in development, but so far has proven to e useful. Simply install in your base envirnoment:
conda install mamba
and then replace conda
with mamba
in commands that install packages. EG:
mamba env create -f environment.yml
To get snakemake to use mamba, pass --conda-frontend mamba
to your command:
snakemake --use-conda --conda-frontend mamba -p -r -j <nprocs>
If you re-run the workflow from certain points, because there are multiple "checkpoints" where the number of output files is unknown until runtime, snakemake will sometimes get confused about a rule's input. The exact error will depend upon where in the workflow it occurs.
One know case gives this error:
[Errno 2] No such file or directory: 't'
but there may be others. They all share the odditiy that the rule's input is marked as TBD in the snakemake output and log. For example:
rule aggregate_prodigal_statistics:
input: <TBD>
output:
test/test-output/25m/1D/v2/kmer_binning/refine_bins/align_cluster_polishing/polished.cds.summary.tsv
jobid: 21
The workaround is to ask snakemake explicitly to create the output file of the failed rule. This is done by jsut appending the full (with path) file name to the snakemake command. Simply copy the file path from the output reported after "TBD". For the example error above, you would run something like this:
snakemake --use-conda --conda-frontend mamba -p -r -j 10
test/test-output/25m/1D/v2/kmer_binning/refine_bins/align_cluster_polishing/polished.cds.summary.tsv
The single step should run OK, then you can resume the rest of the workflow with the all
rule.
This pipeline is described in Genome Research:
Beaulaurier J., Luo E., Eppley J., Den Uyl P., Dai X., Turner D.J., Pendelton M., Juul S., Harrington E., DeLong E.F. Assembly-free single-molecule sequencing recovers complete virus genomes from natural microbial communities. Genome Research (2020). doi:10.1101/gr.251686.119