We have a new bioinformatic resource that largely replaces the functionality of this project! See our new repository here: https://github.com/epi2me-labs/wf-pore-c.
This repository is now unsupported and we do not recommend its use. Please contact Oxford Nanopore: support@nanoporetech.com for help with your application if it is not possible to upgrade to our new resources, or we are missing key features.
This pipeline manages a pore-c workflow starting from raw fastq files and converting them to standard file formats for use by downstream tools. The steps involved are:
- Pre-processing a reference genome or draft assembly to generate auxiliary files used in downstream analyses
- Creating virtual digests of the genome
- Filtering the raw reads to remove any that might break downstream tools
- Align against a reference genome
- Processing results to filter spurious alignments, detect ligation junctions and assign fragments. The results are stored in a parquet table for downstream processing.
- Converting the results to the following formats:
In most cases, it is best to pre-install conda before starting. All other dependencies will be installed automatically when running the pipeline for the first time.
This pipeline requires a computer running Linux (Ubuntu 16). >64Gb of memory would be recommended. The pipeline has been tested on minimal server installs of these operating systems.
Most software dependencies are managed using conda. To install conda, please install miniconda3 and refer to installation instructions. You will need to accept the license agreement during installation and we recommend that you allow the Conda installer to prepend its path to your .bashrc file when asked.
wget https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh
bash Miniconda3-latest-Linux-x86_64.sh
Check if the conda has successfully installed
conda -h
If conda has installed correctly, you should see the follow output. If you do not see the below output, you may need to close and reopen your terminal.
$ conda
usage: conda [-h] [-V] command ...
conda is a tool for managing and deploying applications, environments and packages.
Options:
positional arguments:
command
clean Remove unused packages and caches.
config Modify configuration values in .condarc. This is modeled
after the git config command. Writes to the user .condarc
file ($HOME/.condarc) by default.
create Create a new conda environment from a list of specified
packages.
..............
Clone this git repository to the location where you want to run your analysis and create the conda environment that will be used to run the pipeline
git clone https://github.com/nanoporetech/Pore-C-Snakemake.git
cd pore-c-snakemake
## Creates environment and the dependencies will install automatically
conda env create
conda activate pore_c_snakemake
Note before you run any of the snakemake commands below you need to make sure that you've run conda activate pore_c_snakemake
.
Test data is included in the .test
subfolder (git-lfs
is required to download them). To run the tests use
snakemake --use-conda test -j 4 --config=output_dir=results.test
The results of the test run will appear in the results.test
directory.
The pipeline configuration is split across several files:
* `config/config.yaml` - A yaml file containing settings for the pipeline. Input data is specified in the following tab-delimited files.
* `config/basecall.tsv` - Metadata and locations of the pore-c sequencing run fastqs.
* `config/references.tsv` - Locations of the draft/scaffold/reference assemblies that the pore-c reads will be mapped to.
* `config/phased_vcfs.tsv` - [Optional] The location of phased vcf files that can be used to haplotag poreC reads.
Test your configuration by performing a dry-run via
snakemake --use-conda -n
Execute the workflow locally via
snakemake --use-conda --cores $N
using $N
cores or run it in a cluster environment via
snakemake --use-conda --cluster qsub --jobs 100
or
snakemake --use-conda --drmaa --jobs 100
in combination with any of the modes above. See the Snakemake documentation for further details.
The pipeline defines several targets that can be speficied on the command line:
- all: The default target which builds the
pore_c
contact and concatemer parquet files under themerged_contacts
directory. - cooler: Builds a multi-resolution
.mcool
file. - pairs: Builds a
pairix
-indexed pairs file. - juicer: Builds a
.hic
file compatible with thejuicebox
suite of tools. - salsa: Builds a
.bed
file for use with thesalsa2
scaffolding tool. - mnd: Builds a
.mnd.txt
file compatible with the3d-dna
scaffolding tool [experimental].
To build the files for a particular target:
snakemake --use-conda -j 8 <target>
Once the pipeline has run successfully you should expect the following files in the output directory:
refgenome/
:{refgenome_id}.rg.metadata.csv
- chromosome metadata in csv format.{refgenome_id}.rg.chromsizes
- reference genome chromosome lengths{refgenome_id}.rg.fa.gz
- reference genome compressed with bgzip{refgenome_id}.rg.fa.gz.fai
- samtools indexed reference genome{refgenome_id..rg.fa.gz.bwt
- bwa index reference genome
virtual_digest/
:{enzyme}_{refgenome_id}.vd.fragments.parquet
- A table containing the intervals generated by the virtual digest.{enzyme}_{refgenome_id}.vd.digest_stats.csv
- virtual digest aggregate statistics
basecall/
:{enzyme}_{run_id}.rd.{batch_id}.fq.gz
- basecalls that have passed filtering split into batches of 50,000 (can be changed in config).{enzyme}_{run_id}.rd.catalog.yaml
- an intake catalog containing read metadata.{enzyme}_{run_id}.rd.read_metadata.parquet
- a table of per-read statistics.{enzyme}_{run_id}.rd.summary.csv
- a table of aggregate statistics for the reads.
mapping/
:{enzyme}_{run_id}_{batch_id}_{refgenome_id}.coord_sort.bam
- bam alignment file sorted by genome coordinate with an alignment index added to the query name.{enzyme}_{run_id}_{batch_id}_{refgenome_id}_{phase_set_id}.coord_sort.bam
- whatshap-produced text file mapping alignments to phase sets (dummy file is produced if unphased).
align_table/
:{enzyme}_{run_id}_{batch_id}_{refgenome_id}_{phase_set_id}.at.alignment.parquet
- a parquet file with alignment information extracted from the corresponding bam file{enzyme}_{run_id}_{batch_id}_{refgenome_id}_{phase_set_id}.at.pore_c.parquet
- a parquet file with the same information as the alignment parquet with additional data on fragment assignments and the pass-fail status of each alignment.
contacts/
:{enzyme}_{run_id}_{batch_id}_{refgenome_id}_{phase_set_id}.contacts.parquet
- a table derived from thepore_c.parquet
file consisting of all pairwise contacts (equivalent to a.pairs
file).
merged_contacts/
:{enzyme}_{run_id}_{refgenome_id}_{phase_set_id}.contacts.parquet
- a merged version of the contacts file for a run{enzyme}_{run_id}_{refgenome_id}_{phase_set_id}.concatemers.parquet
- a table with per-read (aka concatemer) statistics{enzyme}_{run_id}_{refgenome_id}_{phase_set_id}.concateme_summary.csv
- a table with per-run statistics
matrix/
{enzyme}_{run_id}_{refgenome_id}_{phase_set_id}.matrix.catalog.yaml
- an intake catalog containing metadata about the aggregate matrix.{enzyme}_{run_id}_{refgenome_id}_{phase_set_id}.matrix.coo.csv.gz
- aggregate read counts in the format 'bin1_id,bin2_id,count' - suitable for use withcooler load
the bin width for this set by the*base*
matrix resolution in the config file.{enzyme}_{run_id}_{refgenome_id}_{phase_set_id}.matrix.cool
- the aggregate contact counts in cooler format{enzyme}_{run_id}_{refgenome_id}_{phase_set_id}.matrix.counts.mcool
- a multi-resolution cool file.
pairs/
:{enzyme}_{run_id}_{refgenome_id}_{phase_set_id}.pairs.pairs.gz
- contains fragment position and fragment pairs in pairs format.
assembly/
:{enzyme}_{run_id}_{refgenome_id}_{phase_set_id}.salsa2.bed
- optional a bed file compatible with the salsa2 scaffolding tool.
juicebox/
:{enzyme}_{run_id}_{refgenome_id}_{phase_set_id}.hicRef
- optional a restriction site format file.{enzyme}_{run_id}_{refgenome_id}_{phase_set_id}.hic
- optional a hic medium format file of pairwise contacts.{enzyme}_{run_id}_{refgenome_id}_{phase_set_id}.mnd.txt
- optional a merged_no_dups format file (experimental).
© 2019 Oxford Nanopore Technologies Ltd.
Bioinformatics-Tutorials is distributed by Oxford Nanopore Technologies under the terms of the MPL-2.0 license.