Skip to content

Commit

Permalink
Merge pull request #74 from ENCODE-DCC/dev_v1.2.1
Browse files Browse the repository at this point in the history
Dev v1.2.1
  • Loading branch information
leepc12 authored Jun 15, 2019
2 parents 9ea5e5b + da055f9 commit 844bec3
Show file tree
Hide file tree
Showing 38 changed files with 2,751 additions and 1,263 deletions.
49 changes: 10 additions & 39 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -37,47 +37,24 @@ You can also run our pipeline on DNAnexus without using Caper or Cromwell. There

## Conda

We no longer recommend Conda for resolving dependencies and plan to phase out Conda support. Instead we recommend using Docker or Singularity. You can install Singularity and use it for our pipeline with Caper (by adding `--use-singularity` to command line arguments).

1) Install [Conda](https://docs.conda.io/en/latest/miniconda.html).

2) Install Conda environment for pipeline.

```bash
$ conda/install_dependencies.sh
```

3) Initialize Conda and re-login.

```bash
$ conda init bash
$ exit
```

4) Configure pipeline's python2 and python3 environments.

```bash
$ conda/config_conda_env.sh
$ conda/config_conda_env_py3.sh
```

5) Update pipeline's Conda environment with pipeline's python source code. You need to run this step everytime you update (`git pull`) this pipeline.

```bash
$ conda/update_conda_env.sh
```
We no longer recommend Conda for resolving dependencies and plan to phase out Conda support. Instead we recommend using Docker or Singularity. You can install Singularity and use it for our pipeline with Caper (by adding `--use-singularity` to command line arguments). Please see [this instruction](docs/install_conda.md).

## Tutorial

Make sure that you have configured Caper correctly.
> **WARNING**: DO NOT RUN THIS ON HPC LOGIN NODES. YOUR JOBS WILL BE KILLED.
> **WARNING**: Do not run Caper on HPC login nodes. Your jobs can be killed.
Run it. Due to `--deepcopy` all files in `examples/caper/ENCSR936XTK_subsampled_chr19_only.json` will be recursively copied into Caper's temporary folder (`--tmp-dir`).
Run it. Due to `--deepcopy` all files (HTTP URLs) in `examples/caper/ENCSR936XTK_subsampled_chr19_only.json` will be recursively copied into Caper's temporary folder (`--tmp-dir`).
```bash
$ caper run chip.wdl -i examples/caper/ENCSR936XTK_subsampled_chr19_only.json --deepcopy --use-singularity
```

If you use Conda or Docker (on cloud platforms) then remove `--use-singularity` from the command line and activate it before running a pipeline.
If you use Docker then replace `--use-singularity` with `--use-docker`.
```bash
$ caper run chip.wdl -i examples/caper/ENCSR936XTK_subsampled_chr19_only.json --deepcopy --use-docker
```

If you use Conda then remove `--use-singularity` from the command line and activate pipeline's Conda env before running a pipeline.
```bash
$ conda activate encode-chip-seq-pipeline
$ caper run chip.wdl -i examples/caper/ENCSR936XTK_subsampled_chr19_only.json --deepcopy
Expand All @@ -87,7 +64,7 @@ To run it on an HPC (e.g. Stanford Sherlock and SCG). See details at [Caper's RE

## Input JSON file

Always use absolute paths in an input JSON.
An input JSON file includes all genomic data files, input parameters and metadata for running pipelines. Always use absolute paths in an input JSON.

[Input JSON file specification](docs/input.md)

Expand All @@ -105,12 +82,6 @@ Find a `metadata.json` on Caper's output directory.
$ croo [METADATA_JSON_FILE]
```

## How to build/download genome database

You need to specify a genome data TSV file in your input JSON. Such TSV can be generated/downloaded with actual genome database files.

Use genome database [downloader](genome/download_genome_data.sh) or [builder](docs/build_genome_database.md) for your own genome.

## Useful tools

There are some useful tools to post-process outputs of the pipeline.
Expand Down
Loading

0 comments on commit 844bec3

Please sign in to comment.