-
Notifications
You must be signed in to change notification settings - Fork 1
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Merge pull request #56 from Cristianetaniguti/joint_EmpiricalReads
Joint empirical reads
- Loading branch information
Showing
7 changed files
with
105 additions
and
11 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
32 changes: 32 additions & 0 deletions
32
pipelines/EmpiricalReads2Map/EmpiricalReads2Map.changelog.md
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,32 @@ | ||
# 1.0.0 | ||
|
||
Initial release | ||
|
||
This workflow build linkage maps from genotyping-by-sequencing (GBS) data. The GBS samples are splitted into chunks to be run in different nodes and optimize the analyses. Set the number of samples by chunk in the 'chunk_size' input. Use 'max_cores' to define number of cores to be used in each node. The workflow runs the combinations: | ||
|
||
* SNP calling: GATK and Freebayes | ||
* Dosage/genotype calling: updog, polyRAD and SuperMASSA | ||
* Linkage map build software: OneMap 3.0 and GUSMap | ||
* Using genotype probabilities from GATK, Freebayes, updog, polyRAD and SuperMASSA, and a global error rate of 5% and 0.001% in the OneMap HMM. | ||
|
||
Resulting in 34 linkage maps. | ||
|
||
The workflow include de options to: | ||
|
||
* Remove or not the read duplicates | ||
* Perform the Hard Filtering in GATK results | ||
* Replace the VCF AD format field by counts from BAM files | ||
* Run MCHap software to build haplotypes based on GATK called markers | ||
* Include or not multiallelic (MNP) markers | ||
* Apply filters using VCFtools | ||
|
||
This workflow requires: | ||
|
||
* Diploid bi-parental F1 population | ||
* Single-end reads | ||
* A reference genome | ||
* Genomic positions for markers order | ||
* Selection of a single chromosome from a reference genome to build the linkage maps | ||
|
||
|
||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,61 @@ | ||
version 1.0 | ||
|
||
import "../../structs/empirical_maps_structs.wdl" | ||
import "../../structs/dna_seq_structs.wdl" | ||
|
||
import "../../pipelines/EmpiricalSNPCalling/EmpiricalSNPCalling.wdl" as snpcalling | ||
import "../../pipelines/EmpiricalMaps/EmpiricalMaps.wdl" as maps | ||
|
||
workflow EmpiricalReads { | ||
|
||
input { | ||
File samples_info | ||
ReferenceFasta references | ||
Dataset dataset | ||
Int max_cores | ||
Int chunk_size | ||
Boolean rm_dupli = true | ||
Boolean gatk_mchap = false | ||
Boolean hardfilters = true | ||
Boolean replaceAD = true | ||
Boolean run_gatk = true | ||
Boolean run_freebayes = true | ||
Int ploidy = 2 | ||
Int n_chrom | ||
String? filters | ||
} | ||
|
||
call snpcalling.SNPCalling { | ||
input: | ||
samples_info = samples_info, | ||
references = references, | ||
max_cores = max_cores, | ||
rm_dupli = rm_dupli, | ||
P1 = dataset.parent1, | ||
P2 = dataset.parent2, | ||
gatk_mchap = gatk_mchap, | ||
hardfilters = hardfilters, | ||
replaceAD = replaceAD, | ||
run_gatk = run_gatk, | ||
run_freebayes = run_freebayes, | ||
ploidy = ploidy, | ||
n_chrom = n_chrom | ||
} | ||
|
||
call maps.Maps { | ||
input: | ||
dataset = dataset, | ||
gatk_vcf_multi = SNPCalling.gatk_multi_vcf, | ||
gatk_mchap = gatk_mchap, | ||
gatk_vcf = SNPCalling.gatk_vcf, | ||
freebayes_vcf = SNPCalling.freebayes_vcf, | ||
gatk_vcf_bam_counts = SNPCalling.gatk_vcf_bam_count, | ||
freebayes_vcf_bam_counts = SNPCalling.freebayes_vcf_bam_count, | ||
filters = filters, | ||
max_cores = max_cores | ||
} | ||
|
||
output { | ||
File EmpiricalReads_results = Maps.EmpiricalReads_results | ||
} | ||
} |
File renamed without changes.
File renamed without changes.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters