Merge remote-tracking branch 'nf-core/master'

This commit is contained in:
Tamara Hodgetts 2021-11-04 11:42:55 +00:00
commit e687c4b97f
67 changed files with 2404 additions and 126 deletions

View file

@ -19,8 +19,9 @@ process ALLELECOUNTER {
} }
input: input:
tuple val(meta), path(bam), path(bai) tuple val(meta), path(input), path(input_index)
path loci path loci
path fasta
output: output:
tuple val(meta), path("*.alleleCount"), emit: allelecount tuple val(meta), path("*.alleleCount"), emit: allelecount
@ -28,11 +29,14 @@ process ALLELECOUNTER {
script: script:
def prefix = options.suffix ? "${meta.id}${options.suffix}" : "${meta.id}" def prefix = options.suffix ? "${meta.id}${options.suffix}" : "${meta.id}"
def reference_options = fasta ? "-r $fasta": ""
""" """
alleleCounter \\ alleleCounter \\
$options.args \\ $options.args \\
-l $loci \\ -l $loci \\
-b $bam \\ -b $input \\
$reference_options \\
-o ${prefix}.alleleCount -o ${prefix}.alleleCount
cat <<-END_VERSIONS > versions.yml cat <<-END_VERSIONS > versions.yml

View file

@ -19,11 +19,11 @@ input:
description: | description: |
Groovy Map containing sample information Groovy Map containing sample information
e.g. [ id:'test', single_end:false ] e.g. [ id:'test', single_end:false ]
- bam: - input:
type: file type: file
description: BAM/CRAM/SAM file description: BAM/CRAM/SAM file
pattern: "*.{bam,cram,sam}" pattern: "*.{bam,cram,sam}"
- bai: - input_index:
type: file type: file
description: BAM/CRAM/SAM index file description: BAM/CRAM/SAM index file
pattern: "*.{bai,crai,sai}" pattern: "*.{bai,crai,sai}"
@ -31,7 +31,9 @@ input:
type: file type: file
description: loci file <CHR><tab><POS1> description: loci file <CHR><tab><POS1>
pattern: "*.{tsv}" pattern: "*.{tsv}"
- fasta:
type: file
description: Input genome fasta file. Required when passing CRAM files.
output: output:
- meta: - meta:
@ -50,3 +52,4 @@ output:
authors: authors:
- "@fullama" - "@fullama"
- "@fbdtemme"

View file

@ -19,13 +19,13 @@ process FREEBAYES {
} }
input: input:
tuple val(meta), path(bam), path(bai) tuple val(meta), path(input_1), path(input_1_index), path(input_2), path(input_2_index)
tuple path(fasta), path(fai) path fasta
path(targets) path fai
path(samples) path targets
path(populations) path samples
path(cnv) path populations
path cnv
output: output:
tuple val(meta), path("*.vcf.gz") , emit: vcf tuple val(meta), path("*.vcf.gz") , emit: vcf
@ -33,10 +33,12 @@ process FREEBAYES {
script: script:
def prefix = options.suffix ? "${meta.id}${options.suffix}" : "${meta.id}" def prefix = options.suffix ? "${meta.id}${options.suffix}" : "${meta.id}"
def input = input_2 ? "${input_1} ${input_2}" : "${input_1}"
def targets_file = targets ? "--target ${targets}" : "" def targets_file = targets ? "--target ${targets}" : ""
def samples_file = samples ? "--samples ${samples}" : "" def samples_file = samples ? "--samples ${samples}" : ""
def populations_file = populations ? "--populations ${populations}" : "" def populations_file = populations ? "--populations ${populations}" : ""
def cnv_file = cnv ? "--cnv-map ${cnv}" : "" def cnv_file = cnv ? "--cnv-map ${cnv}" : ""
if (task.cpus > 1) { if (task.cpus > 1) {
""" """
freebayes-parallel \\ freebayes-parallel \\
@ -47,7 +49,7 @@ process FREEBAYES {
$populations_file \\ $populations_file \\
$cnv_file \\ $cnv_file \\
$options.args \\ $options.args \\
$bam > ${prefix}.vcf $input > ${prefix}.vcf
gzip --no-name ${prefix}.vcf gzip --no-name ${prefix}.vcf
@ -66,7 +68,7 @@ process FREEBAYES {
$populations_file \\ $populations_file \\
$cnv_file \\ $cnv_file \\
$options.args \\ $options.args \\
$bam > ${prefix}.vcf $input > ${prefix}.vcf
gzip --no-name ${prefix}.vcf gzip --no-name ${prefix}.vcf

View file

@ -4,15 +4,18 @@ keywords:
- variant caller - variant caller
- SNP - SNP
- genotyping - genotyping
- variant calling - somatic variant calling
- germline variant calling
- bacterial variant calling
- bayesian - bayesian
tools: tools:
- freebayes: - freebayes:
description: Bayesian haplotype-based polymorphism discovery and genotyping description: Bayesian haplotype-based polymorphism discovery and genotyping
homepage: https://github.com/freebayes/freebayes homepage: https://github.com/freebayes/freebayes
documentation: https://github.com/freebayes/freebayes documentation: https://github.com/freebayes/freebayes
tool_dev_url: https://github.com/freebayes/freebayes tool_dev_url: https://github.com/freebayes/freebayes
doi: "" doi: "arXiv:1207.3907"
licence: ['MIT'] licence: ['MIT']
input: input:
@ -21,11 +24,11 @@ input:
description: | description: |
Groovy Map containing sample information Groovy Map containing sample information
e.g. [ id:'test', single_end:false ] e.g. [ id:'test', single_end:false ]
- bam: - input:
type: file type: file
description: BAM/CRAM/SAM file description: BAM/CRAM/SAM file
pattern: "*.{bam,cram,sam}" pattern: "*.{bam,cram,sam}"
- bai: - input_index:
type: file type: file
description: BAM/CRAM/SAM index file description: BAM/CRAM/SAM index file
pattern: "*.bam.bai" pattern: "*.bam.bai"
@ -59,7 +62,6 @@ input:
seq_name start end sample_name copy_number seq_name start end sample_name copy_number
pattern: "*.bed" pattern: "*.bed"
output: output:
- meta: - meta:
type: map type: map
@ -74,5 +76,8 @@ output:
type: file type: file
description: Compressed VCF file description: Compressed VCF file
pattern: "*.vcf.gz" pattern: "*.vcf.gz"
authors: authors:
- "@maxibor" - "@maxibor"
- "@FriederikeHanssen"
- "@maxulysse"

View file

@ -61,3 +61,4 @@ output:
authors: authors:
- "@yocra3" - "@yocra3"
- "@FriederikeHanssen"

View file

@ -0,0 +1,78 @@
//
// Utility functions used in nf-core DSL2 module files
//
//
// Extract name of software tool from process name using $task.process
//
def getSoftwareName(task_process) {
return task_process.tokenize(':')[-1].tokenize('_')[0].toLowerCase()
}
//
// Extract name of module from process name using $task.process
//
def getProcessName(task_process) {
return task_process.tokenize(':')[-1]
}
//
// Function to initialise default values and to generate a Groovy Map of available options for nf-core modules
//
def initOptions(Map args) {
def Map options = [:]
options.args = args.args ?: ''
options.args2 = args.args2 ?: ''
options.args3 = args.args3 ?: ''
options.publish_by_meta = args.publish_by_meta ?: []
options.publish_dir = args.publish_dir ?: ''
options.publish_files = args.publish_files
options.suffix = args.suffix ?: ''
return options
}
//
// Tidy up and join elements of a list to return a path string
//
def getPathFromList(path_list) {
def paths = path_list.findAll { item -> !item?.trim().isEmpty() } // Remove empty entries
paths = paths.collect { it.trim().replaceAll("^[/]+|[/]+\$", "") } // Trim whitespace and trailing slashes
return paths.join('/')
}
//
// Function to save/publish module results
//
def saveFiles(Map args) {
def ioptions = initOptions(args.options)
def path_list = [ ioptions.publish_dir ?: args.publish_dir ]
// Do not publish versions.yml unless running from pytest workflow
if (args.filename.equals('versions.yml') && !System.getenv("NF_CORE_MODULES_TEST")) {
return null
}
if (ioptions.publish_by_meta) {
def key_list = ioptions.publish_by_meta instanceof List ? ioptions.publish_by_meta : args.publish_by_meta
for (key in key_list) {
if (args.meta && key instanceof String) {
def path = key
if (args.meta.containsKey(key)) {
path = args.meta[key] instanceof Boolean ? "${key}_${args.meta[key]}".toString() : args.meta[key]
}
path = path instanceof String ? path : ''
path_list.add(path)
}
}
}
if (ioptions.publish_files instanceof Map) {
for (ext in ioptions.publish_files) {
if (args.filename.endsWith(ext.key)) {
def ext_list = path_list.collect()
ext_list.add(ext.value)
return "${getPathFromList(ext_list)}/$args.filename"
}
}
} else if (ioptions.publish_files == null) {
return "${getPathFromList(path_list)}/$args.filename"
}
}

View file

@ -0,0 +1,67 @@
// Import generic module functions
include { initOptions; saveFiles; getSoftwareName; getProcessName } from './functions'
params.options = [:]
options = initOptions(params.options)
process GATK4_GENOMICSDBIMPORT {
tag "$meta.id"
label 'process_low'
publishDir "${params.outdir}",
mode: params.publish_dir_mode,
saveAs: { filename -> saveFiles(filename:filename, options:params.options, publish_dir:getSoftwareName(task.process), meta:meta, publish_by_meta:['id']) }
conda (params.enable_conda ? "bioconda::gatk4=4.2.0.0" : null)
if (workflow.containerEngine == 'singularity' && !params.singularity_pull_docker_container) {
container "https://depot.galaxyproject.org/singularity/gatk4:4.2.0.0--0"
} else {
container "quay.io/biocontainers/gatk4:4.2.0.0--0"
}
input:
tuple val(meta), path(vcf), path(tbi), path(intervalfile), val(intervalval), path(wspace)
val run_intlist
val run_updatewspace
val input_map
output:
tuple val(meta), path("*_genomicsdb") , optional:true, emit: genomicsdb
tuple val(meta), path("$updated_db") , optional:true, emit: updatedb
tuple val(meta), path("*.interval_list"), optional:true, emit: intervallist
path "versions.yml" , emit: versions
script:
def prefix = options.suffix ? "${meta.id}${options.suffix}" : "${meta.id}"
// settings for running default create gendb mode
def inputs_command = input_map ? "--sample-name-map ${vcf[0]}" : "${'-V ' + vcf.join(' -V')}"
def dir_command = "--genomicsdb-workspace-path ${prefix}"
def intervals_command = intervalfile ? " -L ${intervalfile} " : " -L ${intervalval} "
// settings changed for running get intervals list mode if run_intlist is true
if (run_intlist) {
inputs_command = ''
dir_command = "--genomicsdb-update-workspace-path ${wspace}"
intervals_command = "--output-interval-list-to-file ${prefix}.interval_list"
}
// settings changed for running update gendb mode. inputs_command same as default, update_db forces module to emit the updated gendb
if (run_updatewspace) {
dir_command = "--genomicsdb-update-workspace-path ${wspace}"
intervals_command = ''
updated_db = wspace.toString()
}
"""
gatk GenomicsDBImport \\
$inputs_command \\
$dir_command \\
$intervals_command \\
$options.args
cat <<-END_VERSIONS > versions.yml
${getProcessName(task.process)}:
${getSoftwareName(task.process)}: \$(echo \$(gatk --version 2>&1) | sed 's/^.*(GATK) v//; s/ .*\$//')
END_VERSIONS
"""
}

View file

@ -0,0 +1,80 @@
name: gatk4_genomicsdbimport
description: merge GVCFs from multiple samples. For use in joint genotyping or somatic panel of normal creation.
keywords:
- gatk4
- genomicsdbimport
- genomicsdb
- panelofnormalscreation
- jointgenotyping
tools:
- gatk4:
description: |
Developed in the Data Sciences Platform at the Broad Institute, the toolkit offers a wide variety of tools
with a primary focus on variant discovery and genotyping. Its powerful processing engine
and high-performance computing features make it capable of taking on projects of any size.
homepage: https://gatk.broadinstitute.org/hc/en-us
documentation: https://gatk.broadinstitute.org/hc/en-us/categories/360002369672s
doi: 10.1158/1538-7445.AM2017-3590
input:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test']
- vcf:
type: list
description: either a list of vcf files to be used to create or update a genomicsdb, or a file that contains a map to vcf files to be used.
pattern: "*.vcf.gz"
- tbi:
type: list
description: list of tbi files that match with the input vcf files
pattern: "*.vcf.gz_tbi"
- wspace:
type: path
description: path to an existing genomicsdb to be used in update db mode or get intervals mode. This WILL NOT specify name of a new genomicsdb in create db mode.
pattern: "/path/to/existing/gendb"
- intervalfile:
type: file
description: file containing the intervals to be used when creating the genomicsdb
pattern: "*.interval_list"
- intervalval:
type: string
description: if an intervals file has not been spcified, the value enetered here will be used as an interval via the "-L" argument
pattern: "example: chr1:1000-10000"
- run_intlist:
type: boolean
description: Specify whether to run get interval list mode, this option cannot be specified at the same time as run_updatewspace.
pattern: "true/false"
- run_updatewspace:
type: boolean
description: Specify whether to run update genomicsdb mode, this option takes priority over run_intlist.
pattern: "true/false"
- input_map:
type: boolean
description: Specify whether the vcf input is providing a list of vcf file(s) or a single file containing a map of paths to vcf files to be used to create or update a genomicsdb.
pattern: "*.sample_map"
output:
- genomicsdb:
type: directory
description: Directory containing the files that compose the genomicsdb workspace, this is only output for create mode, as update changes an existing db
pattern: "*_genomicsdb"
- intervallist:
type: file
description: File containing the intervals used to generate the genomicsdb, only created by get intervals mode.
pattern: "*.interval_list"
- versions:
type: file
description: File containing software versions
pattern: "versions.yml"
authors:
- "@GCJMackenzie"

View file

@ -22,6 +22,10 @@ process GENRICH {
tuple val(meta), path(treatment_bam) tuple val(meta), path(treatment_bam)
path control_bam path control_bam
path blacklist_bed path blacklist_bed
val save_pvalues
val save_pileup
val save_bed
val save_duplicates
output: output:
tuple val(meta), path("*narrowPeak") , emit: peaks tuple val(meta), path("*narrowPeak") , emit: peaks
@ -33,13 +37,13 @@ process GENRICH {
script: script:
def prefix = options.suffix ? "${meta.id}${options.suffix}" : "${meta.id}" def prefix = options.suffix ? "${meta.id}${options.suffix}" : "${meta.id}"
def control = params.control_bam ? "-c $control_bam" : '' def control = control_bam ? "-c $control_bam" : ''
def pvalues = params.pvalues ? "-f ${prefix}.pvalues.bedGraph" : "" def blacklist = blacklist_bed ? "-E $blacklist_bed" : ""
def pileup = params.pileup ? "-k ${prefix}.pileup.bedGraph" : "" def pvalues = save_pvalues ? "-f ${prefix}.pvalues.bedGraph" : ""
def bed = params.bed ? "-b ${prefix}.intervals.bed" : "" def pileup = save_pileup ? "-k ${prefix}.pileup.bedGraph" : ""
def blacklist = params.blacklist_bed ? "-E $blacklist_bed" : "" def bed = save_bed ? "-b ${prefix}.intervals.bed" : ""
def duplicates = "" def duplicates = ""
if (params.save_duplicates) { if (save_duplicates) {
if (options.args.contains('-r')) { if (options.args.contains('-r')) {
duplicates = "-R ${prefix}.duplicates.txt" duplicates = "-R ${prefix}.duplicates.txt"
} else { } else {
@ -58,7 +62,6 @@ process GENRICH {
$pileup \\ $pileup \\
$bed \\ $bed \\
$duplicates \\ $duplicates \\
$blacklist \\
$control $control
cat <<-END_VERSIONS > versions.yml cat <<-END_VERSIONS > versions.yml

View file

@ -15,7 +15,6 @@ tools:
tool_dev_url: https://github.com/jsh58/Genrich tool_dev_url: https://github.com/jsh58/Genrich
doi: "" doi: ""
licence: ['MIT'] licence: ['MIT']
input: input:
- meta: - meta:
type: map type: map
@ -34,7 +33,18 @@ input:
type: file type: file
description: Bed file containing genomic intervals to exclude from the analysis description: Bed file containing genomic intervals to exclude from the analysis
pattern: "*.{bed}" pattern: "*.{bed}"
- save_pvalues:
type: boolean
description: Create bedgraph-ish file for p/q-values file
- save_pileup:
type: boolean
description: Create bedgraph-ish file for pileups and p-values
- save_bed:
type: boolean
description: Create BED file for reads/fragments/intervals
- save_duplicates:
type: boolean
description: Create PCR duplicates file (only works if -r option is set)
output: output:
- meta: - meta:
type: map type: map
@ -65,7 +75,6 @@ output:
type: file type: file
description: File containing software version description: File containing software version
pattern: "*.{version.txt}" pattern: "*.{version.txt}"
authors: authors:
- "@JoseEspinosa" - "@JoseEspinosa"

View file

@ -0,0 +1,78 @@
//
// Utility functions used in nf-core DSL2 module files
//
//
// Extract name of software tool from process name using $task.process
//
def getSoftwareName(task_process) {
return task_process.tokenize(':')[-1].tokenize('_')[0].toLowerCase()
}
//
// Extract name of module from process name using $task.process
//
def getProcessName(task_process) {
return task_process.tokenize(':')[-1]
}
//
// Function to initialise default values and to generate a Groovy Map of available options for nf-core modules
//
def initOptions(Map args) {
def Map options = [:]
options.args = args.args ?: ''
options.args2 = args.args2 ?: ''
options.args3 = args.args3 ?: ''
options.publish_by_meta = args.publish_by_meta ?: []
options.publish_dir = args.publish_dir ?: ''
options.publish_files = args.publish_files
options.suffix = args.suffix ?: ''
return options
}
//
// Tidy up and join elements of a list to return a path string
//
def getPathFromList(path_list) {
def paths = path_list.findAll { item -> !item?.trim().isEmpty() } // Remove empty entries
paths = paths.collect { it.trim().replaceAll("^[/]+|[/]+\$", "") } // Trim whitespace and trailing slashes
return paths.join('/')
}
//
// Function to save/publish module results
//
def saveFiles(Map args) {
def ioptions = initOptions(args.options)
def path_list = [ ioptions.publish_dir ?: args.publish_dir ]
// Do not publish versions.yml unless running from pytest workflow
if (args.filename.equals('versions.yml') && !System.getenv("NF_CORE_MODULES_TEST")) {
return null
}
if (ioptions.publish_by_meta) {
def key_list = ioptions.publish_by_meta instanceof List ? ioptions.publish_by_meta : args.publish_by_meta
for (key in key_list) {
if (args.meta && key instanceof String) {
def path = key
if (args.meta.containsKey(key)) {
path = args.meta[key] instanceof Boolean ? "${key}_${args.meta[key]}".toString() : args.meta[key]
}
path = path instanceof String ? path : ''
path_list.add(path)
}
}
}
if (ioptions.publish_files instanceof Map) {
for (ext in ioptions.publish_files) {
if (args.filename.endsWith(ext.key)) {
def ext_list = path_list.collect()
ext_list.add(ext.value)
return "${getPathFromList(ext_list)}/$args.filename"
}
}
} else if (ioptions.publish_files == null) {
return "${getPathFromList(path_list)}/$args.filename"
}
}

View file

@ -0,0 +1,37 @@
// Import generic module functions
include { initOptions; saveFiles; getSoftwareName; getProcessName } from './functions'
params.options = [:]
options = initOptions(params.options)
process GUNC_DOWNLOADDB {
tag '$db_name'
label 'process_low'
publishDir "${params.outdir}",
mode: params.publish_dir_mode,
saveAs: { filename -> saveFiles(filename:filename, options:params.options, publish_dir:getSoftwareName(task.process), meta:[:], publish_by_meta:[]) }
conda (params.enable_conda ? "bioconda::gunc=1.0.5" : null)
if (workflow.containerEngine == 'singularity' && !params.singularity_pull_docker_container) {
container "https://depot.galaxyproject.org/singularity/gunc:1.0.5--pyhdfd78af_0"
} else {
container "quay.io/biocontainers/gunc:1.0.5--pyhdfd78af_0"
}
input:
val db_name
output:
path "*.dmnd" , emit: db
path "versions.yml" , emit: versions
script:
"""
gunc download_db . -db $db_name $options.args
cat <<-END_VERSIONS > versions.yml
${getProcessName(task.process)}:
${getSoftwareName(task.process)}: \$( gunc --version )
END_VERSIONS
"""
}

View file

@ -0,0 +1,36 @@
name: gunc_downloaddb
description: Download database for GUNC detection of Chimerism and Contamination in Prokaryotic Genomes
keywords:
- download
- prokaryote
- assembly
- genome
- quality control
- chimeras
tools:
- gunc:
description: Python package for detection of chimerism and contamination in prokaryotic genomes.
homepage: https://grp-bork.embl-community.io/gunc/
documentation: https://grp-bork.embl-community.io/gunc/
tool_dev_url: https://github.com/grp-bork/gunc
doi: "10.1186/s13059-021-02393-0"
licence: ['GNU General Public v3 or later (GPL v3+)']
input:
- db_name:
type: string
description: "Which database to download. Options: progenomes or gtdb"
pattern: "progenomes|gtdb"
output:
- versions:
type: file
description: File containing software versions
pattern: "versions.yml"
- db:
type: file
description: GUNC database file
pattern: "*.dmnd"
authors:
- "@jfy133"

View file

@ -0,0 +1,78 @@
//
// Utility functions used in nf-core DSL2 module files
//
//
// Extract name of software tool from process name using $task.process
//
def getSoftwareName(task_process) {
return task_process.tokenize(':')[-1].tokenize('_')[0].toLowerCase()
}
//
// Extract name of module from process name using $task.process
//
def getProcessName(task_process) {
return task_process.tokenize(':')[-1]
}
//
// Function to initialise default values and to generate a Groovy Map of available options for nf-core modules
//
def initOptions(Map args) {
def Map options = [:]
options.args = args.args ?: ''
options.args2 = args.args2 ?: ''
options.args3 = args.args3 ?: ''
options.publish_by_meta = args.publish_by_meta ?: []
options.publish_dir = args.publish_dir ?: ''
options.publish_files = args.publish_files
options.suffix = args.suffix ?: ''
return options
}
//
// Tidy up and join elements of a list to return a path string
//
def getPathFromList(path_list) {
def paths = path_list.findAll { item -> !item?.trim().isEmpty() } // Remove empty entries
paths = paths.collect { it.trim().replaceAll("^[/]+|[/]+\$", "") } // Trim whitespace and trailing slashes
return paths.join('/')
}
//
// Function to save/publish module results
//
def saveFiles(Map args) {
def ioptions = initOptions(args.options)
def path_list = [ ioptions.publish_dir ?: args.publish_dir ]
// Do not publish versions.yml unless running from pytest workflow
if (args.filename.equals('versions.yml') && !System.getenv("NF_CORE_MODULES_TEST")) {
return null
}
if (ioptions.publish_by_meta) {
def key_list = ioptions.publish_by_meta instanceof List ? ioptions.publish_by_meta : args.publish_by_meta
for (key in key_list) {
if (args.meta && key instanceof String) {
def path = key
if (args.meta.containsKey(key)) {
path = args.meta[key] instanceof Boolean ? "${key}_${args.meta[key]}".toString() : args.meta[key]
}
path = path instanceof String ? path : ''
path_list.add(path)
}
}
}
if (ioptions.publish_files instanceof Map) {
for (ext in ioptions.publish_files) {
if (args.filename.endsWith(ext.key)) {
def ext_list = path_list.collect()
ext_list.add(ext.value)
return "${getPathFromList(ext_list)}/$args.filename"
}
}
} else if (ioptions.publish_files == null) {
return "${getPathFromList(path_list)}/$args.filename"
}
}

45
modules/gunc/run/main.nf Normal file
View file

@ -0,0 +1,45 @@
// Import generic module functions
include { initOptions; saveFiles; getSoftwareName; getProcessName } from './functions'
params.options = [:]
options = initOptions(params.options)
process GUNC_RUN {
tag "$meta.id"
label 'process_medium'
publishDir "${params.outdir}",
mode: params.publish_dir_mode,
saveAs: { filename -> saveFiles(filename:filename, options:params.options, publish_dir:getSoftwareName(task.process), meta:meta, publish_by_meta:['id']) }
conda (params.enable_conda ? "bioconda::gunc=1.0.5" : null)
if (workflow.containerEngine == 'singularity' && !params.singularity_pull_docker_container) {
container "https://depot.galaxyproject.org/singularity/gunc:1.0.5--pyhdfd78af_0"
} else {
container "quay.io/biocontainers/gunc:1.0.5--pyhdfd78af_0"
}
input:
tuple val(meta), path(fasta)
path(db)
output:
tuple val(meta), path("*maxCSS_level.tsv") , emit: maxcss_level_tsv
tuple val(meta), path("*all_levels.tsv") , optional: true, emit: all_levels_tsv
path "versions.yml" , emit: versions
script:
def prefix = options.suffix ? "${meta.id}${options.suffix}" : "${meta.id}"
"""
gunc \\
run \\
--input_fasta $fasta \\
--db_file $db \\
--threads $task.cpus \\
$options.args
cat <<-END_VERSIONS > versions.yml
${getProcessName(task.process)}:
${getSoftwareName(task.process)}: \$( gunc --version )
END_VERSIONS
"""
}

53
modules/gunc/run/meta.yml Normal file
View file

@ -0,0 +1,53 @@
name: gunc_run
description: Detection of Chimerism and Contamination in Prokaryotic Genomes
keywords:
- prokaryote
- assembly
- genome
- quality control
- chimeras
tools:
- gunc:
description: Python package for detection of chimerism and contamination in prokaryotic genomes.
homepage: https://grp-bork.embl-community.io/gunc/
documentation: https://grp-bork.embl-community.io/gunc/
tool_dev_url: https://github.com/grp-bork/gunc
doi: "10.1186/s13059-021-02393-0"
licence: ['GNU General Public v3 or later (GPL v3+)']
input:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- fasta:
type: file
description: FASTA file containing contig (bins)
pattern: "*.fa"
- db:
type: file
description: GUNC database file
pattern: "*.dmnd"
output:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- versions:
type: file
description: File containing software versions
pattern: "versions.yml"
- maxcss_levels_tsv:
type: file
description: Output file with scores for a taxonomic level with the highest CSS score
pattern: "*.tsv"
- all_levels_tsv:
type: file
description: Optional output file with results for each taxonomic level
pattern: "*.tsv"
authors:
- "@jfy133"

View file

@ -0,0 +1,78 @@
//
// Utility functions used in nf-core DSL2 module files
//
//
// Extract name of software tool from process name using $task.process
//
def getSoftwareName(task_process) {
return task_process.tokenize(':')[-1].tokenize('_')[0].toLowerCase()
}
//
// Extract name of module from process name using $task.process
//
def getProcessName(task_process) {
return task_process.tokenize(':')[-1]
}
//
// Function to initialise default values and to generate a Groovy Map of available options for nf-core modules
//
def initOptions(Map args) {
def Map options = [:]
options.args = args.args ?: ''
options.args2 = args.args2 ?: ''
options.args3 = args.args3 ?: ''
options.publish_by_meta = args.publish_by_meta ?: []
options.publish_dir = args.publish_dir ?: ''
options.publish_files = args.publish_files
options.suffix = args.suffix ?: ''
return options
}
//
// Tidy up and join elements of a list to return a path string
//
def getPathFromList(path_list) {
def paths = path_list.findAll { item -> !item?.trim().isEmpty() } // Remove empty entries
paths = paths.collect { it.trim().replaceAll("^[/]+|[/]+\$", "") } // Trim whitespace and trailing slashes
return paths.join('/')
}
//
// Function to save/publish module results
//
def saveFiles(Map args) {
def ioptions = initOptions(args.options)
def path_list = [ ioptions.publish_dir ?: args.publish_dir ]
// Do not publish versions.yml unless running from pytest workflow
if (args.filename.equals('versions.yml') && !System.getenv("NF_CORE_MODULES_TEST")) {
return null
}
if (ioptions.publish_by_meta) {
def key_list = ioptions.publish_by_meta instanceof List ? ioptions.publish_by_meta : args.publish_by_meta
for (key in key_list) {
if (args.meta && key instanceof String) {
def path = key
if (args.meta.containsKey(key)) {
path = args.meta[key] instanceof Boolean ? "${key}_${args.meta[key]}".toString() : args.meta[key]
}
path = path instanceof String ? path : ''
path_list.add(path)
}
}
}
if (ioptions.publish_files instanceof Map) {
for (ext in ioptions.publish_files) {
if (args.filename.endsWith(ext.key)) {
def ext_list = path_list.collect()
ext_list.add(ext.value)
return "${getPathFromList(ext_list)}/$args.filename"
}
}
} else if (ioptions.publish_files == null) {
return "${getPathFromList(path_list)}/$args.filename"
}
}

View file

@ -0,0 +1,40 @@
// Import generic module functions
include { initOptions; saveFiles; getSoftwareName; getProcessName } from './functions'
params.options = [:]
options = initOptions(params.options)
def VERSION = '0.1.1'
process HMMCOPY_GCCOUNTER {
label 'process_low'
publishDir "${params.outdir}",
mode: params.publish_dir_mode,
saveAs: { filename -> saveFiles(filename:filename, options:params.options, publish_dir:getSoftwareName(task.process), meta:[:], publish_by_meta:[]) }
conda (params.enable_conda ? "bioconda::hmmcopy=0.1.1" : null)
if (workflow.containerEngine == 'singularity' && !params.singularity_pull_docker_container) {
container "https://depot.galaxyproject.org/singularity/hmmcopy:0.1.1--h2e03b76_5"
} else {
container "quay.io/biocontainers/hmmcopy:0.1.1--h2e03b76_5"
}
input:
path fasta
output:
path "*.gc.wig" , emit: wig
path "versions.yml", emit: versions
script:
"""
gcCounter \\
$options.args \\
${fasta} > ${fasta.baseName}.gc.wig
cat <<-END_VERSIONS > versions.yml
${getProcessName(task.process)}:
${getSoftwareName(task.process)}: \$(echo $VERSION)
END_VERSIONS
"""
}

View file

@ -0,0 +1,33 @@
name: hmmcopy_gccounter
description: gcCounter function from HMMcopy utilities, used to generate GC content in non-overlapping windows from a fasta reference
keywords:
- hmmcopy
- gccounter
- cnv
tools:
- hmmcopy:
description: C++ based programs for analyzing BAM files and preparing read counts -- used with bioconductor-hmmcopy
homepage: https://github.com/shahcompbio/hmmcopy_utils
documentation: https://github.com/shahcompbio/hmmcopy_utils
tool_dev_url: https://github.com/shahcompbio/hmmcopy_utils
doi: ""
licence: ['GPL v3']
input:
- fasta:
type: file
description: Input genome fasta file
output:
- versions:
type: file
description: File containing software versions
pattern: "versions.yml"
- wig:
type: file
description: wig file containing gc content of each window of the genome
pattern: "*.{gc.wig}"
authors:
- "@sppearce"

View file

@ -0,0 +1,78 @@
//
// Utility functions used in nf-core DSL2 module files
//
//
// Extract name of software tool from process name using $task.process
//
def getSoftwareName(task_process) {
return task_process.tokenize(':')[-1].tokenize('_')[0].toLowerCase()
}
//
// Extract name of module from process name using $task.process
//
def getProcessName(task_process) {
return task_process.tokenize(':')[-1]
}
//
// Function to initialise default values and to generate a Groovy Map of available options for nf-core modules
//
def initOptions(Map args) {
def Map options = [:]
options.args = args.args ?: ''
options.args2 = args.args2 ?: ''
options.args3 = args.args3 ?: ''
options.publish_by_meta = args.publish_by_meta ?: []
options.publish_dir = args.publish_dir ?: ''
options.publish_files = args.publish_files
options.suffix = args.suffix ?: ''
return options
}
//
// Tidy up and join elements of a list to return a path string
//
def getPathFromList(path_list) {
def paths = path_list.findAll { item -> !item?.trim().isEmpty() } // Remove empty entries
paths = paths.collect { it.trim().replaceAll("^[/]+|[/]+\$", "") } // Trim whitespace and trailing slashes
return paths.join('/')
}
//
// Function to save/publish module results
//
def saveFiles(Map args) {
def ioptions = initOptions(args.options)
def path_list = [ ioptions.publish_dir ?: args.publish_dir ]
// Do not publish versions.yml unless running from pytest workflow
if (args.filename.equals('versions.yml') && !System.getenv("NF_CORE_MODULES_TEST")) {
return null
}
if (ioptions.publish_by_meta) {
def key_list = ioptions.publish_by_meta instanceof List ? ioptions.publish_by_meta : args.publish_by_meta
for (key in key_list) {
if (args.meta && key instanceof String) {
def path = key
if (args.meta.containsKey(key)) {
path = args.meta[key] instanceof Boolean ? "${key}_${args.meta[key]}".toString() : args.meta[key]
}
path = path instanceof String ? path : ''
path_list.add(path)
}
}
}
if (ioptions.publish_files instanceof Map) {
for (ext in ioptions.publish_files) {
if (args.filename.endsWith(ext.key)) {
def ext_list = path_list.collect()
ext_list.add(ext.value)
return "${getPathFromList(ext_list)}/$args.filename"
}
}
} else if (ioptions.publish_files == null) {
return "${getPathFromList(path_list)}/$args.filename"
}
}

View file

@ -0,0 +1,42 @@
// Import generic module functions
include { initOptions; saveFiles; getSoftwareName; getProcessName } from './functions'
params.options = [:]
options = initOptions(params.options)
def VERSION = '0.1.1'
process HMMCOPY_READCOUNTER {
tag "$meta.id"
label 'process_low'
publishDir "${params.outdir}",
mode: params.publish_dir_mode,
saveAs: { filename -> saveFiles(filename:filename, options:params.options, publish_dir:getSoftwareName(task.process), meta:meta, publish_by_meta:['id']) }
conda (params.enable_conda ? "bioconda::hmmcopy=0.1.1" : null)
if (workflow.containerEngine == 'singularity' && !params.singularity_pull_docker_container) {
container "https://depot.galaxyproject.org/singularity/hmmcopy:0.1.1--h2e03b76_5"
} else {
container "quay.io/biocontainers/hmmcopy:0.1.1--h2e03b76_5"
}
input:
tuple val(meta), path(bam), path(bai)
output:
tuple val(meta), path("*.wig"), emit: wig
path "versions.yml" , emit: versions
script:
def prefix = options.suffix ? "${meta.id}${options.suffix}" : "${meta.id}"
"""
readCounter \\
$options.args \\
${bam} > ${prefix}.wig
cat <<-END_VERSIONS > versions.yml
${getProcessName(task.process)}:
${getSoftwareName(task.process)}: \$(echo $VERSION)
END_VERSIONS
"""
}

View file

@ -0,0 +1,43 @@
name: hmmcopy_readcounter
description: readCounter function from HMMcopy utilities, used to generate read in windows
keywords:
- hmmcopy
- readcounter
- cnv
tools:
- hmmcopy:
description: C++ based programs for analyzing BAM files and preparing read counts -- used with bioconductor-hmmcopy
homepage: https://github.com/shahcompbio/hmmcopy_utils
documentation: https://github.com/shahcompbio/hmmcopy_utils
tool_dev_url: https://github.com/shahcompbio/hmmcopy_utils
doi: ""
licence: ['GPL v3']
input:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- bam:
type: file
description: BAM/CRAM/SAM file
pattern: "*.{bam,cram,sam}"
output:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- versions:
type: file
description: File containing software versions
pattern: "versions.yml"
- wig:
type: file
description: A wig file with the number of reads lying within each window in each chromosome
pattern: "*.wig"
authors:
- "@sppearce"

78
modules/idr/functions.nf Normal file
View file

@ -0,0 +1,78 @@
//
// Utility functions used in nf-core DSL2 module files
//
//
// Extract name of software tool from process name using $task.process
//
def getSoftwareName(task_process) {
return task_process.tokenize(':')[-1].tokenize('_')[0].toLowerCase()
}
//
// Extract name of module from process name using $task.process
//
def getProcessName(task_process) {
return task_process.tokenize(':')[-1]
}
//
// Function to initialise default values and to generate a Groovy Map of available options for nf-core modules
//
def initOptions(Map args) {
def Map options = [:]
options.args = args.args ?: ''
options.args2 = args.args2 ?: ''
options.args3 = args.args3 ?: ''
options.publish_by_meta = args.publish_by_meta ?: []
options.publish_dir = args.publish_dir ?: ''
options.publish_files = args.publish_files
options.suffix = args.suffix ?: ''
return options
}
//
// Tidy up and join elements of a list to return a path string
//
def getPathFromList(path_list) {
def paths = path_list.findAll { item -> !item?.trim().isEmpty() } // Remove empty entries
paths = paths.collect { it.trim().replaceAll("^[/]+|[/]+\$", "") } // Trim whitespace and trailing slashes
return paths.join('/')
}
//
// Function to save/publish module results
//
def saveFiles(Map args) {
def ioptions = initOptions(args.options)
def path_list = [ ioptions.publish_dir ?: args.publish_dir ]
// Do not publish versions.yml unless running from pytest workflow
if (args.filename.equals('versions.yml') && !System.getenv("NF_CORE_MODULES_TEST")) {
return null
}
if (ioptions.publish_by_meta) {
def key_list = ioptions.publish_by_meta instanceof List ? ioptions.publish_by_meta : args.publish_by_meta
for (key in key_list) {
if (args.meta && key instanceof String) {
def path = key
if (args.meta.containsKey(key)) {
path = args.meta[key] instanceof Boolean ? "${key}_${args.meta[key]}".toString() : args.meta[key]
}
path = path instanceof String ? path : ''
path_list.add(path)
}
}
}
if (ioptions.publish_files instanceof Map) {
for (ext in ioptions.publish_files) {
if (args.filename.endsWith(ext.key)) {
def ext_list = path_list.collect()
ext_list.add(ext.value)
return "${getPathFromList(ext_list)}/$args.filename"
}
}
} else if (ioptions.publish_files == null) {
return "${getPathFromList(path_list)}/$args.filename"
}
}

56
modules/idr/main.nf Normal file
View file

@ -0,0 +1,56 @@
// Import generic module functions
include { initOptions; saveFiles; getSoftwareName; getProcessName } from './functions'
params.options = [:]
options = initOptions(params.options)
process IDR {
tag "$prefix"
label 'process_low'
publishDir "${params.outdir}",
mode: params.publish_dir_mode,
saveAs: { filename -> saveFiles(filename:filename, options:params.options, publish_dir:getSoftwareName(task.process), meta:[:], publish_by_meta:[]) }
conda (params.enable_conda ? "bioconda::idr=2.0.4.2" : null)
if (workflow.containerEngine == 'singularity' && !params.singularity_pull_docker_container) {
container "https://depot.galaxyproject.org/singularity/idr:2.0.4.2--py39hcbe4a3b_5"
} else {
container "quay.io/biocontainers/idr:2.0.4.2--py38h9af456f_5"
}
input:
path peaks
val peak_type
val prefix
output:
path "*idrValues.txt", emit: idr
path "*log.txt" , emit: log
path "*.png" , emit: png
path "versions.yml" , emit: versions
script:
if (peaks.toList().size < 2) {
log.error "[ERROR] idr needs at least two replicates only one provided."
}
def peak_types = ['narrowPeak', 'broadPeak', 'bed']
if (!peak_types.contains(peak_type)) {
log.error "[ERROR] Invalid option: '${peak_type}'. Valid options for 'peak_type': ${peak_types.join(', ')}."
}
def idr_vals = prefix ? "${prefix}.idrValues.txt" : "idrValues.txt"
def log_file = prefix ? "${prefix}.log.txt" : "log.txt"
"""
idr \\
--samples $peaks \\
--input-file-type $peak_type \\
--output-file $idr_vals \\
--log-output-file $log_file \\
--plot \\
$options.args
cat <<-END_VERSIONS > versions.yml
${getProcessName(task.process)}:
${getSoftwareName(task.process)}: \$(echo \$(idr --version 2>&1) | sed 's/^.*IDR //; s/ .*\$//')
END_VERSIONS
"""
}

53
modules/idr/meta.yml Normal file
View file

@ -0,0 +1,53 @@
name: idr
description: |
Measures reproducibility of ChIP-seq, ATAC-seq peaks using IDR (Irreproducible
Discovery Rate)
keywords:
- IDR
- peaks
- ChIP-seq
- ATAC-seq
tools:
- idr:
description: |
The IDR (Irreproducible Discovery Rate) framework is a unified approach
to measure the reproducibility of findings identified from replicate
experiments and provide highly stable thresholds based on reproducibility.
homepage: None
documentation: None
tool_dev_url: https://github.com/kundajelab/idr
doi: ""
licence: ['GPL v2']
input:
- peaks:
type: tuple of two files
description: BED, narrowPeak or broadPeak files of replicates
pattern: "*"
- peak_type:
type: value
description: Type of peak file
pattern: "{narrowPeak,broadPeak,bed}"
- prefix:
type: value
description: Prefix for output files
output:
- versions:
type: file
description: File containing software versions
pattern: "versions.yml"
- idr:
type: file
description: Text file containing IDR values
pattern: "*.{txt}"
- log:
type: file
description: Log file
pattern: "*.{txt}"
- png:
type: file
description: Plot generated by idr
pattern: "*{.png}"
authors:
- "@drpatelh"
- "@joseespinosa"

View file

@ -0,0 +1,78 @@
//
// Utility functions used in nf-core DSL2 module files
//
//
// Extract name of software tool from process name using $task.process
//
def getSoftwareName(task_process) {
return task_process.tokenize(':')[-1].tokenize('_')[0].toLowerCase()
}
//
// Extract name of module from process name using $task.process
//
def getProcessName(task_process) {
return task_process.tokenize(':')[-1]
}
//
// Function to initialise default values and to generate a Groovy Map of available options for nf-core modules
//
def initOptions(Map args) {
def Map options = [:]
options.args = args.args ?: ''
options.args2 = args.args2 ?: ''
options.args3 = args.args3 ?: ''
options.publish_by_meta = args.publish_by_meta ?: []
options.publish_dir = args.publish_dir ?: ''
options.publish_files = args.publish_files
options.suffix = args.suffix ?: ''
return options
}
//
// Tidy up and join elements of a list to return a path string
//
def getPathFromList(path_list) {
def paths = path_list.findAll { item -> !item?.trim().isEmpty() } // Remove empty entries
paths = paths.collect { it.trim().replaceAll("^[/]+|[/]+\$", "") } // Trim whitespace and trailing slashes
return paths.join('/')
}
//
// Function to save/publish module results
//
def saveFiles(Map args) {
def ioptions = initOptions(args.options)
def path_list = [ ioptions.publish_dir ?: args.publish_dir ]
// Do not publish versions.yml unless running from pytest workflow
if (args.filename.equals('versions.yml') && !System.getenv("NF_CORE_MODULES_TEST")) {
return null
}
if (ioptions.publish_by_meta) {
def key_list = ioptions.publish_by_meta instanceof List ? ioptions.publish_by_meta : args.publish_by_meta
for (key in key_list) {
if (args.meta && key instanceof String) {
def path = key
if (args.meta.containsKey(key)) {
path = args.meta[key] instanceof Boolean ? "${key}_${args.meta[key]}".toString() : args.meta[key]
}
path = path instanceof String ? path : ''
path_list.add(path)
}
}
}
if (ioptions.publish_files instanceof Map) {
for (ext in ioptions.publish_files) {
if (args.filename.endsWith(ext.key)) {
def ext_list = path_list.collect()
ext_list.add(ext.value)
return "${getPathFromList(ext_list)}/$args.filename"
}
}
} else if (ioptions.publish_files == null) {
return "${getPathFromList(path_list)}/$args.filename"
}
}

View file

@ -0,0 +1,43 @@
include { initOptions; saveFiles; getSoftwareName; getProcessName } from './functions'
params.options = [:]
options = initOptions(params.options)
process MTNUCRATIO {
tag "$meta.id"
label 'process_low'
publishDir "${params.outdir}",
mode: params.publish_dir_mode,
saveAs: { filename -> saveFiles(filename:filename, options:params.options, publish_dir:getSoftwareName(task.process), meta:meta, publish_by_meta:['id']) }
conda (params.enable_conda ? "bioconda::mtnucratio=0.7" : null)
if (workflow.containerEngine == 'singularity' && !params.singularity_pull_docker_container) {
container "https://depot.galaxyproject.org/singularity/mtnucratio:0.7--hdfd78af_2"
} else {
container "quay.io/biocontainers/mtnucratio:0.7--hdfd78af_2"
}
input:
tuple val(meta), path(bam)
val(mt_id)
output:
tuple val(meta), path("*.mtnucratio"), emit: mtnucratio
tuple val(meta), path("*.json") , emit: json
path "versions.yml" , emit: versions
script:
def prefix = options.suffix ? "${meta.id}${options.suffix}" : "${meta.id}"
"""
mtnucratio \\
$options.args \\
$bam \\
$mt_id
cat <<-END_VERSIONS > versions.yml
${getProcessName(task.process)}:
${getSoftwareName(task.process)}: \$(echo \$(mtnucratio --version 2>&1) | head -n1 | sed 's/Version: //')
END_VERSIONS
"""
}

View file

@ -0,0 +1,54 @@
name: mtnucratio
description: A small Java tool to calculate ratios between MT and nuclear sequencing reads in a given BAM file.
keywords:
- mtnucratio
- ratio
- reads
- bam
- mitochondrial to nuclear ratio
- mitochondria
- statistics
tools:
- mtnucratio:
description: A small tool to determine MT to Nuclear ratios for NGS data.
homepage: https://github.com/apeltzer/MTNucRatioCalculator
documentation: https://github.com/apeltzer/MTNucRatioCalculator
tool_dev_url: https://github.com/apeltzer/MTNucRatioCalculator
doi: "10.1186/s13059-016-0918-z"
licence: ['GPL v3']
input:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- bam:
type: file
description: (coordinate) sorted BAM/SAM file
pattern: "*.{bam,sam}"
- mt_id:
type: string
description: Identifier of the contig/chromosome of interest (e.g. chromosome, contig) as in the aligned against reference FASTA file, e.g. mt or chrMT for mitochondria
output:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- versions:
type: file
description: File containing software versions
pattern: "versions.yml"
- mtnucratio:
type: file
description: Text file containing metrics (mtreads, mt_cov_avg, nucreads, nuc_cov_avg, mt_nuc_ratio)
pattern: "*.mtnucratio"
- json:
type: file
description: JSON file, containing metadata map with sample name, tool name and version, and metrics as in txt file
pattern: "*.json"
authors:
- "@louperelo"

View file

@ -0,0 +1,78 @@
//
// Utility functions used in nf-core DSL2 module files
//
//
// Extract name of software tool from process name using $task.process
//
def getSoftwareName(task_process) {
return task_process.tokenize(':')[-1].tokenize('_')[0].toLowerCase()
}
//
// Extract name of module from process name using $task.process
//
def getProcessName(task_process) {
return task_process.tokenize(':')[-1]
}
//
// Function to initialise default values and to generate a Groovy Map of available options for nf-core modules
//
def initOptions(Map args) {
def Map options = [:]
options.args = args.args ?: ''
options.args2 = args.args2 ?: ''
options.args3 = args.args3 ?: ''
options.publish_by_meta = args.publish_by_meta ?: []
options.publish_dir = args.publish_dir ?: ''
options.publish_files = args.publish_files
options.suffix = args.suffix ?: ''
return options
}
//
// Tidy up and join elements of a list to return a path string
//
def getPathFromList(path_list) {
def paths = path_list.findAll { item -> !item?.trim().isEmpty() } // Remove empty entries
paths = paths.collect { it.trim().replaceAll("^[/]+|[/]+\$", "") } // Trim whitespace and trailing slashes
return paths.join('/')
}
//
// Function to save/publish module results
//
def saveFiles(Map args) {
def ioptions = initOptions(args.options)
def path_list = [ ioptions.publish_dir ?: args.publish_dir ]
// Do not publish versions.yml unless running from pytest workflow
if (args.filename.equals('versions.yml') && !System.getenv("NF_CORE_MODULES_TEST")) {
return null
}
if (ioptions.publish_by_meta) {
def key_list = ioptions.publish_by_meta instanceof List ? ioptions.publish_by_meta : args.publish_by_meta
for (key in key_list) {
if (args.meta && key instanceof String) {
def path = key
if (args.meta.containsKey(key)) {
path = args.meta[key] instanceof Boolean ? "${key}_${args.meta[key]}".toString() : args.meta[key]
}
path = path instanceof String ? path : ''
path_list.add(path)
}
}
}
if (ioptions.publish_files instanceof Map) {
for (ext in ioptions.publish_files) {
if (args.filename.endsWith(ext.key)) {
def ext_list = path_list.collect()
ext_list.add(ext.value)
return "${getPathFromList(ext_list)}/$args.filename"
}
}
} else if (ioptions.publish_files == null) {
return "${getPathFromList(path_list)}/$args.filename"
}
}

45
modules/racon/main.nf Normal file
View file

@ -0,0 +1,45 @@
// Import generic module functions
include { initOptions; saveFiles; getSoftwareName; getProcessName } from './functions'
params.options = [:]
options = initOptions(params.options)
process RACON {
tag "$meta.id"
label 'process_high'
publishDir "${params.outdir}",
mode: params.publish_dir_mode,
saveAs: { filename -> saveFiles(filename:filename, options:params.options, publish_dir:getSoftwareName(task.process), meta:meta, publish_by_meta:['id']) }
conda (params.enable_conda ? "bioconda::racon=1.4.20" : null)
if (workflow.containerEngine == 'singularity' && !params.singularity_pull_docker_container) {
container "https://depot.galaxyproject.org/singularity/racon:1.4.20--h9a82719_1"
} else {
container "quay.io/biocontainers/racon:1.4.20--h9a82719_1"
}
input:
tuple val(meta), path(reads), path(assembly), path(paf)
output:
tuple val(meta), path('*_assembly_consensus.fasta.gz') , emit: improved_assembly
path "versions.yml" , emit: versions
script:
def prefix = options.suffix ? "${meta.id}${options.suffix}" : "${meta.id}"
"""
racon -t "${task.cpus}" \\
"${reads}" \\
"${paf}" \\
$options.args \\
"${assembly}" > \\
${prefix}_assembly_consensus.fasta
gzip -n ${prefix}_assembly_consensus.fasta
cat <<-END_VERSIONS > versions.yml
${getProcessName(task.process)}:
${getSoftwareName(task.process)}: \$( racon --version 2>&1 | sed 's/^.*v//' )
END_VERSIONS
"""
}

52
modules/racon/meta.yml Normal file
View file

@ -0,0 +1,52 @@
name: racon
description: Consensus module for raw de novo DNA assembly of long uncorrected reads
keywords:
- assembly
- pacbio
- nanopore
- polish
tools:
- racon:
description: Ultrafast consensus module for raw de novo genome assembly of long uncorrected reads.
homepage: https://github.com/lbcb-sci/racon
documentation: https://github.com/lbcb-sci/racon
tool_dev_url: https://github.com/lbcb-sci/racon
doi: https://doi.org/10.1101/gr.214270.116
licence: ['MIT']
input:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- reads:
type: file
description: List of input FastQ files. Racon expects single end reads
pattern: "*.{fastq,fastq.gz,fq,fq.gz}"
- assembly:
type: file
description: Genome assembly to be improved
pattern: "*.{fasta,fa}"
- paf:
type: file
description: Alignment in PAF format
pattern: "*.paf"
output:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- versions:
type: file
description: File containing software versions
pattern: "versions.yml"
- improved_assembly:
type: file
description: Improved genome assembly
pattern: "*_assembly_consensus.fasta.gz"
authors:
- "@avantonder"

View file

@ -0,0 +1,78 @@
//
// Utility functions used in nf-core DSL2 module files
//
//
// Extract name of software tool from process name using $task.process
//
def getSoftwareName(task_process) {
return task_process.tokenize(':')[-1].tokenize('_')[0].toLowerCase()
}
//
// Extract name of module from process name using $task.process
//
def getProcessName(task_process) {
return task_process.tokenize(':')[-1]
}
//
// Function to initialise default values and to generate a Groovy Map of available options for nf-core modules
//
def initOptions(Map args) {
def Map options = [:]
options.args = args.args ?: ''
options.args2 = args.args2 ?: ''
options.args3 = args.args3 ?: ''
options.publish_by_meta = args.publish_by_meta ?: []
options.publish_dir = args.publish_dir ?: ''
options.publish_files = args.publish_files
options.suffix = args.suffix ?: ''
return options
}
//
// Tidy up and join elements of a list to return a path string
//
def getPathFromList(path_list) {
def paths = path_list.findAll { item -> !item?.trim().isEmpty() } // Remove empty entries
paths = paths.collect { it.trim().replaceAll("^[/]+|[/]+\$", "") } // Trim whitespace and trailing slashes
return paths.join('/')
}
//
// Function to save/publish module results
//
def saveFiles(Map args) {
def ioptions = initOptions(args.options)
def path_list = [ ioptions.publish_dir ?: args.publish_dir ]
// Do not publish versions.yml unless running from pytest workflow
if (args.filename.equals('versions.yml') && !System.getenv("NF_CORE_MODULES_TEST")) {
return null
}
if (ioptions.publish_by_meta) {
def key_list = ioptions.publish_by_meta instanceof List ? ioptions.publish_by_meta : args.publish_by_meta
for (key in key_list) {
if (args.meta && key instanceof String) {
def path = key
if (args.meta.containsKey(key)) {
path = args.meta[key] instanceof Boolean ? "${key}_${args.meta[key]}".toString() : args.meta[key]
}
path = path instanceof String ? path : ''
path_list.add(path)
}
}
}
if (ioptions.publish_files instanceof Map) {
for (ext in ioptions.publish_files) {
if (args.filename.endsWith(ext.key)) {
def ext_list = path_list.collect()
ext_list.add(ext.value)
return "${getPathFromList(ext_list)}/$args.filename"
}
}
} else if (ioptions.publish_files == null) {
return "${getPathFromList(path_list)}/$args.filename"
}
}

View file

@ -0,0 +1,45 @@
// Import generic module functions
include { initOptions; saveFiles; getSoftwareName; getProcessName } from './functions'
params.options = [:]
options = initOptions(params.options)
process SAMTOOLS_FIXMATE {
tag "$meta.id"
label 'process_low'
publishDir "${params.outdir}",
mode: params.publish_dir_mode,
saveAs: { filename -> saveFiles(filename:filename, options:params.options, publish_dir:getSoftwareName(task.process), meta:meta, publish_by_meta:['id']) }
conda (params.enable_conda ? "bioconda::samtools=1.14" : null)
if (workflow.containerEngine == 'singularity' && !params.singularity_pull_docker_container) {
container "https://depot.galaxyproject.org/singularity/samtools:1.14--hb421002_0"
} else {
container "quay.io/biocontainers/samtools:1.14--hb421002_0"
}
input:
tuple val(meta), path(bam)
output:
tuple val(meta), path("*.bam"), emit: bam
path "versions.yml" , emit: versions
script:
def prefix = options.suffix ? "${meta.id}${options.suffix}" : "${meta.id}"
if ("$bam" == "${prefix}.bam") error "Input and output names are the same, use the suffix option to disambiguate!"
"""
samtools \\
fixmate \\
$options.args \\
-@ $task.cpus \\
$bam \\
${prefix}.bam \\
cat <<-END_VERSIONS > versions.yml
${getProcessName(task.process)}:
${getSoftwareName(task.process)}: \$(echo \$(samtools --version 2>&1) | sed 's/^.*samtools //; s/Using.*\$//')
END_VERSIONS
"""
}

View file

@ -0,0 +1,49 @@
name: samtools_fixmate
description: Samtools fixmate is a tool that can fill in information (insert size, cigar, mapq) about paired end reads onto the corresponding other read. Also has options to remove secondary/unmapped alignments and recalculate whether reads are proper pairs.
keywords:
- fixmate
- samtools
- insert size
- repair
- bam
- paired
- read pairs
tools:
- samtools:
description: |
SAMtools is a set of utilities for interacting with and post-processing
short DNA sequence read alignments in the SAM, BAM and CRAM formats, written by Heng Li.
These files are generated as output by short read aligners like BWA.
homepage: http://www.htslib.org/
documentation: http://www.htslib.org/doc/samtools.html
tool_dev_url: https://github.com/samtools/samtools
doi: 10.1093/bioinformatics/btp352
licence: ['MIT']
input:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- bam:
type: file
description: BAM/CRAM/SAM file, must be sorted by name, not coordinate
pattern: "*.{bam,cram,sam}"
output:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- versions:
type: file
description: File containing software versions
pattern: "versions.yml"
- bam:
type: file
description: A BAM/CRAM/SAM file with mate information added and/or proper pairs recalled
pattern: "*.{bam,cram,sam}"
authors:
- "@sppearce"

View file

@ -0,0 +1,78 @@
//
// Utility functions used in nf-core DSL2 module files
//
//
// Extract name of software tool from process name using $task.process
//
def getSoftwareName(task_process) {
return task_process.tokenize(':')[-1].tokenize('_')[0].toLowerCase()
}
//
// Extract name of module from process name using $task.process
//
def getProcessName(task_process) {
return task_process.tokenize(':')[-1]
}
//
// Function to initialise default values and to generate a Groovy Map of available options for nf-core modules
//
def initOptions(Map args) {
def Map options = [:]
options.args = args.args ?: ''
options.args2 = args.args2 ?: ''
options.args3 = args.args3 ?: ''
options.publish_by_meta = args.publish_by_meta ?: []
options.publish_dir = args.publish_dir ?: ''
options.publish_files = args.publish_files
options.suffix = args.suffix ?: ''
return options
}
//
// Tidy up and join elements of a list to return a path string
//
def getPathFromList(path_list) {
def paths = path_list.findAll { item -> !item?.trim().isEmpty() } // Remove empty entries
paths = paths.collect { it.trim().replaceAll("^[/]+|[/]+\$", "") } // Trim whitespace and trailing slashes
return paths.join('/')
}
//
// Function to save/publish module results
//
def saveFiles(Map args) {
def ioptions = initOptions(args.options)
def path_list = [ ioptions.publish_dir ?: args.publish_dir ]
// Do not publish versions.yml unless running from pytest workflow
if (args.filename.equals('versions.yml') && !System.getenv("NF_CORE_MODULES_TEST")) {
return null
}
if (ioptions.publish_by_meta) {
def key_list = ioptions.publish_by_meta instanceof List ? ioptions.publish_by_meta : args.publish_by_meta
for (key in key_list) {
if (args.meta && key instanceof String) {
def path = key
if (args.meta.containsKey(key)) {
path = args.meta[key] instanceof Boolean ? "${key}_${args.meta[key]}".toString() : args.meta[key]
}
path = path instanceof String ? path : ''
path_list.add(path)
}
}
}
if (ioptions.publish_files instanceof Map) {
for (ext in ioptions.publish_files) {
if (args.filename.endsWith(ext.key)) {
def ext_list = path_list.collect()
ext_list.add(ext.value)
return "${getPathFromList(ext_list)}/$args.filename"
}
}
} else if (ioptions.publish_files == null) {
return "${getPathFromList(path_list)}/$args.filename"
}
}

45
modules/seqsero2/main.nf Normal file
View file

@ -0,0 +1,45 @@
// Import generic module functions
include { initOptions; saveFiles; getSoftwareName; getProcessName } from './functions'
params.options = [:]
options = initOptions(params.options)
process SEQSERO2 {
tag "$meta.id"
label 'process_low'
publishDir "${params.outdir}",
mode: params.publish_dir_mode,
saveAs: { filename -> saveFiles(filename:filename, options:params.options, publish_dir:getSoftwareName(task.process), meta:meta, publish_by_meta:['id']) }
conda (params.enable_conda ? "bioconda::seqsero2=1.2.1" : null)
if (workflow.containerEngine == 'singularity' && !params.singularity_pull_docker_container) {
container "https://depot.galaxyproject.org/singularity/seqsero2:1.2.1--py_0"
} else {
container "quay.io/biocontainers/seqsero2:1.2.1--py_0"
}
input:
tuple val(meta), path(seqs)
output:
tuple val(meta), path("results/*_log.txt") , emit: log
tuple val(meta), path("results/*_result.tsv"), emit: tsv
tuple val(meta), path("results/*_result.txt"), emit: txt
path "versions.yml" , emit: versions
script:
def prefix = options.suffix ? "${meta.id}${options.suffix}" : "${meta.id}"
"""
SeqSero2_package.py \\
$options.args \\
-d results/ \\
-n $prefix \\
-p $task.cpus \\
-i $seqs
cat <<-END_VERSIONS > versions.yml
${getProcessName(task.process)}:
${getSoftwareName(task.process)}: \$( echo \$( SeqSero2_package.py --version 2>&1) | sed 's/^.*SeqSero2_package.py //' )
END_VERSIONS
"""
}

52
modules/seqsero2/meta.yml Normal file
View file

@ -0,0 +1,52 @@
name: seqsero2
description: Salmonella serotype prediction from reads and assemblies
keywords:
- fasta
- fastq
- salmonella
- sertotype
tools:
- seqsero2:
description: Salmonella serotype prediction from genome sequencing data
homepage: https://github.com/denglab/SeqSero2
documentation: https://github.com/denglab/SeqSero2
tool_dev_url: https://github.com/denglab/SeqSero2
doi: "10.1128/AEM.01746-19"
licence: ['GPL v2']
input:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- seqs:
type: file
description: FASTQ or FASTA formated sequences
pattern: "*.{fq.gz,fastq.gz,fna.gz,fna,fasta.gz,fasta,fa.gz,fa}"
output:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- versions:
type: file
description: File containing software versions
pattern: "versions.yml"
- log:
type: file
description: A log of serotype antigen results
pattern: "*_log.txt"
- tsv:
type: file
description: Tab-delimited summary of the SeqSero2 results
pattern: "*_result.tsv"
- txt:
type: file
description: Detailed summary of the SeqSero2 results
pattern: "*_result.txt"
authors:
- "@rpetit3"

View file

@ -450,6 +450,10 @@ gatk4/filtermutectcalls:
- modules/gatk4/filtermutectcalls/** - modules/gatk4/filtermutectcalls/**
- tests/modules/gatk4/filtermutectcalls/** - tests/modules/gatk4/filtermutectcalls/**
gatk4/genomicsdbimport:
- modules/gatk4/genomicsdbimport/**
- tests/modules/gatk4/genomicsdbimport/**
gatk4/getpileupsummaries: gatk4/getpileupsummaries:
- modules/gatk4/getpileupsummaries/** - modules/gatk4/getpileupsummaries/**
- tests/modules/gatk4/getpileupsummaries/** - tests/modules/gatk4/getpileupsummaries/**
@ -542,6 +546,14 @@ gubbins:
- modules/gubbins/** - modules/gubbins/**
- tests/modules/gubbins/** - tests/modules/gubbins/**
gunc/downloaddb:
- modules/gunc/downloaddb/**
- tests/modules/gunc/downloaddb/**
gunc/run:
- modules/gunc/run/**
- tests/modules/gunc/run/**
gunzip: gunzip:
- modules/gunzip/** - modules/gunzip/**
- tests/modules/gunzip/** - tests/modules/gunzip/**
@ -569,6 +581,14 @@ hisat2/extractsplicesites:
- modules/hisat2/extractsplicesites/** - modules/hisat2/extractsplicesites/**
- tests/modules/hisat2/extractsplicesites/** - tests/modules/hisat2/extractsplicesites/**
hmmcopy/gccounter:
- modules/hmmcopy/gccounter/**
- tests/modules/hmmcopy/gccounter/**
hmmcopy/readcounter:
- modules/hmmcopy/readcounter/**
- tests/modules/hmmcopy/readcounter/**
hmmer/hmmalign: hmmer/hmmalign:
- modules/hmmer/hmmalign/** - modules/hmmer/hmmalign/**
- tests/modules/hmmer/hmmalign/** - tests/modules/hmmer/hmmalign/**
@ -589,6 +609,10 @@ homer/makeucscfile:
- modules/homer/makeucscfile/** - modules/homer/makeucscfile/**
- tests/modules/homer/makeucscfile/** - tests/modules/homer/makeucscfile/**
idr:
- modules/idr/**
- tests/modules/idr/**
iqtree: iqtree:
- modules/iqtree/** - modules/iqtree/**
- tests/modules/iqtree/** - tests/modules/iqtree/**
@ -778,6 +802,10 @@ msisensor/scan:
- modules/msisensor/scan/** - modules/msisensor/scan/**
- tests/modules/msisensor/scan/** - tests/modules/msisensor/scan/**
mtnucratio:
- modules/mtnucratio/**
- tests/modules/mtnucratio/**
multiqc: multiqc:
- modules/fastqc/** - modules/fastqc/**
- modules/multiqc/** - modules/multiqc/**
@ -923,6 +951,10 @@ quast:
- modules/quast/** - modules/quast/**
- tests/modules/quast/** - tests/modules/quast/**
racon:
- modules/racon/**
- tests/modules/racon/**
rapidnj: rapidnj:
- modules/rapidnj/** - modules/rapidnj/**
- tests/modules/rapidnj/** - tests/modules/rapidnj/**
@ -1007,6 +1039,10 @@ samtools/fastq:
- modules/samtools/fastq/** - modules/samtools/fastq/**
- tests/modules/samtools/fastq/** - tests/modules/samtools/fastq/**
samtools/fixmate:
- modules/samtools/fixmate/**
- tests/modules/samtools/fixmate/**
samtools/flagstat: samtools/flagstat:
- modules/samtools/flagstat/** - modules/samtools/flagstat/**
- tests/modules/samtools/flagstat/** - tests/modules/samtools/flagstat/**
@ -1047,6 +1083,10 @@ seqkit/split2:
- modules/seqkit/split2/** - modules/seqkit/split2/**
- tests/modules/seqkit/split2/** - tests/modules/seqkit/split2/**
seqsero2:
- modules/seqsero2/**
- tests/modules/seqsero2/**
seqtk/mergepe: seqtk/mergepe:
- modules/seqtk/mergepe/** - modules/seqtk/mergepe/**
- tests/modules/seqtk/mergepe/** - tests/modules/seqtk/mergepe/**

View file

@ -11,6 +11,7 @@ params {
genome_gff3 = "${test_data_dir}/genomics/sarscov2/genome/genome.gff3" genome_gff3 = "${test_data_dir}/genomics/sarscov2/genome/genome.gff3"
genome_gff3_gz = "${test_data_dir}/genomics/sarscov2/genome/genome.gff3.gz" genome_gff3_gz = "${test_data_dir}/genomics/sarscov2/genome/genome.gff3.gz"
genome_gtf = "${test_data_dir}/genomics/sarscov2/genome/genome.gtf" genome_gtf = "${test_data_dir}/genomics/sarscov2/genome/genome.gtf"
genome_paf = "${test_data_dir}/genomics/sarscov2/genome/genome.paf"
genome_sizes = "${test_data_dir}/genomics/sarscov2/genome/genome.sizes" genome_sizes = "${test_data_dir}/genomics/sarscov2/genome/genome.sizes"
transcriptome_fasta = "${test_data_dir}/genomics/sarscov2/genome/transcriptome.fasta" transcriptome_fasta = "${test_data_dir}/genomics/sarscov2/genome/transcriptome.fasta"
transcriptome_paf = "${test_data_dir}/genomics/sarscov2/genome/transcriptome.paf" transcriptome_paf = "${test_data_dir}/genomics/sarscov2/genome/transcriptome.paf"
@ -102,6 +103,7 @@ params {
genome_dict = "${test_data_dir}/genomics/homo_sapiens/genome/genome.dict" genome_dict = "${test_data_dir}/genomics/homo_sapiens/genome/genome.dict"
genome_gff3 = "${test_data_dir}/genomics/homo_sapiens/genome/genome.gff3" genome_gff3 = "${test_data_dir}/genomics/homo_sapiens/genome/genome.gff3"
genome_gtf = "${test_data_dir}/genomics/homo_sapiens/genome/genome.gtf" genome_gtf = "${test_data_dir}/genomics/homo_sapiens/genome/genome.gtf"
genome_interval_list = "${test_data_dir}/genomics/homo_sapiens/genome/genome.interval_list"
genome_sizes = "${test_data_dir}/genomics/homo_sapiens/genome/genome.sizes" genome_sizes = "${test_data_dir}/genomics/homo_sapiens/genome/genome.sizes"
genome_bed = "${test_data_dir}/genomics/homo_sapiens/genome/genome.bed" genome_bed = "${test_data_dir}/genomics/homo_sapiens/genome/genome.bed"
genome_header = "${test_data_dir}/genomics/homo_sapiens/genome/genome.header" genome_header = "${test_data_dir}/genomics/homo_sapiens/genome/genome.header"
@ -180,6 +182,7 @@ params {
test2_baserecalibrator_table = "${test_data_dir}/genomics/homo_sapiens/illumina/gatk/test2.baserecalibrator.table" test2_baserecalibrator_table = "${test_data_dir}/genomics/homo_sapiens/illumina/gatk/test2.baserecalibrator.table"
test_pileups_table = "${test_data_dir}/genomics/homo_sapiens/illumina/gatk/test.pileups.table" test_pileups_table = "${test_data_dir}/genomics/homo_sapiens/illumina/gatk/test.pileups.table"
test2_pileups_table = "${test_data_dir}/genomics/homo_sapiens/illumina/gatk/test2.pileups.table" test2_pileups_table = "${test_data_dir}/genomics/homo_sapiens/illumina/gatk/test2.pileups.table"
test_genomicsdb_tar_gz = "${test_data_dir}/genomics/homo_sapiens/illumina/gatk/test_genomicsdb.tar.gz"
test_test2_paired_mutect2_calls_vcf_gz = "${test_data_dir}/genomics/homo_sapiens/illumina/gatk/paired_mutect2_calls/test_test2_paired_mutect2_calls.vcf.gz" test_test2_paired_mutect2_calls_vcf_gz = "${test_data_dir}/genomics/homo_sapiens/illumina/gatk/paired_mutect2_calls/test_test2_paired_mutect2_calls.vcf.gz"
test_test2_paired_mutect2_calls_vcf_gz_tbi = "${test_data_dir}/genomics/homo_sapiens/illumina/gatk/paired_mutect2_calls/test_test2_paired_mutect2_calls.vcf.gz.tbi" test_test2_paired_mutect2_calls_vcf_gz_tbi = "${test_data_dir}/genomics/homo_sapiens/illumina/gatk/paired_mutect2_calls/test_test2_paired_mutect2_calls.vcf.gz.tbi"
@ -199,6 +202,12 @@ params {
test2_genome_vcf_gz_tbi = "${test_data_dir}/genomics/homo_sapiens/illumina/gvcf/test2.genome.vcf.gz.tbi" test2_genome_vcf_gz_tbi = "${test_data_dir}/genomics/homo_sapiens/illumina/gvcf/test2.genome.vcf.gz.tbi"
test2_genome_vcf_idx = "${test_data_dir}/genomics/homo_sapiens/illumina/gvcf/test2.genome.vcf.idx" test2_genome_vcf_idx = "${test_data_dir}/genomics/homo_sapiens/illumina/gvcf/test2.genome.vcf.idx"
test_broadpeak = "${test_data_dir}/genomics/homo_sapiens/illumina/broadpeak/test.broadPeak"
test2_broadpeak = "${test_data_dir}/genomics/homo_sapiens/illumina/broadpeak/test2.broadPeak"
test_narrowpeak = "${test_data_dir}/genomics/homo_sapiens/illumina/narrowpeak/test.narrowPeak"
test2_narrowpeak = "${test_data_dir}/genomics/homo_sapiens/illumina/narrowpeak/test2.narrowPeak"
test_10x_1_fastq_gz = "${test_data_dir}/genomics/homo_sapiens/illumina/10xgenomics/test.10x_1.fastq.gz" test_10x_1_fastq_gz = "${test_data_dir}/genomics/homo_sapiens/illumina/10xgenomics/test.10x_1.fastq.gz"
test_10x_2_fastq_gz = "${test_data_dir}/genomics/homo_sapiens/illumina/10xgenomics/test.10x_2.fastq.gz" test_10x_2_fastq_gz = "${test_data_dir}/genomics/homo_sapiens/illumina/10xgenomics/test.10x_2.fastq.gz"
@ -243,6 +252,7 @@ params {
'bacteroides_fragilis'{ 'bacteroides_fragilis'{
'genome' { 'genome' {
genome_fna_gz = "${test_data_dir}/genomics/bacteroides_fragilis/genome/genome.fna.gz" genome_fna_gz = "${test_data_dir}/genomics/bacteroides_fragilis/genome/genome.fna.gz"
genome_paf = "${test_data_dir}/genomics/bacteroides_fragilis/genome/genome.paf"
} }
'illumina' { 'illumina' {
test1_contigs_fa_gz = "${test_data_dir}/genomics/bacteroides_fragilis/illumina/fasta/test1.contigs.fa.gz" test1_contigs_fa_gz = "${test_data_dir}/genomics/bacteroides_fragilis/illumina/fasta/test1.contigs.fa.gz"
@ -253,6 +263,7 @@ params {
} }
'nanopore' { 'nanopore' {
test_fastq_gz = "${test_data_dir}/genomics/bacteroides_fragilis/nanopore/fastq/test.fastq.gz" test_fastq_gz = "${test_data_dir}/genomics/bacteroides_fragilis/nanopore/fastq/test.fastq.gz"
overlap_paf = "${test_data_dir}/genomics/bacteroides_fragilis/nanopore/overlap.paf"
} }
} }
} }

View file

@ -3,12 +3,24 @@ nextflow.enable.dsl = 2
include { ALLELECOUNTER } from '../../../modules/allelecounter/main.nf' addParams( options: [:] ) include { ALLELECOUNTER } from '../../../modules/allelecounter/main.nf' addParams( options: [:] )
workflow test_allelecounter { workflow test_allelecounter_bam {
input = [ [ id:'test', single_end:false ], // meta map input = [ [ id:'test', single_end:false ], // meta map
file(params.test_data['sarscov2']['illumina']['test_paired_end_sorted_bam'], checkIfExists: true), file(params.test_data['sarscov2']['illumina']['test_paired_end_sorted_bam'], checkIfExists: true),
file(params.test_data['sarscov2']['illumina']['test_paired_end_sorted_bam_bai'], checkIfExists: true) file(params.test_data['sarscov2']['illumina']['test_paired_end_sorted_bam_bai'], checkIfExists: true)
] ]
positions = [ file(params.test_data['sarscov2']['genome']['test_bed'], checkIfExists: true) ] positions = [ file(params.test_data['sarscov2']['genome']['test_bed'], checkIfExists: true) ]
ALLELECOUNTER ( input, positions ) ALLELECOUNTER ( input, positions, [] )
}
workflow test_allelecounter_cram {
input = [ [ id:'test', single_end:false ], // meta map
file(params.test_data['homo_sapiens']['illumina']['test_paired_end_sorted_cram'], checkIfExists: true),
file(params.test_data['homo_sapiens']['illumina']['test_paired_end_sorted_cram_crai'], checkIfExists: true)
]
positions = [ file(params.test_data['homo_sapiens']['genome']['genome_bed'], checkIfExists: true) ]
fasta = [ file(params.test_data['homo_sapiens']['genome']['genome_fasta'], checkIfExists: true) ]
ALLELECOUNTER ( input, positions, fasta )
} }

View file

@ -1,7 +1,15 @@
- name: allelecounter test_allelecounter - name: allelecounter test_allelecounter_bam
command: nextflow run tests/modules/allelecounter -entry test_allelecounter -c tests/config/nextflow.config command: nextflow run tests/modules/allelecounter -entry test_allelecounter_bam -c tests/config/nextflow.config
tags: tags:
- allelecounter - allelecounter
files: files:
- path: output/allelecounter/test.alleleCount - path: output/allelecounter/test.alleleCount
md5sum: 2bbe9d7331b78bdac30fe30dbc5fdaf3 md5sum: 2bbe9d7331b78bdac30fe30dbc5fdaf3
- name: allelecounter test_allelecounter_cram
command: nextflow run tests/modules/allelecounter -entry test_allelecounter_cram -c tests/config/nextflow.config
tags:
- allelecounter
files:
- path: output/allelecounter/test.alleleCount
md5sum: 2f83352a185168c7c98e9e42550b2856

View file

@ -8,28 +8,88 @@ workflow test_freebayes {
input = [ [ id:'test', single_end:false ], // meta map input = [ [ id:'test', single_end:false ], // meta map
file(params.test_data['sarscov2']['illumina']['test_paired_end_sorted_bam'], checkIfExists: true), file(params.test_data['sarscov2']['illumina']['test_paired_end_sorted_bam'], checkIfExists: true),
file(params.test_data['sarscov2']['illumina']['test_paired_end_sorted_bam_bai'], checkIfExists: true)] file(params.test_data['sarscov2']['illumina']['test_paired_end_sorted_bam_bai'], checkIfExists: true),
reference = [file(params.test_data['sarscov2']['genome']['genome_fasta'], checkIfExists: true), [],
file(params.test_data['sarscov2']['genome']['genome_fasta_fai'], checkIfExists: true)] []
]
fasta = file(params.test_data['sarscov2']['genome']['genome_fasta'], checkIfExists: true)
fai = file(params.test_data['sarscov2']['genome']['genome_fasta_fai'], checkIfExists: true)
targets = [] targets = []
samples = [] samples = []
populations = [] populations = []
cnv = [] cnv = []
FREEBAYES ( input, reference, targets, samples, populations, cnv) FREEBAYES (input, fasta, fai, targets, samples, populations, cnv)
} }
workflow test_freebayes_bed { workflow test_freebayes_bed {
input = [ [ id:'test', single_end:false ], // meta map input = [ [ id:'test', single_end:false ], // meta map
file(params.test_data['sarscov2']['illumina']['test_paired_end_sorted_bam'], checkIfExists: true), file(params.test_data['sarscov2']['illumina']['test_paired_end_sorted_bam'], checkIfExists: true),
file(params.test_data['sarscov2']['illumina']['test_paired_end_sorted_bam_bai'], checkIfExists: true)] file(params.test_data['sarscov2']['illumina']['test_paired_end_sorted_bam_bai'], checkIfExists: true),
reference = [file(params.test_data['sarscov2']['genome']['genome_fasta'], checkIfExists: true), [],
file(params.test_data['sarscov2']['genome']['genome_fasta_fai'], checkIfExists: true)] []
]
fasta = file(params.test_data['sarscov2']['genome']['genome_fasta'], checkIfExists: true)
fai = file(params.test_data['sarscov2']['genome']['genome_fasta_fai'], checkIfExists: true)
targets = file(params.test_data['sarscov2']['genome']['test_bed'], checkIfExists: true) targets = file(params.test_data['sarscov2']['genome']['test_bed'], checkIfExists: true)
samples = [] samples = []
populations = [] populations = []
cnv = [] cnv = []
FREEBAYES ( input, reference, targets, samples, populations, cnv) FREEBAYES (input, fasta, fai, targets, samples, populations, cnv)
}
workflow test_freebayes_cram {
input = [ [ id:'test', single_end:false ], // meta map
file(params.test_data['homo_sapiens']['illumina']['test_paired_end_sorted_cram'], checkIfExists: true),
file(params.test_data['homo_sapiens']['illumina']['test_paired_end_sorted_cram_crai'], checkIfExists: true),
[],
[]
]
fasta = file(params.test_data['homo_sapiens']['genome']['genome_fasta'], checkIfExists: true)
fai = file(params.test_data['homo_sapiens']['genome']['genome_fasta_fai'], checkIfExists: true)
targets = []
samples = []
populations = []
cnv = []
FREEBAYES (input, fasta, fai, targets, samples, populations, cnv)
}
workflow test_freebayes_somatic {
input = [ [ id:'test', single_end:false ], // meta map
file(params.test_data['homo_sapiens']['illumina']['test_paired_end_sorted_bam'], checkIfExists: true),
file(params.test_data['homo_sapiens']['illumina']['test_paired_end_sorted_bam_bai'], checkIfExists: true),
file(params.test_data['homo_sapiens']['illumina']['test2_paired_end_sorted_bam'], checkIfExists: true),
file(params.test_data['homo_sapiens']['illumina']['test2_paired_end_sorted_bam_bai'], checkIfExists: true)
]
fasta = file(params.test_data['homo_sapiens']['genome']['genome_fasta'], checkIfExists: true)
fai = file(params.test_data['homo_sapiens']['genome']['genome_fasta_fai'], checkIfExists: true)
targets = []
samples = []
populations = []
cnv = []
FREEBAYES (input, fasta, fai, targets, samples, populations, cnv)
}
workflow test_freebayes_somatic_cram_intervals {
input = [ [ id:'test', single_end:false ], // meta map
file(params.test_data['homo_sapiens']['illumina']['test_paired_end_sorted_cram'], checkIfExists: true),
file(params.test_data['homo_sapiens']['illumina']['test_paired_end_sorted_cram_crai'], checkIfExists: true),
file(params.test_data['homo_sapiens']['illumina']['test2_paired_end_sorted_cram'], checkIfExists: true),
file(params.test_data['homo_sapiens']['illumina']['test2_paired_end_sorted_cram_crai'], checkIfExists: true)
]
fasta = file(params.test_data['homo_sapiens']['genome']['genome_fasta'], checkIfExists: true)
fai = file(params.test_data['homo_sapiens']['genome']['genome_fasta_fai'], checkIfExists: true)
targets = file(params.test_data['homo_sapiens']['genome']['genome_bed'], checkIfExists: true)
samples = []
populations = []
cnv = []
FREEBAYES (input, fasta, fai, targets, samples, populations, cnv)
} }

View file

@ -12,3 +12,23 @@
files: files:
- path: output/freebayes/test.vcf.gz - path: output/freebayes/test.vcf.gz
- name: freebayes test_freebayes_cram
command: nextflow run tests/modules/freebayes -entry test_freebayes_cram -c tests/config/nextflow.config
tags:
- freebayes
files:
- path: output/freebayes/test.vcf.gz
- name: freebayes test_freebayes_somatic
command: nextflow run tests/modules/freebayes -entry test_freebayes_somatic -c tests/config/nextflow.config
tags:
- freebayes
files:
- path: output/freebayes/test.vcf.gz
- name: freebayes test_freebayes_somatic_cram_intervals
command: nextflow run tests/modules/freebayes -entry test_freebayes_somatic_cram_intervals -c tests/config/nextflow.config
tags:
- freebayes
files:
- path: output/freebayes/test.vcf.gz

View file

@ -2,67 +2,16 @@
nextflow.enable.dsl = 2 nextflow.enable.dsl = 2
include { UNTAR } from '../../../../modules/untar/main.nf' addParams( options: [:] )
include { GATK4_CREATESOMATICPANELOFNORMALS } from '../../../../modules/gatk4/createsomaticpanelofnormals/main.nf' addParams( options: [suffix:'.pon'] ) include { GATK4_CREATESOMATICPANELOFNORMALS } from '../../../../modules/gatk4/createsomaticpanelofnormals/main.nf' addParams( options: [suffix:'.pon'] )
workflow test_gatk4_createsomaticpanelofnormals { workflow test_gatk4_createsomaticpanelofnormals {
maindir = file('test_genomicsdb') db = file(params.test_data['homo_sapiens']['illumina']['test_genomicsdb_tar_gz'], checkIfExists: true)
subdir1 = file('test_genomicsdb/chr22$1$40001')
subdir2 = file('test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448')
subdir3 = file('test_genomicsdb/chr22$1$40001/genomicsdb_meta_dir')
subdir2.mkdirs()
subdir3.mkdirs()
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/__tiledb_workspace.tdb' , checkIfExists: true).copyTo(maindir) UNTAR ( db )
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/vcfheader.vcf' , checkIfExists: true).copyTo(maindir)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/vidmap.json' , checkIfExists: true).copyTo(maindir)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/callset.json' , checkIfExists: true).copyTo(maindir)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/.__consolidation_lock' , checkIfExists: true).copyTo(subdir1)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__array_schema.tdb' , checkIfExists: true).copyTo(subdir1)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/genomicsdb_meta_dir/genomicsdb_column_bounds.json' , checkIfExists: true).copyTo(subdir3)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/genomicsdb_meta_dir/genomicsdb_meta_2b25a6c2-cb94-4a4a-9005-acb7c595d322.json' , checkIfExists: true).copyTo(subdir3)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/AD.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/AD_var.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/ALT.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/ALT_var.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/BaseQRankSum.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/DB.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/DP.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/DP_FORMAT.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/END.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/ExcessHet.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/FILTER.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/FILTER_var.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/GQ.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/GT.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/GT_var.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/ID.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/ID_var.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/InbreedingCoeff.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/MIN_DP.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/MLEAC.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/MLEAC_var.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/MLEAF.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/MLEAF_var.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/MQRankSum.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/PGT.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/PGT_var.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/PID.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/PID_var.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/PL.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/PL_var.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/PS.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/QUAL.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/RAW_MQandDP.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/REF.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/REF_var.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/ReadPosRankSum.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/SB.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/__book_keeping.tdb.gz' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/__coords.tdb' , checkIfExists: true).copyTo(subdir2)
file( 'https://raw.githubusercontent.com/nf-core/test-datasets/modules/data/genomics/homo_sapiens/illumina/gatk/test_genomicsdb/chr22$1$40001/__3cf81648-433d-4464-be08-23d082445c9b139814474716928_1630588248448/__tiledb_fragment.tdb' , checkIfExists: true).copyTo(subdir2)
input = [ [ id:'test' ], // meta map input = Channel.of([ id:'test'])
file( maindir , checkIfExists: true)] .combine(UNTAR.out.untar)
fasta = file(params.test_data['homo_sapiens']['genome']['genome_fasta'], checkIfExists: true) fasta = file(params.test_data['homo_sapiens']['genome']['genome_fasta'], checkIfExists: true)
fastaidx = file(params.test_data['homo_sapiens']['genome']['genome_fasta_fai'], checkIfExists: true) fastaidx = file(params.test_data['homo_sapiens']['genome']['genome_fasta_fai'], checkIfExists: true)

View file

@ -0,0 +1,61 @@
#!/usr/bin/env nextflow
nextflow.enable.dsl = 2
include { UNTAR } from '../../../../modules/untar/main.nf' addParams( options: [:] )
include { GATK4_GENOMICSDBIMPORT } from '../../../../modules/gatk4/genomicsdbimport/main.nf' addParams( options: [:] )
workflow test_gatk4_genomicsdbimport_create_genomicsdb {
input = [ [ id:'test_genomicsdb'], // meta map
file(params.test_data['homo_sapiens']['illumina']['test_genome_vcf_gz'], checkIfExists: true) ,
file(params.test_data['homo_sapiens']['illumina']['test_genome_vcf_gz_tbi'], checkIfExists: true) ,
file(params.test_data['homo_sapiens']['genome']['genome_interval_list'], checkIfExists: true) ,
[] ,
[] ]
run_intlist = false
run_updatewspace = false
input_map = false
GATK4_GENOMICSDBIMPORT ( input, run_intlist, run_updatewspace, input_map )
}
workflow test_gatk4_genomicsdbimport_get_intervalslist {
db = file(params.test_data['homo_sapiens']['illumina']['test_genomicsdb_tar_gz'], checkIfExists: true)
UNTAR ( db )
def input = Channel.of([ [ id:'test_genomicsdb'], // meta map
[] ,
[] ,
[] ,
[] ])
.combine(UNTAR.out.untar)
run_intlist = true
run_updatewspace = false
input_map = false
GATK4_GENOMICSDBIMPORT ( input, run_intlist, run_updatewspace, input_map )
}
workflow test_gatk4_genomicsdbimport_update_genomicsdb {
db = file(params.test_data['homo_sapiens']['illumina']['test_genomicsdb_tar_gz'], checkIfExists: true)
UNTAR ( db )
def input = Channel.of([ [ id:'test_genomicsdb'], // meta map
file( params.test_data['homo_sapiens']['illumina']['test2_genome_vcf_gz'] , checkIfExists: true) ,
file( params.test_data['homo_sapiens']['illumina']['test2_genome_vcf_gz_tbi'] , checkIfExists: true) ,
[] ,
[] ])
.combine(UNTAR.out.untar)
run_intlist = false
run_updatewspace = true
input_map = false
GATK4_GENOMICSDBIMPORT ( input, run_intlist, run_updatewspace, input_map )
}

View file

@ -0,0 +1,50 @@
- name: gatk4 genomicsdbimport test_gatk4_genomicsdbimport_create_genomicsdb
command: nextflow run tests/modules/gatk4/genomicsdbimport -entry test_gatk4_genomicsdbimport_create_genomicsdb -c tests/config/nextflow.config
tags:
- gatk4/genomicsdbimport
- gatk4
files:
- path: output/gatk4/test_genomicsdb/__tiledb_workspace.tdb
md5sum: d41d8cd98f00b204e9800998ecf8427e
- path: output/gatk4/test_genomicsdb/callset.json
md5sum: a7d07d1c86449bbb1091ff29368da07a
- path: output/gatk4/test_genomicsdb/chr22$1$40001/.__consolidation_lock
md5sum: d41d8cd98f00b204e9800998ecf8427e
- path: output/gatk4/test_genomicsdb/chr22$1$40001/__array_schema.tdb
- path: output/gatk4/test_genomicsdb/chr22$1$40001/genomicsdb_meta_dir/genomicsdb_column_bounds.json
md5sum: 2502f79658bc000578ebcfddfc1194c0
- path: output/gatk4/test_genomicsdb/vcfheader.vcf
contains:
- "FORMAT=<ID=AD,Number=R,Type=Integer,Description="
- path: output/gatk4/test_genomicsdb/vidmap.json
md5sum: 18d3f68bd2cb6f4474990507ff95017a
- name: gatk4 genomicsdbimport test_gatk4_genomicsdbimport_get_intervalslist
command: nextflow run tests/modules/gatk4/genomicsdbimport -entry test_gatk4_genomicsdbimport_get_intervalslist -c tests/config/nextflow.config
tags:
- gatk4/genomicsdbimport
- gatk4
files:
- path: output/gatk4/test_genomicsdb.interval_list
md5sum: 4c85812ac15fc1cd29711a851d23c0bf
- name: gatk4 genomicsdbimport test_gatk4_genomicsdbimport_update_genomicsdb
command: nextflow run tests/modules/gatk4/genomicsdbimport -entry test_gatk4_genomicsdbimport_update_genomicsdb -c tests/config/nextflow.config
tags:
- gatk4/genomicsdbimport
- gatk4
files:
- path: output/gatk4/test_genomicsdb/__tiledb_workspace.tdb
md5sum: d41d8cd98f00b204e9800998ecf8427e
- path: output/gatk4/test_genomicsdb/callset.json
md5sum: 1ea31b59b9a218dd5681164aff4a5e07
- path: output/gatk4/test_genomicsdb/chr22$1$40001/.__consolidation_lock
md5sum: d41d8cd98f00b204e9800998ecf8427e
- path: output/gatk4/test_genomicsdb/chr22$1$40001/__array_schema.tdb
md5sum: 35e7fdd18149be80c0c3f56fa1f23971
- path: output/gatk4/test_genomicsdb/chr22$1$40001/genomicsdb_meta_dir/genomicsdb_column_bounds.json
md5sum: 2502f79658bc000578ebcfddfc1194c0
- path: output/gatk4/test_genomicsdb/vcfheader.vcf
md5sum: 47a615385a49f9261e088104b903bb9b
- path: output/gatk4/test_genomicsdb/vidmap.json
md5sum: 18d3f68bd2cb6f4474990507ff95017a

View file

@ -2,10 +2,10 @@
nextflow.enable.dsl = 2 nextflow.enable.dsl = 2
include { GENRICH } from '../../../modules/genrich/main.nf' addParams( control_bam: false, pvalues: false, pileup:false, bed:false, blacklist_bed:false, save_duplicates:false, options: ["args": "-p 0.1"] ) include { GENRICH } from '../../../modules/genrich/main.nf' addParams( options: ["args": "-p 0.1"] )
include { GENRICH as GENRICH_BLACKLIST } from '../../../modules/genrich/main.nf' addParams( control_bam: false, pvalues: false, pileup:false, bed:false, blacklist_bed:true, save_duplicates:false, options: ["args": "-p 0.1"] ) include { GENRICH as GENRICH_CTRL } from '../../../modules/genrich/main.nf' addParams( options: ["args": "-p 0.9"] )
include { GENRICH as GENRICH_ALL_OUTPUTS } from '../../../modules/genrich/main.nf' addParams( control_bam: false, pvalues: true, pileup:true, bed:true, blacklist_bed:false, save_duplicates:true, options: ["args": "-r -p 0.1"] ) include { GENRICH as GENRICH_ALL } from '../../../modules/genrich/main.nf' addParams( options: ["args": "-r -p 0.1"] )
include { GENRICH as GENRICH_ATACSEQ } from '../../../modules/genrich/main.nf' addParams( control_bam: false, pvalues: false, pileup:false, bed:false, blacklist_bed:false, save_duplicates:false, options: ["args": "-j -p 0.1"] ) include { GENRICH as GENRICH_ATACSEQ } from '../../../modules/genrich/main.nf' addParams( options: ["args": "-j -p 0.1"] )
workflow test_genrich { workflow test_genrich {
input = [ [ id:'test', single_end:false ], // meta map input = [ [ id:'test', single_end:false ], // meta map
@ -13,7 +13,12 @@ workflow test_genrich {
control = [ ] control = [ ]
blacklist = [ ] blacklist = [ ]
GENRICH ( input, control, blacklist ) save_pvalues = false
save_pileup = false
save_bed = false
save_duplicates = false
GENRICH ( input, control, blacklist, save_pvalues, save_pileup, save_bed, save_duplicates )
} }
workflow test_genrich_ctrl { workflow test_genrich_ctrl {
@ -22,7 +27,12 @@ workflow test_genrich_ctrl {
control = [ file( params.test_data['homo_sapiens']['illumina']['test2_paired_end_name_sorted_bam'], checkIfExists: true) ] control = [ file( params.test_data['homo_sapiens']['illumina']['test2_paired_end_name_sorted_bam'], checkIfExists: true) ]
blacklist = [ ] blacklist = [ ]
GENRICH ( input, control, blacklist ) save_pvalues = false
save_pileup = false
save_bed = false
save_duplicates = false
GENRICH_CTRL ( input, control, blacklist, save_pvalues, save_pileup, save_bed, save_duplicates )
} }
workflow test_genrich_all_outputs { workflow test_genrich_all_outputs {
@ -31,14 +41,39 @@ workflow test_genrich_all_outputs {
control = [ file( params.test_data['homo_sapiens']['illumina']['test2_paired_end_name_sorted_bam'], checkIfExists: true) ] control = [ file( params.test_data['homo_sapiens']['illumina']['test2_paired_end_name_sorted_bam'], checkIfExists: true) ]
blacklist = [ ] blacklist = [ ]
GENRICH_ALL_OUTPUTS ( input, control, blacklist ) save_pvalues = true
save_pileup = true
save_bed = true
save_duplicates = true
GENRICH_ALL ( input, control, blacklist, save_pvalues, save_pileup, save_bed, save_duplicates )
}
workflow test_genrich_blacklist {
input = [ [ id:'test', single_end:false ], // meta map
[ file( params.test_data['homo_sapiens']['illumina']['test_paired_end_name_sorted_bam'], checkIfExists: true) ]]
control = [ ]
blacklist = [ file(params.test_data['sarscov2']['genome']['test_bed'], checkIfExists: true)]
save_pvalues = false
save_pileup = false
save_bed = false
save_duplicates = false
GENRICH ( input, control, blacklist, save_pvalues, save_pileup, save_bed, save_duplicates )
} }
workflow test_genrich_atacseq { workflow test_genrich_atacseq {
input = [ [ id:'test', single_end:false ], // meta map input = [ [ id:'test', single_end:false ], // meta map
[ file( params.test_data['homo_sapiens']['illumina']['test_paired_end_name_sorted_bam'], checkIfExists: true) ]] [ file( params.test_data['homo_sapiens']['illumina']['test_paired_end_name_sorted_bam'], checkIfExists: true) ]]
control = [ file( params.test_data['homo_sapiens']['illumina']['test2_paired_end_name_sorted_bam'], checkIfExists: true) ] control = [ ]
blacklist = [ ] blacklist = [ ]
GENRICH_ATACSEQ ( input, control, blacklist ) save_pvalues = false
save_pileup = false
save_bed = false
save_duplicates = false
GENRICH_ATACSEQ ( input, control, blacklist, save_pvalues, save_pileup, save_bed, save_duplicates )
} }

View file

@ -12,7 +12,7 @@
- genrich - genrich
files: files:
- path: output/genrich/test.narrowPeak - path: output/genrich/test.narrowPeak
md5sum: 6afabdd3f691c7c84c66ff8a23984681 md5sum: 2fcc392360b317f5ebee88cdbc149e05
- name: genrich test_genrich_all_outputs - name: genrich test_genrich_all_outputs
command: nextflow run tests/modules/genrich -entry test_genrich_all_outputs -c tests/config/nextflow.config command: nextflow run tests/modules/genrich -entry test_genrich_all_outputs -c tests/config/nextflow.config
@ -20,15 +20,23 @@
- genrich - genrich
files: files:
- path: output/genrich/test.duplicates.txt - path: output/genrich/test.duplicates.txt
md5sum: a92893f905fd8b3751bc6a960fbfe7ba md5sum: 159d557af7c23bc3cfb802d87fa96c34
- path: output/genrich/test.intervals.bed - path: output/genrich/test.intervals.bed
md5sum: 52edf47e6641c0cc03f9cca7324f7eaa md5sum: 4bea65caa3f4043d703af4b57161112e
- path: output/genrich/test.narrowPeak - path: output/genrich/test.narrowPeak
md5sum: e45eb7d000387975050c2e85c164e5be md5sum: d41d8cd98f00b204e9800998ecf8427e
- path: output/genrich/test.pileup.bedGraph - path: output/genrich/test.pileup.bedGraph
md5sum: e4f7fa664cd4ed2cf3a1a3a9eb415e71 md5sum: 03e53848de695b5794f32f15b2709203
- path: output/genrich/test.pvalues.bedGraph - path: output/genrich/test.pvalues.bedGraph
md5sum: 564859953704983393d4b7d6317060cd md5sum: b14feef34b6d2379a173a734ca963cde
- name: genrich test_genrich_blacklist
command: nextflow run tests/modules/genrich -entry test_genrich_blacklist -c tests/config/nextflow.config
tags:
- genrich
files:
- path: output/genrich/test.narrowPeak
md5sum: 6afabdd3f691c7c84c66ff8a23984681
- name: genrich test_genrich_atacseq - name: genrich test_genrich_atacseq
command: nextflow run tests/modules/genrich -entry test_genrich_atacseq -c tests/config/nextflow.config command: nextflow run tests/modules/genrich -entry test_genrich_atacseq -c tests/config/nextflow.config

View file

@ -0,0 +1,12 @@
#!/usr/bin/env nextflow
nextflow.enable.dsl = 2
include { GUNC_DOWNLOADDB } from '../../../../modules/gunc/downloaddb/main.nf' addParams( options: [:] )
workflow test_gunc_downloaddb {
input = 'progenomes'
GUNC_DOWNLOADDB ( input )
}

View file

@ -0,0 +1,8 @@
- name: gunc downloaddb
command: nextflow run ./tests/modules/gunc/downloaddb -entry test_gunc_downloaddb -c tests/config/nextflow.config
tags:
- gunc
- gunc/downloaddb
files:
- path: output/gunc/gunc_db_progenomes2.1.dmnd
md5sum: 447c9330056b02f29f30fe81fe4af4eb

View file

@ -0,0 +1,17 @@
#!/usr/bin/env nextflow
nextflow.enable.dsl = 2
include { GUNC_RUN } from '../../../../modules/gunc/run/main.nf' addParams( options: [:] )
include { GUNC_DOWNLOADDB } from '../../../../modules/gunc/downloaddb/main.nf' addParams( options: [:] )
workflow test_gunc_run {
input = [ [ id:'test', single_end:false ], // meta map
file(params.test_data['sarscov2']['illumina']['contigs_fasta'], checkIfExists: true) ]
GUNC_DOWNLOADDB('progenomes')
GUNC_RUN ( input, GUNC_DOWNLOADDB.out.db )
}

View file

@ -0,0 +1,8 @@
- name: gunc run
command: nextflow run ./tests/modules/gunc/run -entry test_gunc_run -c tests/config/nextflow.config
tags:
- gunc
- gunc/run
files:
- path: output/gunc/GUNC.progenomes_2.1.maxCSS_level.tsv
md5sum: 0420c1a9f2c50fefaee9fab5d80a551a

View file

@ -0,0 +1,11 @@
#!/usr/bin/env nextflow
nextflow.enable.dsl = 2
include { HMMCOPY_GCCOUNTER } from '../../../../modules/hmmcopy/gccounter/main.nf' addParams( options: [:] )
workflow test_hmmcopy_gccounter {
fasta = file(params.test_data['sarscov2']['genome']['genome_fasta'], checkIfExists: true)
HMMCOPY_GCCOUNTER (fasta)
}

View file

@ -0,0 +1,8 @@
- name: hmmcopy gccounter test_hmmcopy_gccounter
command: nextflow run tests/modules/hmmcopy/gccounter -entry test_hmmcopy_gccounter -c tests/config/nextflow.config
tags:
- hmmcopy
- hmmcopy/gccounter
files:
- path: output/hmmcopy/genome.gc.wig
md5sum: 59ad14bc5aaa903187d7b248c9490deb

View file

@ -0,0 +1,14 @@
#!/usr/bin/env nextflow
nextflow.enable.dsl = 2
include { HMMCOPY_READCOUNTER } from '../../../../modules/hmmcopy/readcounter/main.nf' addParams( options: [:] )
workflow test_hmmcopy_readcounter {
input = [ [ id:'test'], // meta map
[ file(params.test_data['homo_sapiens']['illumina']['test_paired_end_recalibrated_sorted_bam'], checkIfExists: true)],
[ file(params.test_data['homo_sapiens']['illumina']['test_paired_end_recalibrated_sorted_bam_bai'], checkIfExists: true)]
]
HMMCOPY_READCOUNTER ( input )
}

View file

@ -0,0 +1,8 @@
- name: hmmcopy readcounter test_hmmcopy_readcounter
command: nextflow run tests/modules/hmmcopy/readcounter -entry test_hmmcopy_readcounter -c tests/config/nextflow.config
tags:
- hmmcopy
- hmmcopy/readcounter
files:
- path: output/hmmcopy/test.wig
md5sum: 3655d8325baea81b3b690791262c6b57

35
tests/modules/idr/main.nf Normal file
View file

@ -0,0 +1,35 @@
#!/usr/bin/env nextflow
nextflow.enable.dsl = 2
include { IDR } from '../../../modules/idr/main.nf' addParams( options: [:] )
workflow test_idr_narrowpeak {
input = [
file(params.test_data['homo_sapiens']['illumina']['test_narrowpeak'], checkIfExists: true),
file(params.test_data['homo_sapiens']['illumina']['test2_narrowpeak'], checkIfExists: true)
]
IDR ( input, 'narrowPeak', 'test' )
}
workflow test_idr_broadpeak {
input = [
file(params.test_data['homo_sapiens']['illumina']['test_broadpeak'], checkIfExists: true),
file(params.test_data['homo_sapiens']['illumina']['test2_broadpeak'], checkIfExists: true)
]
IDR ( input, 'broadPeak', 'test' )
}
workflow test_idr_noprefix {
input = [
file(params.test_data['homo_sapiens']['illumina']['test_narrowpeak'], checkIfExists: true),
file(params.test_data['homo_sapiens']['illumina']['test2_narrowpeak'], checkIfExists: true)
]
IDR ( input, 'narrowPeak', '' )
}

View file

@ -0,0 +1,35 @@
- name: idr test_idr_narrowpeak
command: nextflow run tests/modules/idr -entry test_idr_narrowpeak -c tests/config/nextflow.config
tags:
- idr
files:
- path: output/idr/test.idrValues.txt
md5sum: 09be837cc6abbc3eb5958b74802eea55
- path: output/idr/test.idrValues.txt.png
md5sum: 4a7143ccc0ccadb37c2317bf626e6d96
- path: output/idr/test.log.txt
md5sum: 6443507ac66b9d3b64bc56b78328083e
- name: idr test_idr_broadpeak
command: nextflow run tests/modules/idr -entry test_idr_broadpeak -c tests/config/nextflow.config
tags:
- idr
files:
- path: output/idr/test.idrValues.txt
md5sum: 387441c716815e4caec3e70a2cc11a4a
- path: output/idr/test.idrValues.txt.png
md5sum: 7204083ca5b920b4215a5991c12cb4e7
- path: output/idr/test.log.txt
md5sum: e6917133112b5cec135c182ffac19237
- name: idr test_idr_noprefix
command: nextflow run tests/modules/idr -entry test_idr_noprefix -c tests/config/nextflow.config
tags:
- idr
files:
- path: output/idr/idrValues.txt
md5sum: 09be837cc6abbc3eb5958b74802eea55
- path: output/idr/idrValues.txt.png
md5sum: 4a7143ccc0ccadb37c2317bf626e6d96
- path: output/idr/log.txt
md5sum: 6443507ac66b9d3b64bc56b78328083e

View file

@ -0,0 +1,14 @@
#!/usr/bin/env nextflow
nextflow.enable.dsl = 2
include { MTNUCRATIO } from '../../../modules/mtnucratio/main.nf' addParams( options: [:] )
workflow test_mtnucratio {
input = [ [ id:'test', single_end:false ], // meta map
file(params.test_data['sarscov2']['illumina']['test_single_end_sorted_bam'], checkIfExists: true)]
mt_id = 'mt_id'
MTNUCRATIO ( input, mt_id )
}

View file

@ -0,0 +1,9 @@
- name: mtnucratio
command: nextflow run tests/modules/mtnucratio -entry test_mtnucratio -c tests/config/nextflow.config
tags:
- mtnucratio
files:
- path: output/mtnucratio/test.single_end.sorted.bam.mtnucratio
md5sum: 19e96849802c70aa0694785f716274b7
- path: output/mtnucratio/test.single_end.sorted.bam.mtnucratiomtnuc.json
md5sum: 14d24be6272854d6762f0dfad5918ef6

View file

@ -0,0 +1,15 @@
#!/usr/bin/env nextflow
nextflow.enable.dsl = 2
include { RACON } from '../../../modules/racon/main.nf' addParams( options: [:] )
workflow test_racon {
input = [ [ id:'test', single_end:true ], // meta map
file(params.test_data['bacteroides_fragilis']['nanopore']['test_fastq_gz'], checkIfExists: true),
file(params.test_data['bacteroides_fragilis']['genome']['genome_fna_gz'], checkIfExists: true),
file(params.test_data['bacteroides_fragilis']['genome']['genome_paf'], checkIfExists: true)
]
RACON ( input )
}

View file

@ -0,0 +1,7 @@
- name: racon test_racon
command: nextflow run tests/modules/racon -entry test_racon -c tests/config/nextflow.config
tags:
- racon
files:
- path: output/racon/test_assembly_consensus.fasta.gz
md5sum: 96a0ba94c6154f6f37b5a76a0207eb6f

View file

@ -0,0 +1,14 @@
#!/usr/bin/env nextflow
nextflow.enable.dsl = 2
include { SAMTOOLS_FIXMATE } from '../../../../modules/samtools/fixmate/main.nf' addParams( options: [args:'-r -c -m'] )
workflow test_samtools_fixmate {
input = [ [ id:'test', single_end:false ], // meta map
file(params.test_data['sarscov2']['illumina']['test_paired_end_bam'], checkIfExists: true) ]
SAMTOOLS_FIXMATE ( input )
}

View file

@ -0,0 +1,8 @@
- name: samtools fixmate test_samtools_fixmate
command: nextflow run tests/modules/samtools/fixmate -entry test_samtools_fixmate -c tests/config/nextflow.config
tags:
- samtools/fixmate
- samtools
files:
- path: output/samtools/test.bam
md5sum: 92c8463710cdcaef2010aa02ed9e01fd

View file

@ -0,0 +1,13 @@
#!/usr/bin/env nextflow
nextflow.enable.dsl = 2
include { SEQSERO2 } from '../../../modules/seqsero2/main.nf' addParams( options: [args: '-m k -t 4'] )
workflow test_seqsero2 {
input = [ [ id:'test', single_end:false ], // meta map
file(params.test_data['sarscov2']['genome']['genome_fasta'], checkIfExists: true) ]
SEQSERO2 ( input )
}

View file

@ -0,0 +1,11 @@
- name: seqsero2 test_seqsero2
command: nextflow run tests/modules/seqsero2 -entry test_seqsero2 -c tests/config/nextflow.config
tags:
- seqsero2
files:
- path: output/seqsero2/results/SeqSero_log.txt
md5sum: d00242dfa734b5abb3622a6048f0b4fb
- path: output/seqsero2/results/SeqSero_result.tsv
contains: ['Sample', 'Predicted', 'Note']
- path: output/seqsero2/results/SeqSero_result.txt
contains: ['Sample', 'Predicted', 'Note']