1
0
Fork 0
mirror of https://github.com/MillironX/taxprofiler.git synced 2024-11-26 00:59:55 +00:00

Merge pull request #130 from nf-core/krakentools/combinekreports

Adds KRAKEN_COMBINEKREPORTS for 'native' multi-sample OTU
This commit is contained in:
James A. Fellows Yates 2022-09-15 13:00:07 +02:00 committed by GitHub
commit d9795a4fdc
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
7 changed files with 163 additions and 13 deletions

View file

@ -294,6 +294,15 @@ process {
]
}
withName: KRAKENTOOLS_COMBINEKREPORTS {
ext.prefix = { "kraken2_${meta.id}_combined_reports" }
publishDir = [
path: { "${params.outdir}/kraken2/" },
mode: params.publish_dir_mode,
pattern: '*.{txt}'
]
}
withName: KRONA_CLEANUP {
ext.prefix = params.perform_runmerging ? { "${meta.id}-${meta.db_name}" } : { "${meta.id}-${meta.run_accession}-${meta.db_name}" }
publishDir = [
@ -358,6 +367,15 @@ process {
]
}
withName: KRAKENTOOLS_COMBINEKREPORTS_CENTRIFUGE {
ext.prefix = { "centrifuge_${meta.id}_combined_reports" }
publishDir = [
path: { "${params.outdir}/centrifuge/" },
mode: params.publish_dir_mode,
pattern: '*.{txt}'
]
}
withName: KAIJU_KAIJU {
ext.prefix = params.perform_runmerging ? { "${meta.id}-${meta.db_name}" } : { "${meta.id}-${meta.run_accession}-${meta.db_name}" }
publishDir = [
@ -369,7 +387,7 @@ process {
}
withName: KAIJU_KAIJU2TABLE {
ext.prefix = { "${meta.id}_combined_reports" }
ext.prefix = { "kaiju_${meta.id}_combined_reports" }
publishDir = [
path: { "${params.outdir}/kaiju/" },
mode: params.publish_dir_mode,

View file

@ -73,6 +73,10 @@
"branch": "master",
"git_sha": "409a308ba46284d8ebb48c2c1befd6f6433db3f7"
},
"krakentools/combinekreports": {
"branch": "master",
"git_sha": "ee0346b4d14ffdc15ce7e093ca1363cd07c9bd78"
},
"krakentools/kreport2krona": {
"branch": "master",
"git_sha": "233fa70811a03a4cecb2ece483b5c8396e2cee1d"
@ -135,7 +139,7 @@
},
"untar": {
"branch": "master",
"git_sha": "5e7b1ef9a5a2d9258635bcbf70fcf37dacd1b247"
"git_sha": "393dbd6ddafe3f18eac02893dd4a21e4d45de679"
}
}
}

View file

@ -0,0 +1,34 @@
process KRAKENTOOLS_COMBINEKREPORTS {
label 'process_low'
conda (params.enable_conda ? "bioconda::krakentools=1.2" : null)
container "${ workflow.containerEngine == 'singularity' && !task.ext.singularity_pull_docker_container ?
'https://depot.galaxyproject.org/singularity/krakentools:1.2--pyh5e36f6f_0':
'quay.io/biocontainers/krakentools:1.2--pyh5e36f6f_0' }"
input:
tuple val(meta), path(kreports)
output:
tuple val(meta), path("*.txt"), emit: txt
path "versions.yml", emit: versions
when:
task.ext.when == null || task.ext.when
script:
def args = task.ext.args ?: ''
prefix = task.ext.prefix ?: "${meta.id}"
def VERSION = '1.2' // WARN: Version information not provided by tool on CLI. Please update this string when bumping container versions.
"""
combine_kreports.py \\
-r ${kreports} \\
-o ${prefix}.txt \\
${args}
cat <<-END_VERSIONS > versions.yml
"${task.process}":
combine_kreports.py: ${VERSION}
END_VERSIONS
"""
}

View file

@ -0,0 +1,43 @@
name: krakentools_combinekreports
description: Takes a Kraken report file and prints out a krona-compatible TEXT file
keywords:
- kraken
- krakentools
- metagenomics
- table
- combining
- merging
tools:
- krakentools:
description: KrakenTools is a suite of scripts to be used for post-analysis of Kraken/KrakenUniq/Kraken2/Bracken results. Please cite the relevant paper if using KrakenTools with any of the listed programs.
homepage: https://github.com/jenniferlu717/KrakenTools
licence: ["GPL v3"]
input:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- kreports:
type: file
description: List of kraken-style report files
pattern: "*.{txt,kreport}"
output:
- meta:
type: map
description: |
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- versions:
type: file
description: File containing software versions
pattern: "versions.yml"
- txt:
type: file
description: Combined kreport file of all input files
pattern: "*.txt"
authors:
- "@jfy133"

View file

@ -25,12 +25,23 @@ process UNTAR {
"""
mkdir output
## Ensures --strip-components only applied when top level of tar contents is a directory
## If just files or multiple directories, place all in output
if [[ \$(tar -tzf ${archive} | grep "/\$" | wc -l) -eq 1 ]]; then
tar \\
-C output --strip-components 1 \\
-xzvf \\
$args \\
$archive \\
$args2
else
tar \\
-C output \\
-xzvf \\
$args \\
$archive \\
$args2
fi
mv output ${untar}

View file

@ -26,9 +26,9 @@ output:
Groovy Map containing sample information
e.g. [ id:'test', single_end:false ]
- untar:
type: file
description:
pattern: "*.*"
type: directory
description: Directory containing contents of archive
pattern: "*/"
- versions:
type: file
description: File containing software versions
@ -36,3 +36,5 @@ output:
authors:
- "@joseespinosa"
- "@drpatelh"
- "@matthdsm"
- "@jfy133"

View file

@ -2,7 +2,9 @@
// Standardise output files e.g. aggregation
//
include { KRAKENTOOLS_COMBINEKREPORTS as KRAKENTOOLS_COMBINEKREPORTS_CENTRIFUGE } from '../../modules/nf-core/modules/krakentools/combinekreports/main'
include { KAIJU_KAIJU2TABLE } from '../../modules/nf-core/modules/kaiju/kaiju2table/main'
include { KRAKENTOOLS_COMBINEKREPORTS } from '../../modules/nf-core/modules/krakentools/combinekreports/main'
include { MOTUS_MERGE } from '../../modules/nf-core/modules/motus/merge/main'
workflow STANDARDISATION_PROFILES {
@ -23,6 +25,8 @@ workflow STANDARDISATION_PROFILES {
ch_input_profiles = profiles
.branch {
motus: it[0]['tool'] == 'motus'
kraken2: it[0]['tool'] == 'kraken2'
centrifuge: it[0]['tool'] == 'centrifuge'
unknown: true
}
@ -43,6 +47,23 @@ workflow STANDARDISATION_PROFILES {
Standardise and aggregate
*/
// CENTRIFUGE
// Collect and replace id for db_name for prefix
// Have to sort by size to ensure first file actually has hits otherwise
// the script fails
ch_profiles_for_centrifuge = ch_input_profiles.centrifuge
.map { [it[0]['db_name'], it[1]] }
.groupTuple(sort: {-it.size()} )
.map {
[[id:it[0]], it[1]]
}
KRAKENTOOLS_COMBINEKREPORTS_CENTRIFUGE ( ch_profiles_for_centrifuge )
ch_standardised_tables = ch_standardised_tables.mix( KRAKENTOOLS_COMBINEKREPORTS_CENTRIFUGE.out.txt )
ch_multiqc_files = ch_multiqc_files.mix( KRAKENTOOLS_COMBINEKREPORTS_CENTRIFUGE.out.txt )
ch_versions = ch_versions.mix( KRAKENTOOLS_COMBINEKREPORTS_CENTRIFUGE.out.versions )
// Kaiju
// Collect and replace id for db_name for prefix
@ -58,6 +79,23 @@ workflow STANDARDISATION_PROFILES {
ch_multiqc_files = ch_multiqc_files.mix( KAIJU_KAIJU2TABLE.out.summary )
ch_versions = ch_versions.mix( KAIJU_KAIJU2TABLE.out.versions )
// Kraken2
// Collect and replace id for db_name for prefix
// Have to sort by size to ensure first file actually has hits otherwise
// the script fails
ch_profiles_for_kraken2 = ch_input_profiles.kraken2
.map { [it[0]['db_name'], it[1]] }
.groupTuple(sort: {-it.size()} )
.map {
[[id:it[0]], it[1]]
}
KRAKENTOOLS_COMBINEKREPORTS ( ch_profiles_for_kraken2 )
ch_standardised_tables = ch_standardised_tables.mix( KRAKENTOOLS_COMBINEKREPORTS.out.txt )
ch_multiqc_files = ch_multiqc_files.mix( KRAKENTOOLS_COMBINEKREPORTS.out.txt )
ch_versions = ch_versions.mix( KRAKENTOOLS_COMBINEKREPORTS.out.versions )
// mOTUs has a 'single' database, and cannot create custom ones.
// Therefore removing db info here, and publish merged at root mOTUs results
// directory