mirror of
https://github.com/MillironX/taxprofiler.git
synced 2024-11-22 05:39:54 +00:00
Fix cardinality issue
This commit is contained in:
parent
6459d8920f
commit
21a76c2449
3 changed files with 18 additions and 49 deletions
|
@ -120,6 +120,7 @@ params {
|
||||||
|
|
||||||
//krakenuniq
|
//krakenuniq
|
||||||
run_krakenuniq = false
|
run_krakenuniq = false
|
||||||
|
krakenuniq_ram_run_size = '16G'
|
||||||
krakenuniq_save_reads = false // added directly to module in profiling.nf
|
krakenuniq_save_reads = false // added directly to module in profiling.nf
|
||||||
krakenuniq_save_readclassifications = false // added directly to module in profiling.nf
|
krakenuniq_save_readclassifications = false // added directly to module in profiling.nf
|
||||||
|
|
||||||
|
|
|
@ -10,11 +10,7 @@
|
||||||
"type": "object",
|
"type": "object",
|
||||||
"fa_icon": "fas fa-terminal",
|
"fa_icon": "fas fa-terminal",
|
||||||
"description": "Define where the pipeline should find input data and save output data.",
|
"description": "Define where the pipeline should find input data and save output data.",
|
||||||
"required": [
|
"required": ["input", "databases", "outdir"],
|
||||||
"input",
|
|
||||||
"databases",
|
|
||||||
"outdir"
|
|
||||||
],
|
|
||||||
"properties": {
|
"properties": {
|
||||||
"input": {
|
"input": {
|
||||||
"type": "string",
|
"type": "string",
|
||||||
|
@ -84,10 +80,7 @@
|
||||||
"shortread_qc_tool": {
|
"shortread_qc_tool": {
|
||||||
"type": "string",
|
"type": "string",
|
||||||
"default": "fastp",
|
"default": "fastp",
|
||||||
"enum": [
|
"enum": ["fastp", "adapterremoval"],
|
||||||
"fastp",
|
|
||||||
"adapterremoval"
|
|
||||||
],
|
|
||||||
"fa_icon": "fas fa-tools",
|
"fa_icon": "fas fa-tools",
|
||||||
"description": "Specify which tool to use for short-read QC"
|
"description": "Specify which tool to use for short-read QC"
|
||||||
},
|
},
|
||||||
|
@ -140,11 +133,7 @@
|
||||||
"shortread_complexityfilter_tool": {
|
"shortread_complexityfilter_tool": {
|
||||||
"type": "string",
|
"type": "string",
|
||||||
"default": "bbduk",
|
"default": "bbduk",
|
||||||
"enum": [
|
"enum": ["bbduk", "prinseqplusplus", "fastp"],
|
||||||
"bbduk",
|
|
||||||
"prinseqplusplus",
|
|
||||||
"fastp"
|
|
||||||
],
|
|
||||||
"fa_icon": "fas fa-hammer",
|
"fa_icon": "fas fa-hammer",
|
||||||
"description": "Specify which tool to use for complexity filtering"
|
"description": "Specify which tool to use for complexity filtering"
|
||||||
},
|
},
|
||||||
|
@ -178,10 +167,7 @@
|
||||||
"shortread_complexityfilter_prinseqplusplus_mode": {
|
"shortread_complexityfilter_prinseqplusplus_mode": {
|
||||||
"type": "string",
|
"type": "string",
|
||||||
"default": "entropy",
|
"default": "entropy",
|
||||||
"enum": [
|
"enum": ["entropy", "dust"],
|
||||||
"entropy",
|
|
||||||
"dust"
|
|
||||||
],
|
|
||||||
"fa_icon": "fas fa-check-square",
|
"fa_icon": "fas fa-check-square",
|
||||||
"description": "Specify the complexity filter mode for PRINSEQ++"
|
"description": "Specify the complexity filter mode for PRINSEQ++"
|
||||||
},
|
},
|
||||||
|
@ -355,15 +341,7 @@
|
||||||
"diamond_output_format": {
|
"diamond_output_format": {
|
||||||
"type": "string",
|
"type": "string",
|
||||||
"default": "tsv",
|
"default": "tsv",
|
||||||
"enum": [
|
"enum": ["blast", "xml", "txt", "daa", "sam", "tsv", "paf"],
|
||||||
"blast",
|
|
||||||
"xml",
|
|
||||||
"txt",
|
|
||||||
"daa",
|
|
||||||
"sam",
|
|
||||||
"tsv",
|
|
||||||
"paf"
|
|
||||||
],
|
|
||||||
"fa_icon": "fas fa-file",
|
"fa_icon": "fas fa-file",
|
||||||
"description": "Specify output format from DIAMOND profiling.",
|
"description": "Specify output format from DIAMOND profiling.",
|
||||||
"help_text": "DIAMOND can produce output in a number of different formats, you can specify here which to produce.\n\nNote that DIAMOND can only produce one format at a time, and depending on which you pick, some downstream steps may not be executed. For example, selecting `daa` or `sam` will mean you will not get a tabular taxonomic profile as with the other tools.\n\nWill be overriden by `--diamond_save_reads.`\n\n> Modifies tool parameter(s):\n> - diamond blastx: `--outfmt`"
|
"help_text": "DIAMOND can produce output in a number of different formats, you can specify here which to produce.\n\nNote that DIAMOND can only produce one format at a time, and depending on which you pick, some downstream steps may not be executed. For example, selecting `daa` or `sam` will mean you will not get a tabular taxonomic profile as with the other tools.\n\nWill be overriden by `--diamond_save_reads.`\n\n> Modifies tool parameter(s):\n> - diamond blastx: `--outfmt`"
|
||||||
|
@ -382,14 +360,7 @@
|
||||||
"kaiju_taxon_rank": {
|
"kaiju_taxon_rank": {
|
||||||
"type": "string",
|
"type": "string",
|
||||||
"default": "species",
|
"default": "species",
|
||||||
"enum": [
|
"enum": ["phylum", "class", "order", "family", "genus", "species"],
|
||||||
"phylum",
|
|
||||||
"class",
|
|
||||||
"order",
|
|
||||||
"family",
|
|
||||||
"genus",
|
|
||||||
"species"
|
|
||||||
],
|
|
||||||
"fa_icon": "fas fa-tag",
|
"fa_icon": "fas fa-tag",
|
||||||
"description": "Specify taxonomic rank to be displayed in Kaiju taxon table",
|
"description": "Specify taxonomic rank to be displayed in Kaiju taxon table",
|
||||||
"help_text": "Specify the taxonomic level(s) to be displayed in the resulting Kaiju taxon table, as generated by the kaiju2table helper tool.\n\nThis can be either a single level (e.g. `species`), or a comma separated list to display the full taxonomic path (e.g. `superkingdom,phylum,class,order,family,genus,species.`).\n\n> Modifies tool parameter(s):\n> - kaiju2table: `-l`"
|
"help_text": "Specify the taxonomic level(s) to be displayed in the resulting Kaiju taxon table, as generated by the kaiju2table helper tool.\n\nThis can be either a single level (e.g. `species`), or a comma separated list to display the full taxonomic path (e.g. `superkingdom,phylum,class,order,family,genus,species.`).\n\n> Modifies tool parameter(s):\n> - kaiju2table: `-l`"
|
||||||
|
@ -428,6 +399,13 @@
|
||||||
"description": "Turn on saving of KrakenUniq-aligned reads",
|
"description": "Turn on saving of KrakenUniq-aligned reads",
|
||||||
"help_text": "Save reads that do and do not have a taxonomic classification in your output results directory in FASTQ format.\n\n> Modifies tool parameter(s):\n> - krakenuniq: `--classified-out` and `--unclassified-out`"
|
"help_text": "Save reads that do and do not have a taxonomic classification in your output results directory in FASTQ format.\n\n> Modifies tool parameter(s):\n> - krakenuniq: `--classified-out` and `--unclassified-out`"
|
||||||
},
|
},
|
||||||
|
"krakenuniq_ram_run_size": {
|
||||||
|
"type": "string",
|
||||||
|
"default": "16G",
|
||||||
|
"description": "Specify how large to chunk database when loading into memory for KrakenUniq",
|
||||||
|
"fa_icon": "fas fa-database",
|
||||||
|
"help_text": "nf-core/taxprofiler utilises a 'low memory' option for KrakenUniq that can reduce the amount of RAM the process requires using the `--preloaded` option.\n\nA further extension to this option is that you can specify how large each chunk of the database should be that gets loaded into memory at any one time. You can specify the amount of RAM to chunk the database to with this parameter, and is particularly useful for people with limited computational resources.\n\nMore information about this parameter can be seen [here](https://github.com/fbreitwieser/krakenuniq/blob/master/README.md#new-release-v07).\n\n> Modifies KrakenUniq parameter: --preload\n\n> \n\n"
|
||||||
|
},
|
||||||
"krakenuniq_save_readclassifications": {
|
"krakenuniq_save_readclassifications": {
|
||||||
"type": "boolean",
|
"type": "boolean",
|
||||||
"fa_icon": "fas fa-save",
|
"fa_icon": "fas fa-save",
|
||||||
|
@ -612,14 +590,7 @@
|
||||||
"description": "Method used to save pipeline results to output directory.",
|
"description": "Method used to save pipeline results to output directory.",
|
||||||
"help_text": "The Nextflow `publishDir` option specifies which intermediate files should be saved to the output directory. This option tells the pipeline what method should be used to move these files. See [Nextflow docs](https://www.nextflow.io/docs/latest/process.html#publishdir) for details.",
|
"help_text": "The Nextflow `publishDir` option specifies which intermediate files should be saved to the output directory. This option tells the pipeline what method should be used to move these files. See [Nextflow docs](https://www.nextflow.io/docs/latest/process.html#publishdir) for details.",
|
||||||
"fa_icon": "fas fa-copy",
|
"fa_icon": "fas fa-copy",
|
||||||
"enum": [
|
"enum": ["symlink", "rellink", "link", "copy", "copyNoFollow", "move"],
|
||||||
"symlink",
|
|
||||||
"rellink",
|
|
||||||
"link",
|
|
||||||
"copy",
|
|
||||||
"copyNoFollow",
|
|
||||||
"move"
|
|
||||||
],
|
|
||||||
"hidden": true
|
"hidden": true
|
||||||
},
|
},
|
||||||
"email_on_fail": {
|
"email_on_fail": {
|
||||||
|
@ -776,10 +747,7 @@
|
||||||
"preprocessing_qc_tool": {
|
"preprocessing_qc_tool": {
|
||||||
"type": "string",
|
"type": "string",
|
||||||
"default": "fastqc",
|
"default": "fastqc",
|
||||||
"enum": [
|
"enum": ["fastqc", "falco"],
|
||||||
"fastqc",
|
|
||||||
"falco"
|
|
||||||
],
|
|
||||||
"help_text": "Falco is designed as a drop-in replacement for FastQC but written in C++ for faster computation. We particularly recommend using falco when using long reads (due to reduced memory constraints), however is also applicable for short reads.",
|
"help_text": "Falco is designed as a drop-in replacement for FastQC but written in C++ for faster computation. We particularly recommend using falco when using long reads (due to reduced memory constraints), however is also applicable for short reads.",
|
||||||
"description": "Specify the tool used for quality control of raw sequencing reads"
|
"description": "Specify the tool used for quality control of raw sequencing reads"
|
||||||
}
|
}
|
||||||
|
|
|
@ -274,8 +274,8 @@ workflow PROFILING {
|
||||||
reads: [ it[0] + it[2], it[1] ]
|
reads: [ it[0] + it[2], it[1] ]
|
||||||
db: it[3]
|
db: it[3]
|
||||||
}
|
}
|
||||||
|
// Hardcode to _always_ produce the report file (which is our basic otput, and goes into raw_profiles)
|
||||||
KRAKENUNIQ_PRELOADEDKRAKENUNIQ ( ch_input_for_krakenuniq.reads, ch_input_for_krakenuniq.db, params.krakenuniq_save_reads, params.krakenuniq_save_readclassification )
|
KRAKENUNIQ_PRELOADEDKRAKENUNIQ ( ch_input_for_krakenuniq.reads, ch_input_for_krakenuniq.db, params.krakenuniq_ram_chunk_size, params.krakenuniq_save_reads, true, params.krakenuniq_save_readclassification )
|
||||||
ch_multiqc_files = ch_multiqc_files.mix( KRAKENUNIQ_PRELOADEDKRAKENUNIQ.out.report )
|
ch_multiqc_files = ch_multiqc_files.mix( KRAKENUNIQ_PRELOADEDKRAKENUNIQ.out.report )
|
||||||
ch_versions = ch_versions.mix( KRAKENUNIQ_PRELOADEDKRAKENUNIQ.out.versions.first() )
|
ch_versions = ch_versions.mix( KRAKENUNIQ_PRELOADEDKRAKENUNIQ.out.versions.first() )
|
||||||
ch_raw_classifications = ch_raw_classifications.mix( KRAKENUNIQ_PRELOADEDKRAKENUNIQ.out.classified_assignment )
|
ch_raw_classifications = ch_raw_classifications.mix( KRAKENUNIQ_PRELOADEDKRAKENUNIQ.out.classified_assignment )
|
||||||
|
|
Loading…
Reference in a new issue