1
0
Fork 0
mirror of https://github.com/MillironX/nf-configs.git synced 2024-11-25 17:29:55 +00:00

Merge remote-tracking branch 'upstream/master'

This commit is contained in:
Alexander Peltzer 2019-11-11 12:21:53 +01:00
commit 426017d727
No known key found for this signature in database
GPG key ID: A4A9322B50AF95A0
6 changed files with 8 additions and 45 deletions

View file

@ -95,7 +95,6 @@ Currently documentation is available for the following clusters:
* [GIS](docs/gis.md)
* [HEBBE](docs/hebbe.md)
* [KRAKEN](docs/kraken.md)
* [MENDEL](docs/mendel.md)
* [MUNIN](docs/munin.md)
* [PASTEUR](docs/pasteur.md)
* [PHOENIX](docs/phoenix.md)

View file

@ -1,24 +0,0 @@
//Profile config names for nf-core/configs
params {
config_profile_description = 'GMI MENDEL cluster profile provided by nf-core/configs'
config_profile_contact = 'Patrick Hüther (@phue)'
config_profile_url = 'http://www.gmi.oeaw.ac.at/'
}
manifest {
nextflowVersion = '>=19.01.0'
}
process {
beforeScript = 'module load Miniconda3/4.6.7'
executor = 'pbspro'
clusterOptions = { "-P $params.project" }
}
params {
max_cpus = 32
max_memory = 128.GB
max_time = 192.h
igenomes_ignore = true
igenomesIgnore = true //deprecated
}

View file

@ -1,18 +0,0 @@
# nf-core/configs: MENDEL Configuration
All nf-core pipelines have been successfully configured for use on the MENDEL CLUSTER at the Gregor Mendel Institute (GMI).
To use, run the pipeline with `-profile conda,mendel`. This will download and launch the [`mendel.config`](../conf/mendel.config) which has been pre-configured with a setup suitable for the MENDEL cluster. A Conda environment will be created automatically and software dependencies will be resolved via [bioconda](https://bioconda.github.io/).
Before running the pipeline you will need to load Conda using the environment module system on MENDEL. You can do this by issuing the commands below:
```bash
## Load Nextflow and Conda environment modules
module purge
module load Nextflow
module load Miniconda/4.6.7
```
>NB: You will need an account to use the HPC cluster in order to run the pipeline. If in doubt contact the HPC team.
>NB: Nextflow will need to submit the jobs via the job scheduler to the HPC cluster and as such the commands above will have to be executed on one of the login nodes. If in doubt contact the HPC team.

View file

@ -10,7 +10,8 @@ To use, run the pipeline with `-profile shh`. This will download and launch the
however this will likely change to a read-only directory in the future that will be managed by IT.
Note that **the configuration file is currently optimised for `nf-core/eager`**. It will submit to the short queue but with a walltime of 2 hours.
This configuration will automatically choose the correct SLURM queue (`short`,`medium`,`long`,`supercruncher`) depending on the time and memory required by each process.
Please note that there is no `supercruncher` queue on CDAG.
>NB: You will need an account and VPN access to use the cluster at MPI-SHH in order to run the pipeline. If in doubt contact IT.

View file

@ -26,6 +26,12 @@ If running on Bianca, you will have no internet connection and these configs wil
Please use the nf-core helper tool on a different system to download the required pipeline files, and transfer them to bianca.
This helper tool bundles the config files in this repo together with the pipeline files, so the profile will still be available.
Note that Bianca only allocates 7 GB memory per core so the max memory needs to be limited:
```bash
--max_memory "112GB"
```
## Getting more memory
If your nf-core pipeline run is running out of memory, you can run on a fat node with more memory using the following nextflow flags:

View file

@ -25,7 +25,6 @@ profiles {
gis { includeConfig "${params.custom_config_base}/conf/gis.config" }
hebbe { includeConfig "${params.custom_config_base}/conf/hebbe.config" }
kraken { includeConfig "${params.custom_config_base}/conf/kraken.config" }
mendel { includeConfig "${params.custom_config_base}/conf/mendel.config" }
munin { includeConfig "${params.custom_config_base}/conf/munin.config" }
pasteur { includeConfig "${params.custom_config_base}/conf/pasteur.config" }
phoenix { includeConfig "${params.custom_config_base}/conf/phoenix.config" }