diff --git a/.travis.yml b/.travis.yml index b200b4a..918a846 100644 --- a/.travis.yml +++ b/.travis.yml @@ -9,10 +9,6 @@ cache: pip matrix: fast_finish: true -before_install: - # PRs to master are only ok if coming from dev branch - - '[ $TRAVIS_PULL_REQUEST = "false" ] || [ $TRAVIS_BRANCH != "master" ] || ([ $TRAVIS_PULL_REQUEST_SLUG = $TRAVIS_REPO_SLUG ] && [ $TRAVIS_PULL_REQUEST_BRANCH = "dev" ])' - install: # Install Nextflow - mkdir /tmp/nextflow && cd /tmp/nextflow @@ -26,4 +22,8 @@ env: script: # Run the pipeline with the test profile and test remote config - - grep "{.*includeConfig.*[a-z]*\.config\"" ${TRAVIS_BUILD_DIR}/nfcore_custom.config | tr -s ' ' | cut -d " " -f 2 | xargs -I {} nextflow run ${TRAVIS_BUILD_DIR}/configtest.nf -profile {} + - | + grep "{.*includeConfig.*[a-z]*\.config\"" ${TRAVIS_BUILD_DIR}/nfcore_custom.config | \ + tr -s ' ' | \ + cut -d " " -f 2 | \ + xargs -I {} nextflow run ${TRAVIS_BUILD_DIR}/configtest.nf -profile {} diff --git a/README.md b/README.md index 6a53642..73b01be 100644 --- a/README.md +++ b/README.md @@ -1,6 +1,4 @@ - - -# [nf-core/configs](https://github.com/nf-core/configs) +# ![nf-core/configs](docs/images/nfcore-configs_logo.png) [![Build Status](https://travis-ci.org/nf-core/configs.svg?branch=master)](https://travis-ci.org/nf-core/configs) @@ -58,7 +56,14 @@ nextflow run /path/to/pipeline/ -c /path/to/my/configs/configs-master/conf/my_co If you decide to upload your custom config file to `nf-core/configs` then this will ensure that your custom config file will be automatically downloaded, and available at run-time to all nf-core pipelines, and to everyone within your organisation. You will simply have to specify `-profile ` in the command used to run the pipeline. See [`nf-core/configs`](https://github.com/nf-core/configs/tree/master/conf) for examples. -Please also make sure to add an extra `params` section with `params. config_profile_name`, `params.config_profile_description`, `params.config_profile_contact` and `params.config_profile_url` set to reasonable values. Users will get information on who wrote the configuration profile then when executing a nf-core pipeline and can report back if there are things missing for example. +Please also make sure to add an extra `params` section with `params.config_profile_description`, `params.config_profile_contact` and `params.config_profile_url` set to reasonable values. Users will get information on who wrote the configuration profile then when executing a nf-core pipeline and can report back if there are things missing for example. + +## Checking user hostnames + +If your cluster has a set of consistent hostnames, nf-core pipelines can check that users are using your profile. +Add one or more hostname substrings to `params.hostnames` under a key that matches the profile name. +If the user's hostname contains this string at the start of a run or when a run fails and their profile +does not contain the profile name, a warning message will be printed. ### Testing @@ -78,6 +83,7 @@ See [`nf-core/configs/docs`](https://github.com/nf-core/configs/tree/master/docs Currently documentation is available for the following clusters: * [BINAC](docs/binac.md) +* [BIGPURPLE](docs/bigpurple.md) * [CCGA](docs/ccga.md) * [CFC](docs/binac.md) * [CRICK](docs/crick.md) @@ -86,6 +92,7 @@ Currently documentation is available for the following clusters: * [MENDEL](docs/mendel.md) * [MUNIN](docs/munin.md) * [PHOENIX](docs/phoenix.md) +* [PRINCE](docs/prince.md) * [SHH](docs/shh.md) * [UCT_HEX](docs/uct_hex.md) * [UPPMAX-DEVEL](docs/uppmax-devel.md) diff --git a/conf/bigpurple.config b/conf/bigpurple.config new file mode 100644 index 0000000..5fdf7e8 --- /dev/null +++ b/conf/bigpurple.config @@ -0,0 +1,27 @@ +singularityDir = "/gpfs/scratch/${USER}/singularity_images_nextflow" + +params { + config_profile_description = """ + NYU School of Medicine BigPurple cluster profile provided by nf-core/configs. + module load both singularity/3.1 and squashfs-tools/4.3 before running the pipeline with this profile!! + Run from your scratch or lab directory - Nextflow makes a lot of files!! + Also consider running the pipeline on a compute node (srun --pty /bin/bash -t=01:00:00) the first time, as it will be pulling the docker image, which will be converted into a singularity image, which is heavy on the login node and will take some time. Subsequent runs can be done on the login node, as the docker image will only be pulled and converted once. By default the images will be stored in $singularityDir + """.stripIndent() + config_profile_contact = 'Tobias Schraink (@tobsecret)' + config_profile_url = 'https://github.com/nf-core/configs/blob/master/docs/bigpurple.md' +} + +singularity { + enabled = true + autoMounts = true + cacheDir = singularityDir +} + +process { + beforeScript = """ + module load singularity/3.1 + module load squashfs-tools/4.3 + """ + .stripIndent() + executor = 'slurm' +} diff --git a/conf/binac.config b/conf/binac.config index 12da68d..0e69d4a 100644 --- a/conf/binac.config +++ b/conf/binac.config @@ -1,23 +1,22 @@ //Profile config names for nf-core/configs params { - config_profile_name = 'BINAC' config_profile_description = 'BINAC cluster profile provided by nf-core/configs.' config_profile_contact = 'Alexander Peltzer (@apeltzer)' config_profile_url = 'https://www.bwhpc-c5.de/wiki/index.php/Category:BwForCluster_BinAC' } + singularity { enabled = true } process { - beforeScript = 'module load devel/singularity/3.0.1' + beforeScript = 'module load devel/singularity/3.0.3' executor = 'pbs' queue = 'short' } params { - igenomesIgnore = false - igenomes_base = '/beegfs/work/igenomes' + igenomes_base = '/nfsmounts/igenomes' max_memory = 128.GB max_cpus = 28 max_time = 48.h diff --git a/conf/ccga.config b/conf/ccga.config index 7dff7c7..b9130e0 100644 --- a/conf/ccga.config +++ b/conf/ccga.config @@ -1,17 +1,21 @@ //Profile config names for nf-core/configs params { - config_profile_name = 'CCGA' config_profile_description = 'CCGA cluster profile provided by nf-core/configs.' config_profile_contact = 'Marc Hoeppner (@marchoeppner)' - config_profile_url = 'https://www.ikmb.uni-kiel.de/' + config_profile_url = 'https://www.ccga.uni-kiel.de/' } /* * ------------------------------------------------- - * Nextflow config file with environment modules for RZCluster in Kiel + * Nextflow config file for CCGA cluster in Kiel * ------------------------------------------------- */ +singularity { + enabled = true + runOptions = "-B /ifs -B /scratch -B /work_beegfs" +} + executor { queueSize=100 } @@ -25,9 +29,12 @@ process { clusterOptions = { "--qos=ikmb_a" } } - + params { // illumina iGenomes reference file paths on RZCluster igenomes_base = '/ifs/data/nfs_share/ikmb_repository/references/iGenomes/references/' saveReference = true + max_memory = 128.GB + max_cpus = 16 + max_time = 120.h } diff --git a/conf/cfc.config b/conf/cfc.config index fc8cc34..2f9047d 100644 --- a/conf/cfc.config +++ b/conf/cfc.config @@ -1,6 +1,5 @@ //Profile config names for nf-core/configs params { - config_profile_name = 'CFC' config_profile_description = 'QBiC Core Facility cluster profile provided by nf-core/configs.' config_profile_contact = 'Alexander Peltzer (@apeltzer)' config_profile_url = 'http://qbic.uni-tuebingen.de/' @@ -15,9 +14,13 @@ process { executor = 'slurm' } +weblog{ + enabled = true + url = 'http://services.qbic.uni-tuebingen.de:8080/workflows' +} + params { - igenomesIgnore = false - igenomes_base = '/sfs/7/igenomes' + igenomes_base = '/nfsmounts/igenomes' max_memory = 60.GB max_cpus = 20 max_time = 140.h diff --git a/conf/crick.config b/conf/crick.config index de8c89b..10ddba5 100755 --- a/conf/crick.config +++ b/conf/crick.config @@ -1,8 +1,7 @@ //Profile config names for nf-core/configs params { - config_profile_name = 'CRICK' config_profile_description = 'The Francis Crick Institute CAMP HPC cluster profile provided by nf-core/configs.' - config_profile_contact = 'Harshil Patel (@drpatelh )' + config_profile_contact = 'Harshil Patel (@drpatelh)' config_profile_url = 'https://www.crick.ac.uk/research/platforms-and-facilities/scientific-computing/technologies' } diff --git a/conf/czbiohub_aws.config b/conf/czbiohub_aws.config index 71d1a9b..1dd8da2 100644 --- a/conf/czbiohub_aws.config +++ b/conf/czbiohub_aws.config @@ -76,7 +76,7 @@ params { fasta = "${params.gencode_base}/mouse/vM21/GRCm38.p6.genome.ERCC92.fa" gtf = "${params.gencode_base}/mouse/vM21/gencode.vM21.annotation.ERCC92.gtf" transcript_fasta = "${params.gencode_base}/mouse/vM21/gencode.vM21.transcripts.ERCC92.fa" - start = "${params.gencode_base}/mouse/vM21/STARIndex/" + star = "${params.gencode_base}/mouse/vM21/STARIndex/" } } diff --git a/conf/gis.config b/conf/gis.config index 3fcb72e..e1645e5 100644 --- a/conf/gis.config +++ b/conf/gis.config @@ -1,11 +1,9 @@ -/* - * ------------------------------------------------- - * Nextflow config file for GIS (Aquila) - * ------------------------------------------------- - * Defines reference genomes, using iGenome paths - * Imported under the default 'standard' Nextflow - * profile in nextflow.config - */ +//Profile config names for nf-core/configs +params { + config_profile_description = 'Genome Institute of Singapore (Aquila) cluster profile provided by nf-core/configs.' + config_profile_contact = 'Andreas Wilm (@andreas-wilm)' + config_profile_url = 'https://www.a-star.edu.sg/gis/' +} process { executor = 'sge' @@ -20,4 +18,3 @@ params { // illumina iGenomes reference file paths on GIS Aquila igenomes_base = '/mnt/projects/rpd/genomes.testing/S3_igenomes/' } - diff --git a/conf/hebbe.config b/conf/hebbe.config index 76b8ac7..500ec31 100644 --- a/conf/hebbe.config +++ b/conf/hebbe.config @@ -1,8 +1,7 @@ //Profile config names for nf-core/configs params { - config_profile_name = 'HEBBE' config_profile_description = 'Gothenburg Hebbe cluster profile provided by nf-core/configs.' - config_profile_contact = 'Phil Ewels (@ewels )' + config_profile_contact = 'Phil Ewels (@ewels)' config_profile_url = 'http://www.c3se.chalmers.se/index.php/Hebbe' } diff --git a/conf/mendel.config b/conf/mendel.config index 0715b42..e658c87 100644 --- a/conf/mendel.config +++ b/conf/mendel.config @@ -1,6 +1,5 @@ //Profile config names for nf-core/configs params { - config_profile_name = 'MENDEL' config_profile_description = 'GMI MENDEL cluster profile provided by nf-core/configs' config_profile_contact = 'Patrick Hüther (@phue)' config_profile_url = 'http://www.gmi.oeaw.ac.at/' diff --git a/conf/munin.config b/conf/munin.config index a24929e..bca56f7 100644 --- a/conf/munin.config +++ b/conf/munin.config @@ -1,6 +1,5 @@ //Profile config names for nf-core/configs params { - config_profile_name = 'MUNIN' config_profile_description = 'Big iron cluster profile provided by nf-core/configs.' config_profile_contact = 'Szilveszter Juhos (@szilva)' config_profile_url = '' @@ -29,5 +28,5 @@ params { max_cpus = 16 max_time = 72.h // illumina iGenomes reference file paths on UPPMAX - igenomes_base = '/data0/btb/references/igenomes/' + igenomes_base = '/data1/references/igenomes/' } diff --git a/conf/phoenix.config b/conf/phoenix.config index 685e1e5..b457763 100644 --- a/conf/phoenix.config +++ b/conf/phoenix.config @@ -1,6 +1,5 @@ //Profile config names for nf-core/configs params { - config_profile_name = 'PHOENIX' config_profile_description = 'University of Adelaide Phoenix HPC cluster profile provided by nf-core/configs' config_profile_contact = 'Yassine Souilmi / Alexander Peltzer (@yassineS, @apeltzer)' config_profile_url = 'https://www.adelaide.edu.au/phoenix/' diff --git a/conf/prince.config b/conf/prince.config new file mode 100644 index 0000000..1c2ea2c --- /dev/null +++ b/conf/prince.config @@ -0,0 +1,27 @@ +singularityDir = "$SCRATCH/singularity_images_nextflow" +singularityModule = "singularity/3.2.1" +squashfsModule = "squashfs/4.3" + +params { + config_profile_description = """ + NYU prince cluster profile provided by nf-core/configs. + Run from your scratch directory, the output files may be large! + Please consider running the pipeline on a compute node the first time, as it will be pulling the docker image, which will be converted into a singularity image, which is heavy on the login node. Subsequent runs can be done on the login node, as the docker image will only be pulled and converted once. By default the images will be stored in $singularityDir + """.stripIndent() + config_profile_contact = 'Tobias Schraink (@tobsecret)' + config_profile_url = 'https://github.com/nf-core/configs/blob/master/docs/prince.md' +} + +singularity { + enabled = true + cacheDir = singularityDir +} + +process { + beforeScript = """ + module load $singularityModule + module load $squashfsModule + """ + .stripIndent() + executor = 'slurm' +} diff --git a/conf/shh.config b/conf/shh.config index 13b7f5c..5cb3396 100644 --- a/conf/shh.config +++ b/conf/shh.config @@ -1,6 +1,5 @@ //Profile config names for nf-core/configs params { - config_profile_name = 'SHH' config_profile_description = 'MPI SHH cluster profile provided by nf-core/configs.' config_profile_contact = 'James Fellows Yates (@jfy133)' config_profile_url = 'https://shh.mpg.de' @@ -8,7 +7,9 @@ params { singularity { enabled = true - cacheDir = "/projects1/users/$USER/nextflow/nf_cache/singularity/" + autoMounts = true + runOptions = '-B /run/shm:/run/shm' + cacheDir = "/projects1/singularity_scratch/cache/" } process { @@ -16,8 +17,14 @@ process { queue = 'short' } -params { - max_memory = 734.GB - max_cpus = 64 - max_time = 2.h +executor { + queueSize = 16 } + +params { + max_memory = 256.GB + max_cpus = 32 + max_time = 2.h + //Illumina iGenomes reference file path + igenomes_base = "/projects1/public_data/igenomes/" +} \ No newline at end of file diff --git a/conf/uct_hex.config b/conf/uct_hex.config index a372000..6e0bbbe 100644 --- a/conf/uct_hex.config +++ b/conf/uct_hex.config @@ -1,6 +1,5 @@ //Profile config names for nf-core/configs params { - config_profile_name = 'uct_hex' config_profile_description = 'University of Cape Town HEX cluster config file provided by nf-core/configs.' config_profile_contact = 'Katie Lennard (@kviljoen)' config_profile_url = 'http://hpc.uct.ac.za/index.php/hex-3/' @@ -22,4 +21,3 @@ executor{ executor = 'pbs' jobName = { "$task.tag" } } - diff --git a/conf/uppmax-devel.config b/conf/uppmax-devel.config index 760ff77..ad71ac4 100644 --- a/conf/uppmax-devel.config +++ b/conf/uppmax-devel.config @@ -1,6 +1,5 @@ // Profile config names for nf-core/configs params { - config_profile_name = 'UPPMAX-devel' config_profile_description = 'Testing & development profile for UPPMAX, provided by nf-core/configs.' config_profile_contact = 'Phil Ewels (@ewels)' config_profile_url = 'https://www.uppmax.uu.se/' diff --git a/conf/uppmax.config b/conf/uppmax.config index 939b8a7..d2fbc5d 100644 --- a/conf/uppmax.config +++ b/conf/uppmax.config @@ -1,6 +1,5 @@ //Profile config names for nf-core/configs params { - config_profile_name = 'UPPMAX' config_profile_description = 'Swedish UPPMAX cluster profile provided by nf-core/configs.' config_profile_contact = 'Phil Ewels (@ewels)' config_profile_url = 'https://www.uppmax.uu.se/' diff --git a/conf/uzh.config b/conf/uzh.config index 70aad1b..c888096 100644 --- a/conf/uzh.config +++ b/conf/uzh.config @@ -1,6 +1,5 @@ //Profile config names for nf-core/configs params{ - config_profile_name = 'UZH' config_profile_description = 'UZH science cloud profile provided by nf-core/configs' config_profile_contact = 'Judith Neukamm/Alexander Peltzer (@JudithNeukamm, @apeltzer)' config_profile_url = 'https://www.id.uzh.ch/en/scienceit/infrastructure/sciencecloud.html' diff --git a/docs/bigpurple.md b/docs/bigpurple.md new file mode 100644 index 0000000..19b0352 --- /dev/null +++ b/docs/bigpurple.md @@ -0,0 +1,24 @@ +# nf-core/configs: BigPurple Configuration +## nf-core pipelines that use this repo +All nf-core pipelines that use this config repo (which is most), can be run on BigPurple. **Before** running a pipeline for the first time, go into an interactive slurm session on a compute node (`srun --pty --time=02:00:00 -c 2`), as the docker image for the pipeline will need to be pulled and converted. Once in the interactive session: + +``` +module load singularity/3.1 +module load squashfs-tools/4.3 +``` + +Now, run the pipeline of your choice with `-profile bigpurple`. This will download and launch the bigpurple.config which has been pre-configured with a setup suitable for the BigPurple cluster. Using this profile, a docker image containing all of the required software will be downloaded, and converted to a singularity image before execution of the pipeline. +An example commandline: + +`nextflow run nf-core/ -profile bigpurple ` + +## nf-core pipelines that do not use this repo +If the pipeline has not yet been configured to use this config, then you will have to do it manually. +git clone this repo, copy the `bigpurple.config` from the conf folder and then you can invoke the pipeline like this: + +`nextflow run nf-core/ -c bigpurple.config ` + + +>NB: You will need an account to use the HPC cluster BigPurple in order to run the pipeline. If in doubt contact MCIT. + +>NB: You will need to install nextflow in your home directory - instructions are on nextflow.io (or ask the writer of this profile). The reason there is no module for nextflow on the cluster, is that the development cycle of nextflow is rapid and it's easy to update yourself: `nextflow self-update` diff --git a/docs/ccga.md b/docs/ccga.md new file mode 100644 index 0000000..1157cb3 --- /dev/null +++ b/docs/ccga.md @@ -0,0 +1,18 @@ +# nf-core/configs: CCGA Configuration + +Deployment and testing of nf-core pipelines at the CCGA cluster is on-going. + +To use, run the pipeline with `-profile ccga`. This will download and launch the [`ccga.config`](../conf/ccga.config) which has been pre-configured with a setup suitable for the CCGA cluster. Using this profile, a docker image containing all of the required software will be downloaded, and converted to a Singularity image before execution of the pipeline. + +Before running the pipeline you will need to load Nextflow and Singularity using the environment module system on the cluster. You can do this by issuing the commands below: + +```bash +## Load Nextflow and Singularity environment modules +module purge +module load IKMB +module load Java/1.8.0 +module load Nextflow +module load singularity3.1.0 +``` + +>NB: Access to the CCGA cluster is restricted to IKMB/CCGA employes. Please talk to Marc Hoeppner to get access (@marchoeppner). diff --git a/docs/czbiohub.md b/docs/czbiohub.md index 5e98f4f..0120283 100644 --- a/docs/czbiohub.md +++ b/docs/czbiohub.md @@ -22,6 +22,7 @@ Now you can run pipelines with abandon! ### 2. Make a GitHub repo for your workflows (optional :) + To make sharing your pipelines and commands easy between your teammates, it's best to share code in a GitHub repository. One way is to store the commands in a Makefile ([example](https://github.com/czbiohub/kh-workflows/blob/master/nf-kmer-similarity/Makefile)) which can contain multiple `nextflow run` commands so that you don't need to remember the S3 bucket or output directory for every single one. [Makefiles](https://kbroman.org/minimal_make/) are broadly used in the software community for running many complex commands. Makefiles can have a lot of dependencies and be confusing, so we're only going to write *simple* Makefiles. ``` @@ -82,6 +83,7 @@ git push origin master ### 3. Run your workflow!! + Remember to specify `-profile czbiohub_aws` to grab the CZ Biohub-specific AWS configurations, and an `--outdir` with an AWS S3 bucket so you don't run out of space on your small AMI ``` @@ -115,6 +117,7 @@ It's important that this command be re-run from the same directory as there is a ## iGenomes specific configuration A local copy of the iGenomes resource has been made available on `s3://czbiohub-reference/igenomes` (in `us-west-2` region) so you should be able to run the pipeline against any reference available in the `igenomes.config` specific to the nf-core pipeline. + You can do this by simply using the `--genome ` parameter. For Human and Mouse, we use [GENCODE](https://www.gencodegenes.org/) gene annotations. This doesn't change how you would specify the genome name, only that the pipelines run with the `czbiohub_aws` profile would be with GENCODE rather than iGenomes. diff --git a/docs/images/nfcore-configs_logo.png b/docs/images/nfcore-configs_logo.png new file mode 100755 index 0000000..eef0738 Binary files /dev/null and b/docs/images/nfcore-configs_logo.png differ diff --git a/docs/prince.md b/docs/prince.md new file mode 100644 index 0000000..12f5e65 --- /dev/null +++ b/docs/prince.md @@ -0,0 +1,20 @@ +# nf-core/configs: Prince Configuration +## nf-core pipelines that use this repo +All nf-core pipelines that use this config repo (which is most), can be run on prince. **Before** running a pipeline for the first time, go into an interactive slurm session on a compute node (`srun --pty --time=02:00:00 -c 2`), as the docker image for the pipeline will need to be pulled and converted. + +Now, run the pipeline of your choice with `-profile prince`. This will download and launch the prince.config which has been pre-configured with a setup suitable for the prince cluster. Using this profile, a docker image containing all of the required software will be downloaded, and converted to a singularity image before execution of the pipeline. This step **takes time**!! +An example commandline: + +`nextflow run nf-core/ -profile prince ` + +## nf-core pipelines that do not use this repo +If the pipeline has not yet been configured to use this config, then you will have to do it manually. +git clone this repo, copy the `prince.config` from the conf folder and then you can invoke the pipeline like this: + +`nextflow run nf-core/ -c prince.config ` + + +>NB: You will need an account to use the HPC cluster Prince in order to run the pipeline. If in doubt contact the HPC admins. + +>NB: Rather than using the nextflow module, I recommend you install nextflow in your home directory - instructions are on nextflow.io (or ask the writer of this profile). The reason this is better than using the module for nextflow on the cluster, is that the development cycle of nextflow is rapid and it's easy to update your installation yourself: `nextflow self-update` + diff --git a/docs/shh.md b/docs/shh.md index bcdda3c..67cf3b9 100644 --- a/docs/shh.md +++ b/docs/shh.md @@ -1,21 +1,17 @@ # nf-core/configs: SHH Configuration -All nf-core pipelines have been successfully configured for use on the Department of Archaeogenetic's SDAG cluster at the [Max Planck Institute for the Science of Human History (MPI-SHH)](http://shh.mpg.de). +All nf-core pipelines have been successfully configured for use on the Department of Archaeogenetic's SDAG/CDAG clusters at the [Max Planck Institute for the Science of Human History (MPI-SHH)](http://shh.mpg.de). -To use, run the pipeline with `-profile shh`. This will download and launch the [`shh.config`](../conf/shh.config) which has been pre-configured with a setup suitable for the SDAG cluster. Using this profile, a docker image containing all of the required software will be downloaded, and converted to a Singularity image before execution of the pipeline. - -Note that the configuration file is currently optimised for `nf-core/eager`. It -will submit to the medium queue but with a walltime of 48 hours. - -## Preparation -Before running the pipeline you will need to create a the following folder in your `/projects1/users/` directory. This will be used to store the singularity software images, which will take up too much space for your home directory. - -This should be named as follows, replacing `` with your username: +To use, run the pipeline with `-profile shh`. This will download and launch the [`shh.config`](../conf/shh.config) which has been pre-configured with a setup suitable for the SDAG and CDAG clusters. Using this profile, a docker image containing all of the required software will be downloaded, and converted to a Singularity image before execution of the pipeline. The image will currently be centrally stored here: ```bash -"/projects1/users//nextflow/nf_cache/singularity/" +/projects1/singularity_scratch/cache/ ``` +however this will likely change to a read-only directory in the future that will be managed by IT. + +Note that **the configuration file is currently optimised for `nf-core/eager`**. It will submit to the short queue but with a walltime of 2 hours. + >NB: You will need an account and VPN access to use the cluster at MPI-SHH in order to run the pipeline. If in doubt contact IT. ->NB: Nextflow will need to submit the jobs via the job scheduler to the HPC cluster and as such the commands above will have to be executed on one of the lhead nodes. If in doubt contact IT. +>NB: Nextflow will need to submit the jobs via SLURM to the clusters and as such the commands above will have to be executed on one of the head nodes. If in doubt contact IT. diff --git a/nfcore_custom.config b/nfcore_custom.config index 72dca30..41b0ade 100644 --- a/nfcore_custom.config +++ b/nfcore_custom.config @@ -28,9 +28,20 @@ profiles { phoenix { includeConfig "${params.custom_config_base}/conf/phoenix.config" } shh { includeConfig "${params.custom_config_base}/conf/shh.config" } uct_hex { includeConfig "${params.custom_config_base}/conf/uct_hex.config" } - uppmax_devel { includeConfig "${params.custom_config_base}/conf/uppmax.config" - includeConfig "${params.custom_config_base}/conf/uppmax-devel.config" - } + uppmax_devel { includeConfig "${params.custom_config_base}/conf/uppmax.config"; includeConfig "${params.custom_config_base}/conf/uppmax-devel.config" } uppmax { includeConfig "${params.custom_config_base}/conf/uppmax.config" } uzh { includeConfig "${params.custom_config_base}/conf/uzh.config" } + prince { includeConfig "${params.custom_config_base}/conf/prince.config" } + bigpurple { includeConfig "${params.custom_config_base}/conf/bigpurple.config" } +} + +// If user hostnames contain one of these substring and they are +// not running the associated profile, it will trigger a warning message +// Should be defined here for all profiles (not within profile config) +params { + // This is a groovy map, not a nextflow parameter set + hostnames = [ + crick: ['.thecrick.org'], + uppmax: ['.uppmax.uu.se'] + ] }