1
0
Fork 0
mirror of https://github.com/MillironX/nf-configs.git synced 2024-11-22 00:26:03 +00:00
nf-configs/docs/uppmax.md

88 lines
4.1 KiB
Markdown
Raw Normal View History

2019-01-07 13:25:06 +00:00
# nf-core/configs: UPPMAX Configuration
All nf-core pipelines have been successfully configured for use on the Swedish UPPMAX clusters.
2020-11-25 21:45:16 +00:00
## Getting help
We have a Slack channel dedicated to UPPMAX users on the nf-core Slack: [https://nfcore.slack.com/channels/uppmax](https://nfcore.slack.com/channels/uppmax)
2019-01-07 13:25:06 +00:00
## Using the UPPMAX config profile
2019-11-25 23:48:17 +00:00
2019-12-05 14:18:01 +00:00
To use, run the pipeline with `-profile uppmax` (one hyphen).
This will download and launch the [`uppmax.config`](../conf/uppmax.config) which has been pre-configured with a setup suitable for the UPPMAX servers.
Using this profile, a docker image containing all of the required software will be downloaded, and converted to a Singularity image before execution of the pipeline.
2019-01-07 13:25:06 +00:00
In addition to this config profile, you will also need to specify an UPPMAX project id.
You can do this with the `--project` flag (two hyphens) when launching nextflow. For example:
```bash
2020-11-26 00:43:09 +00:00
nextflow run nf-core/PIPELINE -profile uppmax --project snic2018-1-234 # ..rest of pipeline flags
2019-01-07 13:25:06 +00:00
```
2020-11-25 21:45:16 +00:00
> NB: If you're not sure what your UPPMAX project ID is, try running `groups` or checking SUPR.
2019-01-07 13:25:06 +00:00
Before running the pipeline you will need to either install Nextflow or load it using the environment module system.
2020-11-25 21:45:16 +00:00
This config enables Nextflow to manage the pipeline jobs via the Slurm job scheduler and using Singularity for software management.
2019-01-07 13:25:06 +00:00
Just run Nextflow on a login node and it will handle everything else.
2020-11-25 21:45:16 +00:00
Remember to use `-bg` to launch Nextflow in the background, so that the pipeline doesn't exit if you leave your terminal session.
2019-01-07 13:25:06 +00:00
## Using iGenomes references
2019-11-25 23:48:17 +00:00
2019-01-07 13:25:06 +00:00
A local copy of the iGenomes resource has been made available on all UPPMAX clusters so you should be able to run the pipeline against any reference available in the `igenomes.config`.
You can do this by simply using the `--genome <GENOME_ID>` parameter.
## Getting more memory
2019-11-25 23:48:17 +00:00
2019-01-07 13:25:06 +00:00
If your nf-core pipeline run is running out of memory, you can run on a fat node with more memory using the following nextflow flags:
```bash
--clusterOptions "-C mem256GB" --max_memory "256GB"
```
This raises the ceiling of available memory from the default of `128.GB` to `256.GB`.
Rackham has nodes with 128GB, 256GB and 1TB memory available.
Note that each job will still start with the same request as normal, but restarted attempts with larger requests will be able to request greater amounts of memory.
All jobs will be submitted to fat nodes using this method, so it's only for use in extreme circumstances.
2020-11-25 21:45:16 +00:00
## Different UPPMAX clusters
2019-12-05 14:18:01 +00:00
The UPPMAX nf-core configuration profile uses the `hostname` of the active environment to automatically apply the following resource limits:
2019-12-05 14:18:01 +00:00
* `bianca`
* cpus available: 16 cpus
* memory available: 109 GB
* `irma`
* cpus available: 16 cpus
* memory available: 250 GB
* `rackham`
* cpus available: 20 cpus
* memory available: 125 GB
## Development config
2019-11-25 23:48:17 +00:00
2019-12-05 14:18:01 +00:00
If doing pipeline development work on UPPMAX, the `devel` profile allows for faster testing.
2019-01-07 13:25:06 +00:00
Applied after main UPPMAX config, it overwrites certain parts of the config and submits jobs to the `devcore` queue, which has much faster queue times.
All jobs are limited to 1 hour to be eligible for this queue and only one job allowed at a time.
It is not suitable for use with real data.
To use it, submit with `-profile uppmax,devel`.
2020-11-26 08:53:51 +00:00
## Running on Bianca
2020-11-26 12:58:16 +00:00
For security reasons, there is no internet access on Bianca so you can't download from or upload files to the cluster directly. Before running a nf-core pipeline on Bianca you will first have to download the pipeline and singularity images needed elsewhere and transfer them via the wharf area to your Bianca project.
2020-11-26 08:53:51 +00:00
2020-11-26 12:58:16 +00:00
You can follow the guide for downloading pipelines [for offline use](https://nf-co.re/tools#downloading-pipelines-for-offline-use). Note that you will have to download the singularity images as well.
2020-11-26 08:53:51 +00:00
2020-11-26 10:41:33 +00:00
Next transfer the pipeline and the singularity images to your project. Before running the pipeline you will have to indicate to nextflow where the singularity images are located by setting `NXF_SINGULARITY_CACHEDIR` :
2020-11-26 08:53:51 +00:00
`export NXF_SINGULARITY_CACHEDIR=Your_Location_For_The_Singularity_directory/.`
You should now be able to run your nf-core pipeline on bianca.