1
0
Fork 0
mirror of https://github.com/MillironX/nf-configs.git synced 2024-12-22 10:38:16 +00:00

Merge pull request #344 from njspix/vai

Add VAI configuration
This commit is contained in:
James A. Fellows Yates 2022-03-14 14:36:00 +01:00 committed by GitHub
commit a2e2e94333
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
5 changed files with 37 additions and 0 deletions

View file

@ -71,6 +71,7 @@ jobs:
- 'utd_ganymede'
- 'utd_sysbio'
- 'uzh'
- 'vai'
steps:
- uses: actions/checkout@v1
- name: Install Nextflow

View file

@ -137,6 +137,7 @@ Currently documentation is available for the following systems:
* [UTD_GANYMEDE](docs/utd_ganymede.md)
* [UTD_SYSBIO](docs/utd_sysbio.md)
* [UZH](docs/uzh.md)
* [VAI](docs/vai.md)
### Uploading to `nf-core/configs`

20
conf/vai.config Normal file
View file

@ -0,0 +1,20 @@
params {
config_profile_description = 'Van Andel Institute HPC profile provided by nf-core/configs.'
config_profile_contact = 'Nathan Spix (@njspix)'
config_profile_url = 'https://vanandelinstitute.sharepoint.com/sites/SC/SitePages/HPC3-High-Performance-Cluster-and-Cloud-Computing.aspx'
max_memory = 250.GB
max_cpus = 40
max_time = 640.h
}
process {
beforeScript = 'alias singularity=/varidata/research/software/singularity/singularity-ce-3.8.2/bin/singularity'
executor = 'pbs'
queue = { task.time <= 48.h ? 'shortq' : 'longq' }
maxRetries = 2
}
singularity {
enabled = true
autoMounts = true
}

14
docs/vai.md Normal file
View file

@ -0,0 +1,14 @@
# nf-core/configs: VAI configuration
All nf-core pipelines have been successfully configured for use on the HPC cluster at Van Andel Institute.
To use, run the pipeline with `-profile vai`. This will download and launch the [`vai.config`](../conf/vai.config) which has been pre-configured with a setup suitable for the VAI HPC. Using this profile, a docker image containing all of the required software will be downloaded, and converted to a Singularity image before execution of the pipeline.
```bash
module load singularity
NXF_OPTS="-Xmx500m" MALLOC_ARENA_MAX=4 nextflow run <pipeline>
```
>NB: You will need an account to use the HPC in order to run the pipeline. If in doubt contact IT.
>NB: Nextflow will need to submit the jobs via the job scheduler to the HPC cluster and as such the commands above will have to be executed on the login node. If in doubt contact IT.
>NB: The submit node limits the amount of memory available to each user. The `NXF_OPTS` and `MALLOC_ARENA_MAX` parameters above prevent Nextflow from allocating more memory than the scheduler will allow.

View file

@ -65,4 +65,5 @@ profiles {
utd_ganymede { includeConfig "${params.custom_config_base}/conf/utd_ganymede.config" }
utd_sysbio { includeConfig "${params.custom_config_base}/conf/utd_sysbio.config" }
uzh { includeConfig "${params.custom_config_base}/conf/uzh.config" }
vai { includeConfig "${params.custom_config_base}/conf/vai.config" }
}