mirror of
https://github.com/MillironX/nf-configs.git
synced 2024-12-22 10:38:16 +00:00
docs: Add sysbio
This commit is contained in:
parent
472082254a
commit
2a958f5ce3
3 changed files with 21 additions and 0 deletions
|
@ -130,6 +130,7 @@ Currently documentation is available for the following systems:
|
|||
* [UCT_HPC](docs/uct_hpc.md)
|
||||
* [UPPMAX](docs/uppmax.md)
|
||||
* [UTD_GANYMEDE](docs/utd_ganymede.md)
|
||||
* [UTD_SYSBIO](docs/utd_sysbio.md)
|
||||
* [UZH](docs/uzh.md)
|
||||
|
||||
### Uploading to `nf-core/configs`
|
||||
|
|
18
docs/utd_sysbio.config
Normal file
18
docs/utd_sysbio.config
Normal file
|
@ -0,0 +1,18 @@
|
|||
# nf-core/configs: UTD Sysbio Configuration
|
||||
|
||||
All nf-core pipelines have been successfully configured for use on the Sysbio HPC cluster at the [The Univeristy of Texas at Dallas](https://www.utdallas.edu/).
|
||||
|
||||
To use, run the pipeline with `-profile utd_sysbio`. This will download and launch the [`utd_sysbio.config`](../conf/utd_sysbio.config) which has been pre-configured with a setup suitable for the Sysbio HPC cluster. Using this profile, a docker image containing all of the required software will be downloaded, and converted to a Singularity image before execution of the pipeline.
|
||||
|
||||
Before running the pipeline you will need to load Singularity using the environment module system on Sysbio. You can do this by issuing the commands below:
|
||||
|
||||
```bash
|
||||
## Singularity environment modules
|
||||
module purge
|
||||
module load singularity
|
||||
```
|
||||
|
||||
All of the intermediate files required to run the pipeline will be stored in the `work/` directory. It is recommended to delete this directory after the pipeline has finished successfully because it can get quite large, and all of the main output files will be saved in the `results/` directory anyway.
|
||||
|
||||
>NB: You will need an account to use the HPC cluster on Sysbio in order to run the pipeline. If in doubt contact OIT.
|
||||
>NB: Nextflow will need to submit the jobs via SLURM to the HPC cluster and as such the commands above will have to be executed on one of the login nodes. If in doubt contact OIT.
|
|
@ -49,6 +49,7 @@ profiles {
|
|||
uct_hpc { includeConfig "${params.custom_config_base}/conf/uct_hpc.config" }
|
||||
uppmax { includeConfig "${params.custom_config_base}/conf/uppmax.config" }
|
||||
utd_ganymede { includeConfig "${params.custom_config_base}/conf/utd_ganymede.config" }
|
||||
utd_sysbio { includeConfig "${params.custom_config_base}/conf/utd_sysbio.config" }
|
||||
uzh { includeConfig "${params.custom_config_base}/conf/uzh.config" }
|
||||
jax { includeConfig "${params.custom_config_base}/conf/jax.config" }
|
||||
}
|
||||
|
@ -70,5 +71,6 @@ params {
|
|||
genouest: ['.genouest.org'],
|
||||
uppmax: ['.uppmax.uu.se'],
|
||||
utd_ganymede: ['ganymede.utdallas.edu']
|
||||
utd_sysbio: ['sysbio.utdallas.edu']
|
||||
]
|
||||
}
|
||||
|
|
Loading…
Reference in a new issue