From 15a0571eae0defa43de4c0e28532724aa4c6c98c Mon Sep 17 00:00:00 2001 From: "James A. Fellows Yates" Date: Wed, 2 Mar 2022 08:50:42 +0100 Subject: [PATCH 1/2] Fix markdown --- docs/cheaha.md | 8 +++----- 1 file changed, 3 insertions(+), 5 deletions(-) diff --git a/docs/cheaha.md b/docs/cheaha.md index b4a0eb9..27d42b4 100644 --- a/docs/cheaha.md +++ b/docs/cheaha.md @@ -15,10 +15,8 @@ module load Nextflow All of the intermediate files required to run the pipeline will be stored in the `work/` directory. It is recommended to delete this directory after the pipeline has finished successfully because it can get quite large, and all of the main output files will be saved in the `results/` directory anyway. ->NB: You will need an account to use the HPC cluster on Cheaha in order to run the pipeline. If in doubt contact UAB IT Research Computing. - ->NB: Nextflow will need to submit the jobs via SLURM to the HPC cluster and as such the commands above will have to be executed on one of the login nodes (or alternatively in an interactive partition, but be aware of time limit). If in doubt contact UAB IT Research Computing. - +>NB: You will need an account to use the HPC cluster on Cheaha in order to run the pipeline. If in doubt contact UAB IT Research Computing.

+>NB: Nextflow will need to submit the jobs via SLURM to the HPC cluster and as such the commands above will have to be executed on one of the login nodes (or alternatively in an interactive partition, but be aware of time limit). If in doubt contact UAB IT Research Computing.

>NB: Instead of using `module load Nextflow`, you may instead create a conda environment (e.g: `conda create -p $USER_DATA/nf-core_nextflow_env nf-core nextflow`) if you would like to have a more personalized environment of Nextflow (versions which may not be modules yet) and nf-core tools. This __requires__ you to instead do the following: ```bash @@ -29,4 +27,4 @@ module load Anaconda3 conda activate $USER_DATA/nf-core_nextflow_env ``` ->NB: while the jobs for each process of the pipeline are sent to the appropriate nodes, the current session must remain active while the pipeline is running. We recommend to use `screen` prior to loading any modules/environments. Once the pipeline starts you can detach the screen session by typing `Ctrl-a d` so you can safely logout of HPC, while keeping the pipeline active (and you may resume the screen session with `screen -r`). Other similar tools (e.g. `tmux`) may also be used. \ No newline at end of file +>NB: while the jobs for each process of the pipeline are sent to the appropriate nodes, the current session must remain active while the pipeline is running. We recommend to use `screen` prior to loading any modules/environments. Once the pipeline starts you can detach the screen session by typing `Ctrl-a d` so you can safely logout of HPC, while keeping the pipeline active (and you may resume the screen session with `screen -r`). Other similar tools (e.g. `tmux`) may also be used. From 121452afadd5991cbe12e0ff2f1c0193b81edd5f Mon Sep 17 00:00:00 2001 From: "James A. Fellows Yates" Date: Wed, 2 Mar 2022 08:51:20 +0100 Subject: [PATCH 2/2] Linting --- docs/marvin.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/marvin.md b/docs/marvin.md index 5ac5666..1d07aaf 100644 --- a/docs/marvin.md +++ b/docs/marvin.md @@ -1,9 +1,9 @@ # nf-core/configs: Marvin Configuration -All nf-core pipelines have been successfully configured for use on the [Marvin HPC cluster of the Universitat Pompeu Fabra (UPF).] (https://www.upf.edu/web/sct-sit/marvin-cluster). To use, run the pipeline with -profile marvin. +All nf-core pipelines have been successfully configured for use on the [Marvin HPC cluster of the Universitat Pompeu Fabra (UPF)](https://www.upf.edu/web/sct-sit/marvin-cluster). To use, run the pipeline with -profile marvin. NB: You will need an account to use the HPC cluster on the Marvin cluster in order to run the pipeline. If in doubt contact IT. NB: Nextflow will need to submit the jobs via the SLURM scheduler to the HPC cluster and as such the commands above will have to be executed on one of the head nodes. If in doubt contact IT. -## Pipeline Specific profiles +## Pipeline Specific profiles There are no specific profiles added for now