Skip to content

Commit b071181

Browse files
authored
Merge branch 'dev' into dev
2 parents 2048b10 + 3020cbb commit b071181

File tree

8 files changed

+29
-36
lines changed

8 files changed

+29
-36
lines changed

docs/images/tower_run.png

135 KB
Loading

docs/images/tower_run1.png

69.7 KB
Loading

docs/images/tower_run2.png

144 KB
Loading

docs/images/tower_workspaces.png

49.8 KB
Loading

docs/markdown/clusters/running_jobs.md

Lines changed: 8 additions & 22 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,6 @@ These are general statements on how to submit jobs to our clusters.
77
* [Dependencies](#dependencies)
88
* [Install Nextflow](#install-nextflow)
99
* [Nextflow version](#nextflow-version)
10-
* [Loading singularity modules](#loading-singularity-modules)
1110
* [Pipeline profiles](#pipeline-profiles)
1211
* [Example bash file](#example-bash-file)
1312
* [Screen sessions](#screen-sessions)
@@ -24,12 +23,11 @@ of the functionality might not be available, for example the `cfc` and `binac` p
2423

2524
Please start all your nextflow jobs from the head node.
2625
Nextflow interacts directly with the Slurm scheduler and will take care of submitting individual jobs to the nodes.
27-
If you submit via an interactive job, weird errors can occur, e.g. the cache directory for the containers is mounted read only on the compute nodes, so you can't pull new containers from a node.
26+
If you submit via an interactive job, strange errors can occur, e.g. the cache directory for the containers is mounted read-only on the compute nodes, so you can't pull new containers from a node.
2827

2928
### Dependencies
3029

31-
To run Nextflow pipelines in our clusters, you will need Nextflow, java and singularity installed.
32-
Luckily, our sysadmin made a module for singularity and java in our clusters already, so you will just need to load these modules.
30+
To run Nextflow pipelines in our clusters, you will need Nextflow, java and singularity installed. Java and singularity are already installed in all cluster nodes so these do not need to be installed. You also do not need to load these modules.
3331

3432
You will still have to install Nextflow for your user, that's very simple and described in the next section.
3533

@@ -41,7 +39,7 @@ You will still have to install Nextflow for your user, that's very simple and de
4139
wget -qO- get.nextflow.io | bash
4240
```
4341

44-
* Optionally, move the nextflow file in a directory accessible by your `$PATH` variable (this is only required to avoid to remember and type the Nextflow full path each time you need to run it).
42+
* Optionally, move the nextflow file in a directory accessible by your `$PATH` variable (required only to avoid remembering and typing the Nextflow full path each time you need to run it).
4543

4644
For more information, visit the [Nextflow documentation](https://www.nextflow.io/docs/latest/en/latest/getstarted.html).
4745

@@ -65,17 +63,6 @@ If not, set it by running:
6563
NXF_VER=19.10.0
6664
```
6765

68-
### Loading singularity modules
69-
70-
We currently load Singularity as a module on both BinAC and CFC to make sure that all paths are set accordingly and load required configuration parameters tailored for the respective system.
71-
Please do use *only* these Singularity versions and *NOT* any other (custom) singularity versions out there.
72-
These singularity modules will already load the required java module
73-
so you don't need to take care of that.
74-
75-
```bash
76-
module load devel/singularity/3.4.2
77-
```
78-
7966
### Pipeline profiles
8067

8168
#### On CFC
@@ -105,7 +92,7 @@ nextflow run custom/pipeline -c custom.config
10592
Please use the [binac profile](https://github.com/nf-core/configs/blob/master/conf/binac.config) by adding `-profile binac` to run your analyses. For example:
10693

10794
```bash
108-
nextflow run nf-core/rnaseq -r 1.4.2 -profile cfc_dev
95+
nextflow run nf-core/rnaseq -r 1.4.2 -profile binac
10996
```
11097

11198
For Nextflow pipelines not part of nf-core and not created with the nf-core create command, these profiles will not be available.
@@ -122,7 +109,6 @@ Here is an example bash file:
122109
```bash
123110
#!/bin/bash
124111
module purge
125-
module load devel/singularity/3.4.2
126112
nextflow run nf-core/sarek -r 2.6.2 -profile cfc,test
127113
```
128114

@@ -177,9 +163,9 @@ Here are some useful commands for the Slurm scheduler.
177163
## Submitting custom jobs
178164

179165
> *Important note*: running scripts without containerizing them is never 100% reproducible, even when using conda environments.
180-
It is ok for testing, but talk to your group leader about the possibilities of containerizing the analysis or adding your scripts to a pipeline.
166+
It is ok to test pipelines, but talk to your group leader about the possibilities of containerizing the analysis or adding your scripts to a pipeline.
181167

182-
To run custom scripts (R or python or whatever you need) in the cluster, it is mandatory to use a dependency management system. This ensures at least some reproducibility for the results. You have two possibilities: use a clean conda environment and eexport it as an `environment.yml` file, or working on Rstudio and then using Rmaggedon.
168+
To run custom scripts (R or Python, or any other tool needed) in the cluster, it is mandatory to use a dependency management system. This ensures at least some reproducibility for the results. You have two possibilities: (1) use a clean conda environment and export it as an `environment.yml` file, or (2) working in Rstudio and then using Rmaggedon.
183169

184170
* *Using conda*: create a conda environment and install there all the necessary dependencies. Once you have them all, export the dependencies to a yml file containing the project code:
185171

@@ -211,7 +197,7 @@ srun -N 1 --ntasks-per-node=8 --mem=16G --time=12000 --pty bash
211197

212198
Change the resources as needed:
213199

214-
* N are the number of nodes
200+
* `-N` are the number of nodes
215201
* `--ntasks-per-node` are the number of cpus
216202
* `--mem` is the memory required
217203
* `--time` is the time required in seconds
@@ -229,7 +215,7 @@ You should see your job listed when running `squeue`.
229215

230216
### Submitting a bash script with `sbatch`
231217

232-
If you have a batch script, you can submit it to the cluster with the `sbatch` command.
218+
Please mind the [above-mentioned instructions](#submitting-nextflow-pipelines) for submitting Nextflow pipelines. If you have a batch script that is not a Nextflow pipeline run, you can submit it to the cluster with the `sbatch` command.
233219

234220
```bash
235221
sbatch <your_script.sh>

docs/markdown/clusters/tower.md

Lines changed: 13 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# Nextflow tower
22

3-
To be able to follow the Nextflow workflow rusn via tower, you can add Tower access credentials in your Nextflow configuration file (`~/.nextflow/config`) using the following snippet:
3+
To be able to follow the Nextflow workflow runs via tower, you can add Tower access credentials in your Nextflow configuration file (`~/.nextflow/config`) using the following snippet:
44

55
```console
66
tower {
@@ -11,12 +11,23 @@ tower {
1111
}
1212
```
1313

14+
Your access token can be created on [this page](http://cfgateway1.zdv.uni-tuebingen.de/tokens).
15+
16+
The workspace ID can be found on the organisation's Workspaces overview page. [Here](http://cfgateway1.zdv.uni-tuebingen.de/orgs/QBiC/workspaces) you can find QBiC's workspaces:
17+
18+
![workspaces](../../images/tower_workspaces.png)
19+
1420
To submit a pipeline to a different Workspace using the Nextflow command line tool, you can provide the workspace ID as an environment variable. For example
1521

1622
```console
1723
export TOWER_WORKSPACE_ID=000000000000000
1824
```
1925

20-
The workspace ID can be found on the organisation's Workspaces overview page.
26+
If you are outside of the University, access to the tower is only possible via VPN. When you started your run, you can now track its progress [here](http://cfgateway1.zdv.uni-tuebingen.de) after selecting your workspace and your run. Here is an example of what it looks like:
27+
28+
![example run](../../images/tower_run.png)
29+
![example run](../../images/tower_run1.png)
30+
![example run](../../images/tower_run2.png)
31+
You can select your run on the left. You will see the name of the run, your command line and the progress and stats of the run.
2132

2233
For more info on how to use tower please refere to the [Tower docs](https://help.tower.nf/).
Lines changed: 8 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -1,35 +1,32 @@
11
# 16S amplicon sequencing
22

3-
This is how to perform 16S amplicon analyses.
3+
This is how to perform 16S amplicon analyses. A video explanation of the biology, the bioinformatics problem and the analysis pipeline can be found for version 2.1.0 in the [nf-core bytesize talk 25](https://nf-co.re/events/2021/bytesize-25-nf-core-ampliseq).
44

55
## Ampliseq pipeline
66

77
To perform 16S amplicon sequencing analyses we employ the [nf-core/ampliseq](https://github.com/nf-core/ampliseq) pipeline.
88

99
### Quick start
1010

11-
* Latest stable release `-r 2.1.0`
11+
* Latest stable release `-r 2.1.1`
1212

1313
A typical command would look like this
1414

1515
```bash
16-
nextflow run nf-core/ampliseq -profile cfc -r 2.1.0 \
16+
nextflow run nf-core/ampliseq -profile cfc -r 2.1.1 \
1717
--input “data” \
1818
--FW_primer "GTGYCAGCMGCCGCGGTAA" \
1919
--RV_primer "GGACTACNVGGGTWTCTAAT" \
2020
--metadata "metadata.tsv" \
21-
--trunc_qmin 35 \
22-
--classifier_removeHash
21+
--trunc_qmin 35
2322
```
2423

25-
See [here](https://nf-co.re/ampliseq/1.2.0/parameters#manifest) the info on how to create the `metadata.tsv` file.
24+
Sequencing data can be analysed with the pipeline using a folder containing `.fastq.gz` files with [direct fastq input](https://nf-co.re/ampliseq/2.1.1/usage#direct-fastq-input) or [samplesheet input](https://nf-co.re/ampliseq/2.1.1/usage#samplesheet-input), also see [here](https://nf-co.re/ampliseq/2.1.1/parameters#input).
2625

27-
If data are distributed on multiple sequencing runs, please use `--multipleSequencingRuns` and note the different requirements for metadata file and folder structure in the [pipeline documentation](https://nf-co.re/ampliseq/1.2.0/parameters#multiplesequencingruns)
28-
29-
### Known bugs
26+
See [here](https://nf-co.re/ampliseq/2.1.1/parameters#metadata) the info on how to create the `metadata.tsv` file.
3027

31-
* All versions include a known bug that is why the `--classifier_removeHash` param should be used.
28+
If data are distributed on multiple sequencing runs, please use `--multipleSequencingRuns` and note the different requirements for metadata file and folder structure in the [pipeline documentation](https://nf-co.re/ampliseq/1.2.0/parameters#multiplesequencingruns)
3229

3330
## Reporting
3431

35-
There are no details about reporting yet.
32+
There are no details about reporting yet. Please refer to the [output documentation](https://nf-co.re/ampliseq/2.1.1/output).

docs/markdown/pipelines/pipeline_releases.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,4 +23,3 @@ These versions will be revised periodically and tested in our infrastructure bef
2323
| Bacterial assembly | [nf-core/bacass](https://nf-co.re/bacass/2.0.0) | 2.0.0 |
2424
| RNA fusion | [nf-core/rnafusion](https://nf-co.re/rnafusion/1.2.0) | 1.2.0 |
2525
| scRNAseq cellranger | [qbic-pipelines/cellranger](https://github.com/qbic-pipelines/cellranger) | 1.0.1 |
26-

0 commit comments

Comments
 (0)