Skip to content

Commit b9a891c

Browse files
committed
Merge remote-tracking branch 'upstream/dev' into 3.0
2 parents 23453e8 + 3441df4 commit b9a891c

7 files changed

Lines changed: 14 additions & 53 deletions

File tree

conf/modules.config

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1105,7 +1105,7 @@ process{
11051105
publishDir = [
11061106
mode: params.publish_dir_mode,
11071107
path: { "${params.outdir}/variant_calling/tiddit/${meta.id}/" },
1108-
pattern: "*{vcf,vcf.gz}"
1108+
pattern: "*vcf.gz"
11091109
]
11101110
}
11111111
}

docs/usage.md

Lines changed: 3 additions & 42 deletions
Original file line numberDiff line numberDiff line change
@@ -680,36 +680,6 @@ Recent updates to Samtools have been introduced, which can speed-up performance
680680
The current workflow does not handle duplex UMIs (i.e. where opposite strands of a duplex molecule have been tagged with a different UMI), and best practices have been proposed to process this type of data.
681681
Both changes will be implemented in a future release.
682682
683-
## How to run sarek when no(t all) reference files are in igenomes
684-
685-
For common genomes, such as GRCh38 and GRCh37, the pipeline is shipped with (almost) all necessary reference files. However, sometimes it is necessary to use custom references for some or all files:
686-
687-
### No igenomes reference files are used
688-
689-
If none of your required genome files are in igenomes, `--igenomes_ignore` must be set to ignore any igenomes input and `--genome null`. The `fasta` file is the only required input file and must be provided to run the pipeline. All other possible reference file can be provided in addition. For details, see the paramter documentation.
690-
691-
Minimal example for custom genomes:
692-
693-
```
694-
nextflow run nf-core/sarek --genome null --igenomes_ignore --fasta <custom.fasta>
695-
```
696-
697-
### Overwrite specific reference files
698-
699-
If you don't want to use some of the provided reference genomes, they can be overwritten by either providing a new file or setting the respective file parameter to `false`, if it should be ignored:
700-
701-
Example for using a custom known indels file:
702-
703-
```
704-
nextflow run nf-core/sarek --known_indels <my_known_indels.vcf.gz> --genome GRCh38.GATK
705-
```
706-
707-
Example for not using known indels, but all other provided reference file:
708-
709-
```
710-
nextflow run nf-core/sarek --known_indels false --genome GRCh38.GATK
711-
```
712-
713683
### Where do the used reference genomes originate from
714684
715685
_under construction - help needed_
@@ -874,18 +844,9 @@ nextflow run download_cache.nf --cadd_cache </path/to/CADD/cache> --cadd_version
874844
Resource requests are difficult to generalize and are often dependent on input data size. Currently, the number of cpus and memory requested by default were adapted from tests on 5 ICGC paired whole-genome sequencing samples with approximately 40X and 80X depth.
875845
For targeted data analysis, this is overshooting by a lot. In this case resources for each process can be limited by either setting `--max_memory` and `-max_cpus` or tailoring the request by process name as described [here](#resource-requests). If you are using sarek for a certain data type regulary, and would like to make these requests available to others on your system, an institution-specific, pipeline-specific config file can be added [here](https://github.com/nf-core/configs/tree/master/conf/pipeline/sarek).
876846
877-
For mapping, sarek follows the parameter suggestions provided in this [paper](https://www.nature.com/articles/s41467-018-06159-4):
878-
879-
`-K 100000000` : for deterministic pipeline results, for more info see [here](https://github.com/CCDG/Pipeline-Standardization/issues/2)
880-
881-
`-Y`: force soft-clipping rather than default hard-clipping of supplementary alignments
882-
883-
In addition, currently, reads with tumor status in the sample sheet are mapped with a mismatch penalty of `-B 3`.
884-
885847
## Spark related issues
886848
887-
If you have problems running processes that make use of Spark such as `MarkDuplicates`.
888-
You are probably experiencing issues with the limit of open files in your system.
849+
If you have problems running processes that make use of Spark such, for instance, as `MarkDuplicates`, then that might be due to a limit on the number of simultaneously open files on your system.
889850
You can check your current limit by typing the following:
890851
891852
```bash
@@ -926,9 +887,9 @@ Currently, when running spark-based tools in combination with docker, it is requ
926887
927888
Sarek can process UMI-reads, using [fgbio](http://fulcrumgenomics.github.io/fgbio/tools/latest/) tools.
928889
929-
In order to use reads containing UMI tags as your initial input, you need to include `--umi_read_structure [structure]` in your parameters.
890+
In order to use reads containing UMI tags as your initial input, you need to include `--umi_read_structure <UMI_string>` in your parameters.
930891
931-
This will enable pre-processing of the reads and UMI consensus reads calling, which will then be used to continue the workflow from the mapping steps. For post-UMI processing depending on the experimental setup, duplicate marking and base quality recalibration can be skipped with [`--skip_tools`].
892+
This will enable pre-processing of the reads and UMI consensus reads calling, which will then be used to continue the workflow from the mapping steps. For post-UMI processing depending on the experimental setup, duplicate marking and base quality recalibration can be skipped with `--skip_tools`.
932893
933894
### UMI Read Structure
934895

modules.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -223,7 +223,7 @@
223223
"git_sha": "e745e167c1020928ef20ea1397b6b4d230681b4d"
224224
},
225225
"svdb/merge": {
226-
"git_sha": "313d76e00525c8e975dabce0c34973fd53c3f4dd"
226+
"git_sha": "6b011dd14821385c282e030c6d993d2116aed23f"
227227
},
228228
"tabix/bgziptabix": {
229229
"git_sha": "9dbaffff88013bc21460a989cde7a5aa172c2e0b"

modules/nf-core/modules/svdb/merge/main.nf

Lines changed: 7 additions & 5 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

modules/nf-core/modules/svdb/merge/meta.yml

Lines changed: 1 addition & 1 deletion
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

subworkflows/nf-core/variantcalling/tiddit/tiddit_somatic/main.nf

Lines changed: 0 additions & 2 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

tests/test_tools.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -989,7 +989,7 @@
989989
- path: results/variant_calling/tiddit/sample4_vs_sample3/sample4_vs_sample3.tiddit.ploidies.tab
990990
- path: results/variant_calling/tiddit/sample4_vs_sample3/sample4_vs_sample3.tiddit.tumor.vcf.gz
991991
- path: results/variant_calling/tiddit/sample4_vs_sample3/sample4_vs_sample3.tiddit.tumor.vcf.gz.tbi
992-
- path: results/variant_calling/tiddit/sample4_vs_sample3/sample4_vs_sample3.tiddit_sv_merge.vcf
992+
- path: results/variant_calling/tiddit/sample4_vs_sample3/sample4_vs_sample3.tiddit_sv_merge.vcf.gz
993993
- name: Run variant calling on germline sample with tiddit
994994
command: nextflow run main.nf -profile test,tools_germline,docker --tools tiddit
995995
tags:

0 commit comments

Comments
 (0)