diff --git a/README.md b/README.md index e5d5b8d..018def7 100644 --- a/README.md +++ b/README.md @@ -18,6 +18,7 @@ Takes TIF images (tiled or prestitched) and outputs a validated BIDS Microscopy - Python >= 3.11 - Lightsheet data: - Raw Ultramicroscope Blaze OME TIFF files (include `blaze` in the acquisition tag) + - can be 2D or 3D TIFF files - Prestitched TIFF files (include `prestitched` in the acquisition tag) @@ -63,10 +64,8 @@ or for snakemake<8.0, use: snakemake -c all --use-singularity ``` -Note: if you run the workflow on a system with large memory, you will need to set the heap size for the stitching and fusion rules. This can be done with e.g.: `--set-resources bigstitcher_spark_stitching:mem_mb=60000 bigstitcher_spark_fusion:mem_mb=100000` +Note: if you run the workflow on a system with large memory, you will need to set the heap size for the stitching and fusion rules. This can be done with e.g.: `--set-resources bigstitcher_stitching:mem_mb=60000 bigstitcher_fusion:mem_mb=100000` 7. If you want to run the workflow using a batch job submission server, please see the executor plugins here: https://snakemake.github.io/snakemake-plugin-catalog/ - -Alternate usage of this workflow (making use of conda) is described in the [Snakemake Workflow Catalog](https://snakemake.github.io/snakemake-workflow-catalog?repo=khanlab/SPIMprep). diff --git a/config/config.yml b/config/config.yml index edc16ba..9858eb7 100644 --- a/config/config.yml +++ b/config/config.yml @@ -1,6 +1,5 @@ datasets: 'config/datasets.tsv' - root: 'bids' # can use a s3:// or gcs:// prefix to write output to cloud storage work: 'work' @@ -13,6 +12,7 @@ cores_per_rule: 32 #import wildcards: tilex, tiley, channel, zslice (and prefix - unused) import_blaze: raw_tif_pattern: "{prefix}_Blaze[{tilex} x {tiley}]_C{channel}_xyz-Table Z{zslice}.ome.tif" + raw_tif_pattern_zstack: "{prefix}_Blaze[{tilex} x {tiley}]_C{channel}.ome.tif" intensity_rescaling: 0.5 #raw images seem to be at the upper end of uint16 (over-saturated) -- causes wrapping issues when adjusting with flatfield correction etc. this rescales the raw data as it imports it.. import_prestitched: @@ -36,12 +36,10 @@ bigstitcher: downsample_in_x: 4 downsample_in_y: 4 downsample_in_z: 1 - method: "phase_corr" #unused - methods: #unused + methods: #unused, only for reference phase_corr: "Phase Correlation" optical_flow: "Lucas-Kanade" filter_pairwise_shifts: - enabled: 1 #unused min_r: 0.7 max_shift_total: 50 global_optimization: @@ -64,7 +62,7 @@ bigstitcher: block_size_factor_z: 32 ome_zarr: - desc: sparkstitchedflatcorr + desc: stitchedflatcorr max_downsampling_layers: 5 # e.g. 4 levels: { 0: orig, 1: ds2, 2: ds4, 3: ds8, 4: ds16} rechunk_size: #z, y, x - 1 @@ -155,5 +153,5 @@ report: containers: - spimprep: 'docker://khanlab/spimprep-deps:main' + spimprep: 'docker://khanlab/spimprep-deps:v0.1.0' diff --git a/resources/qc/ff_html_temp.html b/resources/qc/ff_html_temp.html index 658b26d..ba6ee7a 100644 --- a/resources/qc/ff_html_temp.html +++ b/resources/qc/ff_html_temp.html @@ -43,13 +43,13 @@
Slice-{{ image.slice }}
Slice-{{ image.slice }}