Skip to content
Snippets Groups Projects
Commit c3b1a849 authored by Christopher Schmied's avatar Christopher Schmied
Browse files

Added annotation for .yaml file

parent 7c52fdd9
No related branches found
No related tags found
No related merge requests found
common: { common: {
# directory that contains the bean shell scripts # directory that contains the bean shell scripts and Snakefile
bsh_directory: "/projects/pilot_spim/Christopher/snakemake-workflows/spim_registration/timelapse/", bsh_directory: "/projects/pilot_spim/Christopher/snakemake-workflows/spim_registration/timelapse/",
directory_cuda: "/sw/users/schmied/cuda/", # Directory that contains the cuda libraries
directory_cuda: "/sw/users/schmied/cuda/",
# Directory that contains the current working Fiji
#fiji-app: "/sw/users/schmied/packages/2015-06-08_Fiji.app.cuda/ImageJ-linux64", #fiji-app: "/sw/users/schmied/packages/2015-06-08_Fiji.app.cuda/ImageJ-linux64",
fiji-app: "/sw/users/schmied/packages/2015-05-29_Fiji_2.3.9_SNAP.app.cuda/ImageJ-linux64", fiji-app: "/sw/users/schmied/packages/2015-05-29_Fiji_2.3.9_SNAP.app.cuda/ImageJ-linux64",
fiji-prefix: "/sw/bin/xvfb-run -a", fiji-prefix: "/sw/bin/xvfb-run -a", # calls xvfb for Fiji headless mode
first_xml_filename: "Dual_Channel", # xml file names without .xml suffix
hdf5_xml_filename: '"hdf5_Dual_Channel"', first_xml_filename: "Dual_Channel", # Name of the xml file for the .czi or .tif files
merged_xml: "hdf5_Dual_Channel_merge", hdf5_xml_filename: '"hdf5_Dual_Channel"', # Name of .xml file for the hdf5 data after resave_hdf5
ntimepoints: 2, merged_xml: "hdf5_Dual_Channel_merge", # Name of .xml file after merge
angles: "0,72,144,216,288", # Describe the dataset
channels: "green,red", ntimepoints: 2, # number of timepoints of dataset
illumination: "0", angles: "0,72,144,216,288", # angles
pixel_distance_x: '0.28590', channels: "green,red", # channels
pixel_distance_y: '0.28590', illumination: "0", # illuminations
pixel_distance_z: '1.50000', pixel_distance_x: '0.28590', # Manual calibration x
pixel_unit: "um", pixel_distance_y: '0.28590', # Manual calibration y
transformation_switch: "timelapse_dublicate", pixel_distance_z: '1.50000', # Manual calibration z
# transformation_switch: "timelapse", pixel_unit: "um", # unit of manual calibration
# Use switches to decide which processing steps you need:
# transformation_switch: "timelapse" standard processing
# after timelapse registration directly goes into fusion, timelapse_dublicate
# "timelapse_dublicate" for dual channel processing one channel contains the beads
# dublicates transformations
transformation_switch: "timelapse_dublicate",
# Switches between content based fusion and deconvoltion
# "deconvolution" > for deconvolution
# "fusion" > for content based fusion
fusion_switch: "deconvolution" fusion_switch: "deconvolution"
# fusion_switch: "fusion"
} }
define_xml_czi: { define_xml_czi: {
first_czi: "2015-02-20_LZ2_Stock48_Stock58.czi", # essential first_czi: "2015-02-20_LZ2_Stock48_Stock58.czi", # master .czi file
rotation_around: "X-Axis", rotation_around: "X-Axis", # axis of acquistion
bsh_file: "define_czi.bsh" bsh_file: "define_czi.bsh" # .bsh script for defining .czi file
} }
define_xml_tif: { define_xml_tif: {
image_file_pattern: 'img_TL{{t}}_Angle{{a}}.tif', # essential # file pattern of .tif files
imglib_container: '"ArrayImg (faster)"', # for multi channel give spim_TL{tt}_Angle{a}_Channel{c}.tif
multiple_angles: '"YES (one file per angle)"', # # SPIM file pattern: for padded zeros use tt
multiple_channels: '"NO (one channel)"', image_file_pattern: 'img_TL{{t}}_Angle{{a}}.tif',
multiple_illumination_directions: '"NO (one illumination direction)"', # Settings for ImageJ Opener
multiple_timepoints: '"YES (one file per time-point)"',
type_of_dataset: '"Image Stacks (ImageJ Opener)"', type_of_dataset: '"Image Stacks (ImageJ Opener)"',
bsh_file: "define_tif_zip.bsh" multiple_timepoints: '"YES (one file per time-point)"', # or NO (one time-point)
multiple_angles: '"YES (one file per angle)"', # or NO (one angle)
multiple_channels: '"NO (one channel)"', # or "\"NO (one channel)\""
multiple_illumination_directions: '"NO (one illumination direction)"', # or YES (one file per illumination direction)
imglib_container: '"ArrayImg (faster)"', # '"ArrayImg (faster)"'
bsh_file: "define_tif_zip.bsh"
} }
resave_hdf5: { resave_hdf5: {
hdf5_chunk_sizes: '"{{ {{32,32,4}}, {{32,32,4}}, {{16,16,16}}, {{16,16,16}} }}"', # data dependent # Subsampling and resolution settings for hdf5: data dependent
subsampling_factors: '"{{ {{1,1,1}}, {{2,2,1}}, {{4,4,1}}, {{8,8,1}} }}"', # data dependent hdf5_chunk_sizes: '"{{ {{32,32,4}}, {{32,32,4}}, {{16,16,16}}, {{16,16,16}} }}"',
subsampling_factors: '"{{ {{1,1,1}}, {{2,2,1}}, {{4,4,1}}, {{8,8,1}} }}"',
# Standard settings for cluster processing
setups_per_partition: '0',
timepoints_per_partition: '1',
resave_timepoint: '"All Timepoints"',
resave_angle: '"All angles"', resave_angle: '"All angles"',
resave_channel: '"All channels"', resave_channel: '"All channels"',
resave_illumination: '"All illuminations"', resave_illumination: '"All illuminations"',
resave_timepoint: '"All Timepoints"',
setups_per_partition: '0',
timepoints_per_partition: '1',
bsh_file: "export.bsh" bsh_file: "export.bsh"
} }
registration: { registration: {
reg_process_channel: '"Single channel (Select from List)"', # essential '"All channels"'; '"Single channel (Select from List)"' # reg_process_channel:
reg_processing_channel: '"red"', # essential # # Single Channel: '"All channels"'
reg_interest_points_channel: '"beads"', # essential # Dual Channel: '"All channels"'
type_of_detection: '"Difference-of-Mean (Integral image based)"', # Difference-of-Gaussian # Dual Channel one Channel contains beads: '"Single channel (Select from List)"'
reg_radius_1: '2', # essential reg_process_channel: '"Single channel (Select from List)"',
reg_radius_2: '3', # essential # reg_processing_channel:
reg_threshold: '0.005', # essential # Dual Channel setting for 1 Channel contains the beads
sigma: '1.8', # essetial reg_processing_channel: '"red"',
threshold_gaussian: '0.0080', # essential # reg_interest_points_channel:
# Single Channel: '"beads"'
# Dual Channel: '"beads,beads"'
# Dual Channel: Channel does not contain the beads '"[DO NOT register this channel],beads"'
reg_interest_points_channel: '"beads"',
# type of detection: '"Difference-of-Mean (Integral image based)"' or '"Difference-of-Gaussian"'
type_of_detection: '"Difference-of-Mean (Integral image based)"',
# Settings for Difference-of-Mean
# For multiple channels 'value1,value2' delimiter is ,
reg_radius_1: '2',
reg_radius_2: '3',
reg_threshold: '0.005',
# Settings for Difference-of-Gaussian
# For multiple channels 'value1,value2' delimiter is ,
sigma: '1.8',
threshold_gaussian: '0.0080',
# Processing setting for Difference-of-Gaussian detection
# compute_on:
compute_on: '"GPU accurate (Nvidia CUDA via JNA)"', compute_on: '"GPU accurate (Nvidia CUDA via JNA)"',
separableconvolution: '"libSeparableConvolutionCUDALib.so"', separableconvolution: '"libSeparableConvolutionCUDALib.so"',
downsample_detection: "No", # Downsampling settings
downsample_detection: "No", # "No" or "Yes"
downsample_xy: '"Match Z Resolution (less downsampling)"', downsample_xy: '"Match Z Resolution (less downsampling)"',
downsample_z: "1x", downsample_z: "1x",
label_interest_points: '"beads"', # Standard Settings for bead based registration
label_interest_points: '"beads"',
reg_process_timepoint: '"Single Timepoint (Select from List)"', reg_process_timepoint: '"Single Timepoint (Select from List)"',
reg_process_angle: '"All angles"', reg_process_angle: '"All angles"',
reg_process_illumination: '"All illuminations"', reg_process_illumination: '"All illuminations"',
...@@ -82,7 +117,7 @@ registration: { ...@@ -82,7 +117,7 @@ registration: {
model_to_regularize_with: "Rigid", model_to_regularize_with: "Rigid",
lambda: '0.10', lambda: '0.10',
imglib_container: '"ArrayImg (faster)"', imglib_container: '"ArrayImg (faster)"',
bsh_file: "registration.bsh" bsh_file: "registration.bsh" # .bsh script for registration
} }
xml_merge: { xml_merge: {
...@@ -90,108 +125,135 @@ xml_merge: { ...@@ -90,108 +125,135 @@ xml_merge: {
} }
timelapse: { timelapse: {
reference_timepoint: '0', # essential reference_timepoint: '0', # Reference timepoint
# Standard settings for timelapse registration
type_of_registration_timelapse: '"Match against one reference timepoint (no global optimization)"', type_of_registration_timelapse: '"Match against one reference timepoint (no global optimization)"',
timelapse_process_timepoints: '"All Timepoints"', timelapse_process_timepoints: '"All Timepoints"',
bsh_file: "timelapse_registration.bsh" bsh_file: "timelapse_registration.bsh"
} }
dublicate_transformations: { dublicate_transformations: {
source_dublication: "red", # essential for 1 usecase # If dual channel processing and only one channel contains beads
target_dublication: "green", # essential for 1 usecase # this allows you to dublicate the transformation for the
duplicate_which_transformations: '"Replace all transformations"', # channel that does not contain beas
bsh_file: "dublicate_transformations.bsh" source_dublication: "red", # source channel
target_dublication: "green", # target channel
duplicate_which_transformations: '"Replace all transformations"', # mode of dublication
bsh_file: "dublicate_transformations.bsh" # .bsh script for dublication
} }
fusion: { fusion: {
downsample: '1', # essential # content based multiview fusion
minimal_x: '220', # essential maybe automate # supports multi channel without new settings
minimal_y: '40', # essential maybe automate downsample: '1', # set downsampling
minimal_z: '-290', # essential maybe automate # Cropping parameters of full resolution
maximal_x: '976', # essential maybe automate minimal_x: '220',
maximal_y: '1892', # essential maybe automate minimal_y: '40',
maximal_z: '472', # essential maybe automate minimal_z: '-290',
fused_image: '"Append to current XML Project"', maximal_x: '976',
imglib2_container_fusion: '"ArrayImg"', maximal_y: '1892',
imglib2_data_container: '"ArrayImg (faster)"', maximal_z: '472',
interpolation: '"Linear Interpolation"', # fused_image: '"Append to current XML Project"', does not work yet
pixel_type: '"16-bit unsigned integer"', process_timepoint: '"Single Timepoint (Select from List)"',
process_angle: '"All angles"', process_angle: '"All angles"',
process_channel: '"All channels"', process_channel: '"All channels"',
process_illumination: '"All illuminations"', process_illumination: '"All illuminations"',
process_timepoint: '"Single Timepoint (Select from List)"', imglib2_container_fusion: '"ArrayImg"',
interpolation: '"Linear Interpolation"',
pixel_type: '"16-bit unsigned integer"',
imglib2_data_container: '"ArrayImg (faster)"',
process_views_in_paralell: '"All"', process_views_in_paralell: '"All"',
xml_output: '"Save every XML with user-provided unique id"', xml_output: '"Save every XML with user-provided unique id"',
bsh_file: "fusion.bsh" bsh_file: "fusion.bsh"
} }
external_transform: { external_transform: {
# BUG # Downsamples for deconvolution
# BUG: external transformation breaks .xml file
# channel setting: '"all_channels"' # channel setting: '"all_channels"'
channel_setting: '"green,red"', channel_setting: '"green,red"',
transform_timepoint: '"All Timepoints"', transform_timepoint: '"All Timepoints"',
transform_angle: '"All angles"', transform_angle: '"All angles"',
transform_channel: '"All channels"', transform_channel: '"All channels"',
# illumination setting only one illumination side
transform_illumination: '"All illuminations"', transform_illumination: '"All illuminations"',
apply_transformation: '"Current view transformations (appends to current transforms)"', apply_transformation: '"Current view transformations (appends to current transforms)"',
define_mode_transform: '"Matrix"', define_mode_transform: '"Matrix"',
matrix_transform: '"0.5, 0.0, 0.0, 0.0, 0.0, 0.5, 0.0, 0.0, 0.0, 0.0, 0.5, 0.0"', # essential # Matrix for downsampling
matrix_transform: '"0.5, 0.0, 0.0, 0.0, 0.0, 0.5, 0.0, 0.0, 0.0, 0.0, 0.5, 0.0"',
transformation: '"Rigid"', transformation: '"Rigid"',
bsh_file: "transform.bsh" bsh_file: "transform.bsh"
} }
deconvolution: { deconvolution: {
iterations: '1', # essential iterations: '1', # number of iterations
minimal_x_deco: '190', # essential maybe automate # Cropping parameters: take downsampling into account
minimal_y_deco: '-16', # essential maybe automate minimal_x_deco: '190',
minimal_z_deco: '-348', # essential maybe automate minimal_y_deco: '-16',
maximal_x_deco: '1019', # essential maybe automate minimal_z_deco: '-348',
maximal_y_deco: '1941', # essential maybe automate maximal_x_deco: '1019',
maximal_z_deco: '486', # essential maybe automate maximal_y_deco: '1941',
detections_to_extract_psf_for_channel: '"beads,beads"', maximal_z_deco: '486',
# Channel settings for deconvolution
# Single Channel: '"beads"'
# Dual Channel: '"beads,beads"'
# Dual Channel one channel contains beads: '"[Same PSF as channel red],beads"'
detections_to_extract_psf_for_channel: '"[Same PSF as channel red],beads"',
# Settings for GPU or CPU processing
# '"CPU (Java)"' or '"GPU (Nvidia CUDA via JNA)"'
compute_on: '"GPU (Nvidia CUDA via JNA)"',
cudafourierconvolution: "libFourierConvolutionCUDALib.so", # GPU processing name of cuda library
# Standard settings for deconvolution
process_timepoint: '"Single Timepoint (Select from List)"', process_timepoint: '"Single Timepoint (Select from List)"',
process_angle: '"All angles"', process_angle: '"All angles"',
process_channel: '"All channels"', process_channel: '"All channels"',
process_illumination: '"All illuminations"', process_illumination: '"All illuminations"',
type_of_iteration: '"Efficient Bayesian - Optimization I (fast, precise)"',
Tikhonov_parameter: '0.0006', Tikhonov_parameter: '0.0006',
compute: '"in 512x512x512 blocks"', compute: '"in 512x512x512 blocks"',
compute_on: '"GPU (Nvidia CUDA via JNA)"',
cudafourierconvolution: "libFourierConvolutionCUDALib.so",
imglib2_container: '"ArrayImg"',
osem_acceleration: '"1 (balanced)"', osem_acceleration: '"1 (balanced)"',
psf_estimation: '"Extract from beads"', psf_estimation: '"Extract from beads"',
psf_size_x: '19', psf_size_x: '19',
psf_size_y: '19', psf_size_y: '19',
psf_size_z: '25', psf_size_z: '25',
type_of_iteration: '"Efficient Bayesian - Optimization I (fast, precise)"', imglib2_container: '"ArrayImg"',
bsh_file: "deconvolution.bsh" bsh_file: "deconvolution.bsh"
} }
hdf5_output: { hdf5_output: {
# Will be obsolete # writes new hdf5 dataset for fusion output: will be obsolete
# Naming pattern of output
# Single Channel: TP{t}_Chgreen_Ill0_Ang0,72,144,216,288.tif > Ch{name} is added here
# Dual Channel: TP{t}_Ch{0}_Ill0_Ang0,72,144,216,288.tif > Ch{name} is added here
output_image_file_pattern: 'TP{{t}}_Ch{{c}}_Ill0_Ang0,72,144,216,288.tif', output_image_file_pattern: 'TP{{t}}_Ch{{c}}_Ill0_Ang0,72,144,216,288.tif',
# channel setting
output_multiple_channels: '"YES (one file per channel)"', # '"YES (one file per channel)"' or '"NO (one channel)"'
output_channels: "green,red",
# .xml file names
output_xml: '"fused_Dual_Channel"', output_xml: '"fused_Dual_Channel"',
output_hdf5_xml: '"hdf5_fused_Dual_Channel"', output_hdf5_xml: '"hdf5_fused_Dual_Channel"',
output_multiple_channels: '"YES (one file per channel)"', # "\"YES (one file per channel)\"" or "\"NO (one channel)\"" output_timepoints: '0-1', # Timepoints format: '1-2'
output_timepoints: '0-1', # pixel size of output: take downsampling into account!
output_pixel_distance_x: 0.28590, output_pixel_distance_x: 0.28590,
output_pixel_distance_y: 0.28590, output_pixel_distance_y: 0.28590,
output_pixel_distance_z: 0.28590, output_pixel_distance_z: 0.28590,
output_pixel_unit: 'um', output_pixel_unit: 'um',
output_channels: "green,red", # give if 16Bit data or 32Bit data
output_data_type: "16Bit", # output of fusion is 16Bit, of deconvolution it is 32Bit
output_data_type: "16Bit", # "32Bit" or "16Bit"
# if data is 32Bit then the data is converted into 16Bit data
convert_32bit: '"[Use min/max of first image (might saturate intenities over time)]"', convert_32bit: '"[Use min/max of first image (might saturate intenities over time)]"',
# subsampling and chunk size settings: dataset dependent
subsampling_output: '"{{ {{1,1,1}}, {{2,2,2}}, {{4,4,4}}, {{8,8,8}} }}"', # data dependent
chunk_sizes_output: '"{{ {{16,16,16}}, {{16,16,16}}, {{16,16,16}}, {{16,16,16}} }}"', # data dependent
# subsampling_output: '"{{ {{1,1,1}}, {{2,2,2}} }}"',
# chunk_sizes_output: '"{{ {{16,16,16}}, {{16,16,16}} }}"',
# Standard settings for hdf5_output
output_type_of_dataset: '"Image Stacks (ImageJ Opener)"', output_type_of_dataset: '"Image Stacks (ImageJ Opener)"',
output_multiple_timepoints: '"YES (one file per time-point)"', output_multiple_timepoints: '"YES (one file per time-point)"',
output_multiple_angles: '"NO (one angle)"', output_multiple_angles: '"NO (one angle)"',
output_illumination_directions: '"NO (one illumination direction)"', output_illumination_directions: '"NO (one illumination direction)"',
output_imglib_container: '"ArrayImg (faster)"', output_imglib_container: '"ArrayImg (faster)"',
subsampling_output: '"{{ {{1,1,1}}, {{2,2,2}}, {{4,4,4}}, {{8,8,8}} }}"', # data dependent bsh_file_define: "define_output.bsh", # .bsh script for defining the dataset
chunk_sizes_output: '"{{ {{16,16,16}}, {{16,16,16}}, {{16,16,16}}, {{16,16,16}} }}"', # data dependent bsh_file_hdf5: "export_output.bsh" # .bsh script for resaving into hdf5
# subsampling_output: '"{{ {{1,1,1}}, {{2,2,2}} }}"',
# chunk_sizes_output: '"{{ {{16,16,16}}, {{16,16,16}} }}"',
bsh_file_define: "define_output.bsh",
bsh_file_hdf5: "export_output.bsh"
} }
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment