Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
S
snakemake-workflows
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Wiki
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Build
Pipelines
Jobs
Pipeline schedules
Artifacts
Deploy
Releases
Container Registry
Model registry
Operate
Environments
Monitor
Incidents
Analyze
Value stream analytics
Contributor analytics
CI/CD analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
FIJI
snakemake-workflows
Commits
c3b1a849
Commit
c3b1a849
authored
9 years ago
by
Christopher Schmied
Browse files
Options
Downloads
Patches
Plain Diff
Added annotation for .yaml file
parent
7c52fdd9
No related branches found
No related tags found
No related merge requests found
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
spim_registration/timelapse/tomancak_test_cluster.yaml
+153
-91
153 additions, 91 deletions
spim_registration/timelapse/tomancak_test_cluster.yaml
with
153 additions
and
91 deletions
spim_registration/timelapse/tomancak_test_cluster.yaml
+
153
−
91
View file @
c3b1a849
common
:
{
common
:
{
# directory that contains the bean shell scripts
# directory that contains the bean shell scripts and Snakefile
bsh_directory
:
"
/projects/pilot_spim/Christopher/snakemake-workflows/spim_registration/timelapse/"
,
bsh_directory
:
"
/projects/pilot_spim/Christopher/snakemake-workflows/spim_registration/timelapse/"
,
directory_cuda
:
"
/sw/users/schmied/cuda/"
,
# Directory that contains the cuda libraries
directory_cuda
:
"
/sw/users/schmied/cuda/"
,
# Directory that contains the current working Fiji
#fiji-app: "/sw/users/schmied/packages/2015-06-08_Fiji.app.cuda/ImageJ-linux64",
#fiji-app: "/sw/users/schmied/packages/2015-06-08_Fiji.app.cuda/ImageJ-linux64",
fiji-app
:
"
/sw/users/schmied/packages/2015-05-29_Fiji_2.3.9_SNAP.app.cuda/ImageJ-linux64"
,
fiji-app
:
"
/sw/users/schmied/packages/2015-05-29_Fiji_2.3.9_SNAP.app.cuda/ImageJ-linux64"
,
fiji-prefix
:
"
/sw/bin/xvfb-run
-a"
,
fiji-prefix
:
"
/sw/bin/xvfb-run
-a"
,
# calls xvfb for Fiji headless mode
first_xml_filename
:
"
Dual_Channel"
,
# xml file names without .xml suffix
hdf5_xml_filename
:
'
"hdf5_Dual_Channel"'
,
first_xml_filename
:
"
Dual_Channel"
,
# Name of the xml file for the .czi or .tif files
merged_xml
:
"
hdf5_Dual_Channel_merge"
,
hdf5_xml_filename
:
'
"hdf5_Dual_Channel"'
,
# Name of .xml file for the hdf5 data after resave_hdf5
ntimepoints
:
2
,
merged_xml
:
"
hdf5_Dual_Channel_merge"
,
# Name of .xml file after merge
angles
:
"
0,72,144,216,288"
,
# Describe the dataset
channels
:
"
green,red"
,
ntimepoints
:
2
,
# number of timepoints of dataset
illumination
:
"
0"
,
angles
:
"
0,72,144,216,288"
,
# angles
pixel_distance_x
:
'
0.28590'
,
channels
:
"
green,red"
,
# channels
pixel_distance_y
:
'
0.28590'
,
illumination
:
"
0"
,
# illuminations
pixel_distance_z
:
'
1.50000'
,
pixel_distance_x
:
'
0.28590'
,
# Manual calibration x
pixel_unit
:
"
um"
,
pixel_distance_y
:
'
0.28590'
,
# Manual calibration y
transformation_switch
:
"
timelapse_dublicate"
,
pixel_distance_z
:
'
1.50000'
,
# Manual calibration z
# transformation_switch: "timelapse",
pixel_unit
:
"
um"
,
# unit of manual calibration
# Use switches to decide which processing steps you need:
# transformation_switch: "timelapse" standard processing
# after timelapse registration directly goes into fusion, timelapse_dublicate
# "timelapse_dublicate" for dual channel processing one channel contains the beads
# dublicates transformations
transformation_switch
:
"
timelapse_dublicate"
,
# Switches between content based fusion and deconvoltion
# "deconvolution" > for deconvolution
# "fusion" > for content based fusion
fusion_switch
:
"
deconvolution"
fusion_switch
:
"
deconvolution"
# fusion_switch: "fusion"
}
}
define_xml_czi
:
{
define_xml_czi
:
{
first_czi
:
"
2015-02-20_LZ2_Stock48_Stock58.czi"
,
#
essential
first_czi
:
"
2015-02-20_LZ2_Stock48_Stock58.czi"
,
#
master .czi file
rotation_around
:
"
X-Axis"
,
rotation_around
:
"
X-Axis"
,
# axis of acquistion
bsh_file
:
"
define_czi.bsh"
bsh_file
:
"
define_czi.bsh"
# .bsh script for defining .czi file
}
}
define_xml_tif
:
{
define_xml_tif
:
{
image_file_pattern
:
'
img_TL{{t}}_Angle{{a}}.tif'
,
# essential
# file pattern of .tif files
imglib_container
:
'
"ArrayImg
(faster)"'
,
# for multi channel give spim_TL{tt}_Angle{a}_Channel{c}.tif
multiple_angles
:
'
"YES
(one
file
per
angle)"'
,
# # SPIM file pattern: for padded zeros use tt
multiple_channels
:
'
"NO
(one
channel)"'
,
image_file_pattern
:
'
img_TL{{t}}_Angle{{a}}.tif'
,
multiple_illumination_directions
:
'
"NO
(one
illumination
direction)"'
,
# Settings for ImageJ Opener
multiple_timepoints
:
'
"YES
(one
file
per
time-point)"'
,
type_of_dataset
:
'
"Image
Stacks
(ImageJ
Opener)"'
,
type_of_dataset
:
'
"Image
Stacks
(ImageJ
Opener)"'
,
bsh_file
:
"
define_tif_zip.bsh"
multiple_timepoints
:
'
"YES
(one
file
per
time-point)"'
,
# or NO (one time-point)
multiple_angles
:
'
"YES
(one
file
per
angle)"'
,
# or NO (one angle)
multiple_channels
:
'
"NO
(one
channel)"'
,
# or "\"NO (one channel)\""
multiple_illumination_directions
:
'
"NO
(one
illumination
direction)"'
,
# or YES (one file per illumination direction)
imglib_container
:
'
"ArrayImg
(faster)"'
,
# '"ArrayImg (faster)"'
bsh_file
:
"
define_tif_zip.bsh"
}
}
resave_hdf5
:
{
resave_hdf5
:
{
hdf5_chunk_sizes
:
'
"{{
{{32,32,4}},
{{32,32,4}},
{{16,16,16}},
{{16,16,16}}
}}"'
,
# data dependent
# Subsampling and resolution settings for hdf5: data dependent
subsampling_factors
:
'
"{{
{{1,1,1}},
{{2,2,1}},
{{4,4,1}},
{{8,8,1}}
}}"'
,
# data dependent
hdf5_chunk_sizes
:
'
"{{
{{32,32,4}},
{{32,32,4}},
{{16,16,16}},
{{16,16,16}}
}}"'
,
subsampling_factors
:
'
"{{
{{1,1,1}},
{{2,2,1}},
{{4,4,1}},
{{8,8,1}}
}}"'
,
# Standard settings for cluster processing
setups_per_partition
:
'
0'
,
timepoints_per_partition
:
'
1'
,
resave_timepoint
:
'
"All
Timepoints"'
,
resave_angle
:
'
"All
angles"'
,
resave_angle
:
'
"All
angles"'
,
resave_channel
:
'
"All
channels"'
,
resave_channel
:
'
"All
channels"'
,
resave_illumination
:
'
"All
illuminations"'
,
resave_illumination
:
'
"All
illuminations"'
,
resave_timepoint
:
'
"All
Timepoints"'
,
setups_per_partition
:
'
0'
,
timepoints_per_partition
:
'
1'
,
bsh_file
:
"
export.bsh"
bsh_file
:
"
export.bsh"
}
}
registration
:
{
registration
:
{
reg_process_channel
:
'
"Single
channel
(Select
from
List)"'
,
# essential '"All channels"'; '"Single channel (Select from List)"'
# reg_process_channel:
reg_processing_channel
:
'
"red"'
,
# essential
# # Single Channel: '"All channels"'
reg_interest_points_channel
:
'
"beads"'
,
# essential
# Dual Channel: '"All channels"'
type_of_detection
:
'
"Difference-of-Mean
(Integral
image
based)"'
,
# Difference-of-Gaussian
# Dual Channel one Channel contains beads: '"Single channel (Select from List)"'
reg_radius_1
:
'
2'
,
# essential
reg_process_channel
:
'
"Single
channel
(Select
from
List)"'
,
reg_radius_2
:
'
3'
,
# essential
# reg_processing_channel:
reg_threshold
:
'
0.005'
,
# essential
# Dual Channel setting for 1 Channel contains the beads
sigma
:
'
1.8'
,
# essetial
reg_processing_channel
:
'
"red"'
,
threshold_gaussian
:
'
0.0080'
,
# essential
# reg_interest_points_channel:
# Single Channel: '"beads"'
# Dual Channel: '"beads,beads"'
# Dual Channel: Channel does not contain the beads '"[DO NOT register this channel],beads"'
reg_interest_points_channel
:
'
"beads"'
,
# type of detection: '"Difference-of-Mean (Integral image based)"' or '"Difference-of-Gaussian"'
type_of_detection
:
'
"Difference-of-Mean
(Integral
image
based)"'
,
# Settings for Difference-of-Mean
# For multiple channels 'value1,value2' delimiter is ,
reg_radius_1
:
'
2'
,
reg_radius_2
:
'
3'
,
reg_threshold
:
'
0.005'
,
# Settings for Difference-of-Gaussian
# For multiple channels 'value1,value2' delimiter is ,
sigma
:
'
1.8'
,
threshold_gaussian
:
'
0.0080'
,
# Processing setting for Difference-of-Gaussian detection
# compute_on:
compute_on
:
'
"GPU
accurate
(Nvidia
CUDA
via
JNA)"'
,
compute_on
:
'
"GPU
accurate
(Nvidia
CUDA
via
JNA)"'
,
separableconvolution
:
'
"libSeparableConvolutionCUDALib.so"'
,
separableconvolution
:
'
"libSeparableConvolutionCUDALib.so"'
,
downsample_detection
:
"
No"
,
# Downsampling settings
downsample_detection
:
"
No"
,
# "No" or "Yes"
downsample_xy
:
'
"Match
Z
Resolution
(less
downsampling)"'
,
downsample_xy
:
'
"Match
Z
Resolution
(less
downsampling)"'
,
downsample_z
:
"
1x"
,
downsample_z
:
"
1x"
,
label_interest_points
:
'
"beads"'
,
# Standard Settings for bead based registration
label_interest_points
:
'
"beads"'
,
reg_process_timepoint
:
'
"Single
Timepoint
(Select
from
List)"'
,
reg_process_timepoint
:
'
"Single
Timepoint
(Select
from
List)"'
,
reg_process_angle
:
'
"All
angles"'
,
reg_process_angle
:
'
"All
angles"'
,
reg_process_illumination
:
'
"All
illuminations"'
,
reg_process_illumination
:
'
"All
illuminations"'
,
...
@@ -82,7 +117,7 @@ registration: {
...
@@ -82,7 +117,7 @@ registration: {
model_to_regularize_with
:
"
Rigid"
,
model_to_regularize_with
:
"
Rigid"
,
lambda
:
'
0.10'
,
lambda
:
'
0.10'
,
imglib_container
:
'
"ArrayImg
(faster)"'
,
imglib_container
:
'
"ArrayImg
(faster)"'
,
bsh_file
:
"
registration.bsh"
bsh_file
:
"
registration.bsh"
# .bsh script for registration
}
}
xml_merge
:
{
xml_merge
:
{
...
@@ -90,108 +125,135 @@ xml_merge: {
...
@@ -90,108 +125,135 @@ xml_merge: {
}
}
timelapse
:
{
timelapse
:
{
reference_timepoint
:
'
0'
,
# essential
reference_timepoint
:
'
0'
,
# Reference timepoint
# Standard settings for timelapse registration
type_of_registration_timelapse
:
'
"Match
against
one
reference
timepoint
(no
global
optimization)"'
,
type_of_registration_timelapse
:
'
"Match
against
one
reference
timepoint
(no
global
optimization)"'
,
timelapse_process_timepoints
:
'
"All
Timepoints"'
,
timelapse_process_timepoints
:
'
"All
Timepoints"'
,
bsh_file
:
"
timelapse_registration.bsh"
bsh_file
:
"
timelapse_registration.bsh"
}
}
dublicate_transformations
:
{
dublicate_transformations
:
{
source_dublication
:
"
red"
,
# essential for 1 usecase
# If dual channel processing and only one channel contains beads
target_dublication
:
"
green"
,
# essential for 1 usecase
# this allows you to dublicate the transformation for the
duplicate_which_transformations
:
'
"Replace
all
transformations"'
,
# channel that does not contain beas
bsh_file
:
"
dublicate_transformations.bsh"
source_dublication
:
"
red"
,
# source channel
target_dublication
:
"
green"
,
# target channel
duplicate_which_transformations
:
'
"Replace
all
transformations"'
,
# mode of dublication
bsh_file
:
"
dublicate_transformations.bsh"
# .bsh script for dublication
}
}
fusion
:
{
fusion
:
{
downsample
:
'
1'
,
# essential
# content based multiview fusion
minimal_x
:
'
220'
,
# essential maybe automate
# supports multi channel without new settings
minimal_y
:
'
40
'
,
#
essential maybe automate
downsample
:
'
1
'
,
#
set downsampling
minimal_z
:
'
-290'
,
# essential maybe automate
# Cropping parameters of full resolution
m
ax
imal_x
:
'
976'
,
# essential maybe automate
m
in
imal_x
:
'
220'
,
m
ax
imal_y
:
'
1892'
,
# essential maybe automate
m
in
imal_y
:
'
40'
,
m
ax
imal_z
:
'
472'
,
# essential maybe automate
m
in
imal_z
:
'
-290'
,
fused_image
:
'
"Append
to
current
XML
Project"'
,
maximal_x
:
'
976'
,
imglib2_container_fusion
:
'
"ArrayImg"'
,
maximal_y
:
'
1892'
,
imglib2_data_container
:
'
"ArrayImg
(faster)"'
,
maximal_z
:
'
472'
,
interpolation
:
'
"Linear
Interpolation"'
,
# fused_image: '"Append to current XML Project"', does not work yet
p
ixel_type
:
'
"16-bit
unsigned
integer
"'
,
p
rocess_timepoint
:
'
"Single
Timepoint
(Select
from
List)
"'
,
process_angle
:
'
"All
angles"'
,
process_angle
:
'
"All
angles"'
,
process_channel
:
'
"All
channels"'
,
process_channel
:
'
"All
channels"'
,
process_illumination
:
'
"All
illuminations"'
,
process_illumination
:
'
"All
illuminations"'
,
process_timepoint
:
'
"Single
Timepoint
(Select
from
List)"'
,
imglib2_container_fusion
:
'
"ArrayImg"'
,
interpolation
:
'
"Linear
Interpolation"'
,
pixel_type
:
'
"16-bit
unsigned
integer"'
,
imglib2_data_container
:
'
"ArrayImg
(faster)"'
,
process_views_in_paralell
:
'
"All"'
,
process_views_in_paralell
:
'
"All"'
,
xml_output
:
'
"Save
every
XML
with
user-provided
unique
id"'
,
xml_output
:
'
"Save
every
XML
with
user-provided
unique
id"'
,
bsh_file
:
"
fusion.bsh"
bsh_file
:
"
fusion.bsh"
}
}
external_transform
:
{
external_transform
:
{
# BUG
# Downsamples for deconvolution
# BUG: external transformation breaks .xml file
# channel setting: '"all_channels"'
# channel setting: '"all_channels"'
channel_setting
:
'
"green,red"'
,
channel_setting
:
'
"green,red"'
,
transform_timepoint
:
'
"All
Timepoints"'
,
transform_timepoint
:
'
"All
Timepoints"'
,
transform_angle
:
'
"All
angles"'
,
transform_angle
:
'
"All
angles"'
,
transform_channel
:
'
"All
channels"'
,
transform_channel
:
'
"All
channels"'
,
# illumination setting only one illumination side
transform_illumination
:
'
"All
illuminations"'
,
transform_illumination
:
'
"All
illuminations"'
,
apply_transformation
:
'
"Current
view
transformations
(appends
to
current
transforms)"'
,
apply_transformation
:
'
"Current
view
transformations
(appends
to
current
transforms)"'
,
define_mode_transform
:
'
"Matrix"'
,
define_mode_transform
:
'
"Matrix"'
,
matrix_transform
:
'
"0.5,
0.0,
0.0,
0.0,
0.0,
0.5,
0.0,
0.0,
0.0,
0.0,
0.5,
0.0"'
,
# essential
# Matrix for downsampling
matrix_transform
:
'
"0.5,
0.0,
0.0,
0.0,
0.0,
0.5,
0.0,
0.0,
0.0,
0.0,
0.5,
0.0"'
,
transformation
:
'
"Rigid"'
,
transformation
:
'
"Rigid"'
,
bsh_file
:
"
transform.bsh"
bsh_file
:
"
transform.bsh"
}
}
deconvolution
:
{
deconvolution
:
{
iterations
:
'
1'
,
# essential
iterations
:
'
1'
,
# number of iterations
minimal_x_deco
:
'
190'
,
# essential maybe automate
# Cropping parameters: take downsampling into account
minimal_y_deco
:
'
-16'
,
# essential maybe automate
minimal_x_deco
:
'
190'
,
minimal_z_deco
:
'
-348'
,
# essential maybe automate
minimal_y_deco
:
'
-16'
,
maximal_x_deco
:
'
1019'
,
# essential maybe automate
minimal_z_deco
:
'
-348'
,
maximal_y_deco
:
'
1941'
,
# essential maybe automate
maximal_x_deco
:
'
1019'
,
maximal_z_deco
:
'
486'
,
# essential maybe automate
maximal_y_deco
:
'
1941'
,
detections_to_extract_psf_for_channel
:
'
"beads,beads"'
,
maximal_z_deco
:
'
486'
,
# Channel settings for deconvolution
# Single Channel: '"beads"'
# Dual Channel: '"beads,beads"'
# Dual Channel one channel contains beads: '"[Same PSF as channel red],beads"'
detections_to_extract_psf_for_channel
:
'
"[Same
PSF
as
channel
red],beads"'
,
# Settings for GPU or CPU processing
# '"CPU (Java)"' or '"GPU (Nvidia CUDA via JNA)"'
compute_on
:
'
"GPU
(Nvidia
CUDA
via
JNA)"'
,
cudafourierconvolution
:
"
libFourierConvolutionCUDALib.so"
,
# GPU processing name of cuda library
# Standard settings for deconvolution
process_timepoint
:
'
"Single
Timepoint
(Select
from
List)"'
,
process_timepoint
:
'
"Single
Timepoint
(Select
from
List)"'
,
process_angle
:
'
"All
angles"'
,
process_angle
:
'
"All
angles"'
,
process_channel
:
'
"All
channels"'
,
process_channel
:
'
"All
channels"'
,
process_illumination
:
'
"All
illuminations"'
,
process_illumination
:
'
"All
illuminations"'
,
type_of_iteration
:
'
"Efficient
Bayesian
-
Optimization
I
(fast,
precise)"'
,
Tikhonov_parameter
:
'
0.0006'
,
Tikhonov_parameter
:
'
0.0006'
,
compute
:
'
"in
512x512x512
blocks"'
,
compute
:
'
"in
512x512x512
blocks"'
,
compute_on
:
'
"GPU
(Nvidia
CUDA
via
JNA)"'
,
cudafourierconvolution
:
"
libFourierConvolutionCUDALib.so"
,
imglib2_container
:
'
"ArrayImg"'
,
osem_acceleration
:
'
"1
(balanced)"'
,
osem_acceleration
:
'
"1
(balanced)"'
,
psf_estimation
:
'
"Extract
from
beads"'
,
psf_estimation
:
'
"Extract
from
beads"'
,
psf_size_x
:
'
19'
,
psf_size_x
:
'
19'
,
psf_size_y
:
'
19'
,
psf_size_y
:
'
19'
,
psf_size_z
:
'
25'
,
psf_size_z
:
'
25'
,
type_of_iteration
:
'
"Efficient
Bayesian
-
Optimization
I
(fast,
precise)
"'
,
imglib2_container
:
'
"ArrayImg
"'
,
bsh_file
:
"
deconvolution.bsh"
bsh_file
:
"
deconvolution.bsh"
}
}
hdf5_output
:
{
hdf5_output
:
{
# Will be obsolete
# writes new hdf5 dataset for fusion output: will be obsolete
# Naming pattern of output
# Single Channel: TP{t}_Chgreen_Ill0_Ang0,72,144,216,288.tif > Ch{name} is added here
# Dual Channel: TP{t}_Ch{0}_Ill0_Ang0,72,144,216,288.tif > Ch{name} is added here
output_image_file_pattern
:
'
TP{{t}}_Ch{{c}}_Ill0_Ang0,72,144,216,288.tif'
,
output_image_file_pattern
:
'
TP{{t}}_Ch{{c}}_Ill0_Ang0,72,144,216,288.tif'
,
# channel setting
output_multiple_channels
:
'
"YES
(one
file
per
channel)"'
,
# '"YES (one file per channel)"' or '"NO (one channel)"'
output_channels
:
"
green,red"
,
# .xml file names
output_xml
:
'
"fused_Dual_Channel"'
,
output_xml
:
'
"fused_Dual_Channel"'
,
output_hdf5_xml
:
'
"hdf5_fused_Dual_Channel"'
,
output_hdf5_xml
:
'
"hdf5_fused_Dual_Channel"'
,
output_
multiple_channels
:
'
"YES
(one
file
per
channel)"'
,
# "\"YES (one file per channel)\"" or "\"NO (one channel)\""
output_
timepoints
:
'
0-1'
,
# Timepoints format: '1-2'
output_timepoints
:
'
0-1'
,
# pixel size of output: take downsampling into account!
output_pixel_distance_x
:
0.28590
,
output_pixel_distance_x
:
0.28590
,
output_pixel_distance_y
:
0.28590
,
output_pixel_distance_y
:
0.28590
,
output_pixel_distance_z
:
0.28590
,
output_pixel_distance_z
:
0.28590
,
output_pixel_unit
:
'
um'
,
output_pixel_unit
:
'
um'
,
output_channels
:
"
green,red"
,
# give if 16Bit data or 32Bit data
output_data_type
:
"
16Bit"
,
# output of fusion is 16Bit, of deconvolution it is 32Bit
output_data_type
:
"
16Bit"
,
# "32Bit" or "16Bit"
# if data is 32Bit then the data is converted into 16Bit data
convert_32bit
:
'
"[Use
min/max
of
first
image
(might
saturate
intenities
over
time)]"'
,
convert_32bit
:
'
"[Use
min/max
of
first
image
(might
saturate
intenities
over
time)]"'
,
# subsampling and chunk size settings: dataset dependent
subsampling_output
:
'
"{{
{{1,1,1}},
{{2,2,2}},
{{4,4,4}},
{{8,8,8}}
}}"'
,
# data dependent
chunk_sizes_output
:
'
"{{
{{16,16,16}},
{{16,16,16}},
{{16,16,16}},
{{16,16,16}}
}}"'
,
# data dependent
# subsampling_output: '"{{ {{1,1,1}}, {{2,2,2}} }}"',
# chunk_sizes_output: '"{{ {{16,16,16}}, {{16,16,16}} }}"',
# Standard settings for hdf5_output
output_type_of_dataset
:
'
"Image
Stacks
(ImageJ
Opener)"'
,
output_type_of_dataset
:
'
"Image
Stacks
(ImageJ
Opener)"'
,
output_multiple_timepoints
:
'
"YES
(one
file
per
time-point)"'
,
output_multiple_timepoints
:
'
"YES
(one
file
per
time-point)"'
,
output_multiple_angles
:
'
"NO
(one
angle)"'
,
output_multiple_angles
:
'
"NO
(one
angle)"'
,
output_illumination_directions
:
'
"NO
(one
illumination
direction)"'
,
output_illumination_directions
:
'
"NO
(one
illumination
direction)"'
,
output_imglib_container
:
'
"ArrayImg
(faster)"'
,
output_imglib_container
:
'
"ArrayImg
(faster)"'
,
subsampling_output
:
'
"{{
{{1,1,1}},
{{2,2,2}},
{{4,4,4}},
{{8,8,8}}
}}"'
,
# data dependent
bsh_file_define
:
"
define_output.bsh"
,
# .bsh script for defining the dataset
chunk_sizes_output
:
'
"{{
{{16,16,16}},
{{16,16,16}},
{{16,16,16}},
{{16,16,16}}
}}"'
,
# data dependent
bsh_file_hdf5
:
"
export_output.bsh"
# .bsh script for resaving into hdf5
# subsampling_output: '"{{ {{1,1,1}}, {{2,2,2}} }}"',
# chunk_sizes_output: '"{{ {{16,16,16}}, {{16,16,16}} }}"',
bsh_file_define
:
"
define_output.bsh"
,
bsh_file_hdf5
:
"
export_output.bsh"
}
}
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment