-
Notifications
You must be signed in to change notification settings - Fork 54
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Two questions about converting larger than memory ND data into ome-zarr #255
Comments
The The Scaler class only has one way of downsampling for dask arrays, which uses ome-zarr-py/ome_zarr/writer.py Line 496 in 2c4d489
There was some discussion on the logic for that on the PR: #192 (comment) There is a PR currently open to fix a bug with the resizing of the edge tiles in a dask array at #244. There's also an issue raised about this at #237. No, there's no channels constructor helper methods. Just the example at https://ngff.openmicroscopy.org/latest/#omero-md. Apologies for the minimal docs there. |
Coincidentally, I have an immediate need to parameterize the |
Hi all, Thank you both for the info. We are trying with some smaller data first and hit a few technical snags. We'll work on them on our own and come back with more questions. Thanks again! |
Hi all, We ended up writing lazy downsampling code for these large datasets, as the current state of this project attempts to load the entire full-resolution array into memory to calculate the downsamples. Because we generate the data from our own microscopes and are now doing the downsampling on our own, it makes more sense to re-arrange the existing zarr store and then add the various OME format attributes. Otherwise, we are needlessly copying data between two zarr stores. On that note, addressing issue #258 would help us a lot, because we could then validate. Thanks for the guidance! I'll try to find a place to host the completed ome-zarr to see how viewing such a large dataset remotely works once everything is working. |
Hi all,
Thanks for the hard work on this package and overall on ome-ngff. We are very excited to learn that Dask arrays are now supported!
We have 4D data of shape
CZYX
, where typicallyc=17
anddtype=np.uint16
. The data is generated by iterative multiplexed light-sheet imaging. The 'zyx' dimension is the same for each channel and is usually large (ranging from[256,50000,50000]
to[1000,100000,100000]
). The full resolution data for each channel is stored as a Zarr array on disk and can be stacked together using Dask.Two questions regarding converting this data to ome-zarr:
Scaler()
function inome-zarr-py
performs lazy down-sampling.Thanks!
The text was updated successfully, but these errors were encountered: