Replies: 1 comment
-
This seems like a bug given the documentation of open_dataset:
Can you open an issue with a reproducible example please? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Often the chunk sizes in
zarr
(recommended1Mb
uncompressed: https://zarr.readthedocs.io/en/stable/tutorial.html) are smaller than recommended for efficient computation with dask (100-1000Mb
) https://blog.dask.org/2021/11/02/choosing-dask-chunk-sizes.In dask
da.from_zarr
chunksize="auto"
chooses appropriate chunk sizes for dask by combining multiple zarr chunks.However this isn't the default behaviour of the
"auto"
inxarray.open_zarr
which uses the defaultzarr
chunksize.In order to rechunk sensibly in xarray, it's necessary to calculate new chunk sizes. But it is possible to do so by specifying the number of chunks to merge over? I have the following helper function, but this seems a common enough use case to be a standard feature?
eg
Beta Was this translation helpful? Give feedback.
All reactions