Often we have many arrays stored on disk that we want to stack together and think of as one large array. This is common with geospatial data in which we might have many HDF5/NetCDF files on disk, one for every day, but we want to do operations that span multiple days.
To solve this problem we use the functions da.stack
and da.concatenate
.
We stack many existing Dask arrays into a new array, creating a new dimension as we go.
>>> import dask.array as da
>>> data = [from_array(np.ones((4, 4)), chunks=(2, 2))
... for i in range(3)] # A small stack of dask arrays
>>> x = da.stack(data, axis=0)
>>> x.shape
(3, 4, 4)
>>> da.stack(data, axis=1).shape
(4, 3, 4)
>>> da.stack(data, axis=-1).shape
(4, 4, 3)
This creates a new dimension with length equal to the number of slices
We concatenate existing arrays into a new array, extending them along an existing dimension
>>> import dask.array as da
>>> import numpy as np
>>> data = [from_array(np.ones((4, 4)), chunks=(2, 2))
... for i in range(3)] # small stack of dask arrays
>>> x = da.concatenate(data, axis=0)
>>> x.shape
(12, 4)
>>> da.concatenate(data, axis=1).shape
(4, 12)