It was rejected as being unpythonic[1] , even though the base functionality to save a particular data frame is already present.
Can what dask is doing, be adapted to a simple case scenario of saving a workspace snapshot?
[1] https://github.com/pydata/pandas/issues/12381#issuecomment-1...
I wonder if this would work if the dask arrays are not equal in length, for example if the files were time series of unequal duration.
Also, are there any plans for dask to support distributed numpy functions requiring kernel computation at the array boundaries? For example scipy.signal.lfilt? I believe it would require ghosting or further inter-dask-array communication that is not yet present.
> Reduction speed: The computation of normalized temperature, z, took a surprisingly long time. I’d like to look into what is holding up that computation.