issues: 824917345
This data as json
id | node_id | number | title | user | state | locked | assignee | milestone | comments | created_at | updated_at | closed_at | author_association | active_lock_reason | draft | pull_request | body | reactions | performed_via_github_app | state_reason | repo | type |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
824917345 | MDU6SXNzdWU4MjQ5MTczNDU= | 5010 | DataArrays inside apply_ufunc with dask=parallelized | 20629530 | closed | 0 | 3 | 2021-03-08T20:19:41Z | 2021-03-08T20:37:15Z | 2021-03-08T20:35:01Z | CONTRIBUTOR | Is your feature request related to a problem? Please describe.
Currently, when using apply_ufunc with Describe the solution you'd like I'd want to profit from both the tools of xarray and the power of dask parallelization. I'd like to be able to do something like this: ```python3 def func(da): """Example of an operation not (easily) possible with numpy.""" return da.groupby('time').mean() xr.apply_ufunc(
da,
func,
input_core_dims=[['time']],
pass_xr=True,
dask='parallelized'
)
Describe alternatives you've considered The alternative is to reduce the size of the datasets (looping on other dimensions), but that defeats the purpose of dask. Another alternative I am currently testing, is to add a layer between apply_ufunc and the |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/5010/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | 13221727 | issue |