issues
1 row where user = 47371188 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date), closed_at (date)
id | node_id | number | title | user | state | locked | assignee | milestone | comments | created_at | updated_at ▲ | closed_at | author_association | active_lock_reason | draft | pull_request | body | reactions | performed_via_github_app | state_reason | repo | type |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
488547784 | MDU6SXNzdWU0ODg1NDc3ODQ= | 3277 | xarray, chunking and rolling operation adds chunking along new dimension (previously worked) | p-d-moore 47371188 | closed | 0 | 5 | 2019-09-03T11:25:23Z | 2021-03-26T19:50:49Z | 2021-03-26T19:50:49Z | NONE | I was testing the latest version of xarray (0.12.3) from the conda-forge channel and this broke some code I had. Under the defaults installation not using conda-forge (xarray=0.12.1), the following code works correctly with desired output: Test code```python import pandas as pd import xarray as xr import numpy as np s_date = '1990-01-01' e_date = '2019-05-01' days = pd.date_range(start=s_date, end=e_date, freq='B', name='day') items = pd.Index([str(i) for i in range(300)], name = 'item') dat = xr.DataArray(np.random.rand(len(days), len(items)), coords=[days, items]) dat_chunk = dat.chunk({'item': 20}) dat_mean = dat_chunk.rolling(day=10).mean() print(dat_chunk) print(' ') print(dat_mean) dat_std_avg = dat_mean.rolling(day=250).std() print(' ') print(dat_std_avg) ``` Output (correct) with xarray=0.12.1 - note the chunksizes``` <xarray.DataArray (day: 7653, item: 300)> dask.array<shape=(7653, 300), dtype=float64, chunksize=(7653, 20)> Coordinates: * day (day) datetime64[ns] 1990-01-01 1990-01-02 ... 2019-05-01 * item (item) object '0' '1' '2' '3' '4' ... '295' '296' '297' '298' '299' <xarray.DataArray '_trim-8c9287bf114d61cb3ad74780465cd19f' (day: 7653, item: 300)> dask.array<shape=(7653, 300), dtype=float64, chunksize=(7653, 20)> Coordinates: * day (day) datetime64[ns] 1990-01-01 1990-01-02 ... 2019-05-01 * item (item) object '0' '1' '2' '3' '4' ... '295' '296' '297' '298' '299' <xarray.DataArray '_trim-2ee90b6c2f29f71a7798a204a4ad3305' (day: 7653, item: 300)> dask.array<shape=(7653, 300), dtype=float64, chunksize=(7653, 20)> Coordinates: * day (day) datetime64[ns] 1990-01-01 1990-01-02 ... 2019-05-01 * item (item) object '0' '1' '2' '3' '4' ... '295' '296' '297' '298' '299' ``` Output (now failing) with xarray=0.12.3 (note the chunksizes)``` <xarray.DataArray (day: 7653, item: 300)> dask.array<shape=(7653, 300), dtype=float64, chunksize=(7653, 20)> Coordinates: * day (day) datetime64[ns] 1990-01-01 1990-01-02 ... 2019-05-01 * item (item) object '0' '1' '2' '3' '4' ... '295' '296' '297' '298' '299' <xarray.DataArray (day: 7653, item: 300)> dask.array<shape=(7653, 300), dtype=float64, chunksize=(5, 20)> Coordinates: * day (day) datetime64[ns] 1990-01-01 1990-01-02 ... 2019-05-01 * item (item) object '0' '1' '2' '3' '4' ... '295' '296' '297' '298' '299' ValueError Traceback (most recent call last) ... ValueError: For window size 250, every chunk should be larger than 125, but the smallest chunk size is 5. Rechunk your array with a larger chunk size or a chunk size that more evenly divides the shape of your array. ``` Problem DescriptionUsing dask + rolling + xarray=0.12.3 appears to add undesirable chunking in a new dimension which was not the case previously using xarray=0.12.1 This additional chunking made the the queuing of a further rolling operation fail with a ValueError. This (at the very least) makes queuing dask based delayed operations difficult when multiple rolling operations are used. Output of
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/3277/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issues] ( [id] INTEGER PRIMARY KEY, [node_id] TEXT, [number] INTEGER, [title] TEXT, [user] INTEGER REFERENCES [users]([id]), [state] TEXT, [locked] INTEGER, [assignee] INTEGER REFERENCES [users]([id]), [milestone] INTEGER REFERENCES [milestones]([id]), [comments] INTEGER, [created_at] TEXT, [updated_at] TEXT, [closed_at] TEXT, [author_association] TEXT, [active_lock_reason] TEXT, [draft] INTEGER, [pull_request] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [state_reason] TEXT, [repo] INTEGER REFERENCES [repos]([id]), [type] TEXT ); CREATE INDEX [idx_issues_repo] ON [issues] ([repo]); CREATE INDEX [idx_issues_milestone] ON [issues] ([milestone]); CREATE INDEX [idx_issues_assignee] ON [issues] ([assignee]); CREATE INDEX [idx_issues_user] ON [issues] ([user]);