issue_comments
7 rows where issue = 221387277 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date)
issue 1
- decode_cf() loads chunked arrays · 7 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
338019155 | https://github.com/pydata/xarray/issues/1372#issuecomment-338019155 | https://api.github.com/repos/pydata/xarray/issues/1372 | MDEyOklzc3VlQ29tbWVudDMzODAxOTE1NQ== | rabernat 1197350 | 2017-10-19T19:56:12Z | 2017-10-19T20:06:53Z | MEMBER | I just hit this issue. I tried to reproduce it with a synthetic dataset, as in @shoyer's example, but I couldn't. I can only reproduce it with data loaded from netcdf4 via open_mfdataset. I downloaded one year of air-temperature data from NARR: ftp://ftp.cdc.noaa.gov/Datasets/NARR/Dailies/pressure/ I load it this way (preprocessing is necessary to resolve conflict between If I try to decode cf, it returns instantly.
This is already a weird situation, since the data is not in memory, but it is not a dask array either. If I try to do anything beyond this with the data, it triggers eager computation. Even if I just call In my case, I could get around this problem if the preprocess function in |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
decode_cf() loads chunked arrays 221387277 | |
293751509 | https://github.com/pydata/xarray/issues/1372#issuecomment-293751509 | https://api.github.com/repos/pydata/xarray/issues/1372 | MDEyOklzc3VlQ29tbWVudDI5Mzc1MTUwOQ== | shoyer 1217238 | 2017-04-13T01:27:00Z | 2017-04-13T01:27:00Z | MEMBER |
Yes, that sounds right to me. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
decode_cf() loads chunked arrays 221387277 | |
293749896 | https://github.com/pydata/xarray/issues/1372#issuecomment-293749896 | https://api.github.com/repos/pydata/xarray/issues/1372 | MDEyOklzc3VlQ29tbWVudDI5Mzc0OTg5Ng== | mrocklin 306380 | 2017-04-13T01:15:19Z | 2017-04-13T01:15:19Z | MEMBER | Optimization has come up a bit recently. In other projects like dask-glm I've actually been thinking about just turning it off entirely. The spread of desires here is wide. I think it would be good to make it a bit more customizable so that different applications can more easily customize their optimization suite (see https://github.com/dask/dask/issues/2206) But for this application in particular presumably all of these already tasks |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
decode_cf() loads chunked arrays 221387277 | |
293748654 | https://github.com/pydata/xarray/issues/1372#issuecomment-293748654 | https://api.github.com/repos/pydata/xarray/issues/1372 | MDEyOklzc3VlQ29tbWVudDI5Mzc0ODY1NA== | shoyer 1217238 | 2017-04-13T01:05:42Z | 2017-04-13T01:06:17Z | MEMBER | Ah, so here's the thing: You might ask why this separate lazy compute machinery exists. The answer is that dask fails to optimize element-wise operations like See https://github.com/dask/dask/issues/746 for discussion and links to PRs about this. @jcrist had a solution that worked, but it slowed down every dask array operations by 20%, which wasn't a great win. I wonder if this is worth revisiting with a simpler, less general optimization pass that doesn't bother with broadcasting. See the subclasses of If we could optimize all these operations (and ideally chain them), then we could drop all the lazy loading stuff from xarray in favor of dask, which would be a real win. @mrocklin any thoughts? |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
decode_cf() loads chunked arrays 221387277 | |
293743835 | https://github.com/pydata/xarray/issues/1372#issuecomment-293743835 | https://api.github.com/repos/pydata/xarray/issues/1372 | MDEyOklzc3VlQ29tbWVudDI5Mzc0MzgzNQ== | shoyer 1217238 | 2017-04-13T00:27:32Z | 2017-04-13T00:27:32Z | MEMBER | A simple test case: ``` In [16]: ds = xr.Dataset({'foo': ('x', np.arange(100))}) In [17]: chunked = ds.chunk() In [18]: chunked.foo Out[18]: <xarray.DataArray 'foo' (x: 100)> dask.array<xarray-foo, shape=(100,), dtype=int64, chunksize=(100,)> Dimensions without coordinates: x In [19]: xr.decode_cf(chunked).foo Out[19]: <xarray.DataArray 'foo' (x: 100)> array([ 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99]) Dimensions without coordinates: x ``` |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
decode_cf() loads chunked arrays 221387277 | |
293740415 | https://github.com/pydata/xarray/issues/1372#issuecomment-293740415 | https://api.github.com/repos/pydata/xarray/issues/1372 | MDEyOklzc3VlQ29tbWVudDI5Mzc0MDQxNQ== | spencerahill 6200806 | 2017-04-13T00:02:56Z | 2017-04-13T00:02:56Z | CONTRIBUTOR | +1 we have come across this recently also in aospy |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
decode_cf() loads chunked arrays 221387277 | |
293725434 | https://github.com/pydata/xarray/issues/1372#issuecomment-293725434 | https://api.github.com/repos/pydata/xarray/issues/1372 | MDEyOklzc3VlQ29tbWVudDI5MzcyNTQzNA== | shoyer 1217238 | 2017-04-12T22:28:17Z | 2017-04-12T22:28:17Z | MEMBER |
|
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
decode_cf() loads chunked arrays 221387277 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 4