issue_comments
7 rows where author_association = "MEMBER" and issue = 138332032 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date)
issue 1
- Array size changes following loading of numpy array · 7 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
193591506 | https://github.com/pydata/xarray/issues/783#issuecomment-193591506 | https://api.github.com/repos/pydata/xarray/issues/783 | MDEyOklzc3VlQ29tbWVudDE5MzU5MTUwNg== | mrocklin 306380 | 2016-03-08T03:44:36Z | 2016-03-08T03:44:36Z | MEMBER | Ah ha! Excellent. Thanks @shoyer . I'll give this a shot tomorrow (or perhaps ask @jcrist to look into it if he has time). |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Array size changes following loading of numpy array 138332032 | |
193576856 | https://github.com/pydata/xarray/issues/783#issuecomment-193576856 | https://api.github.com/repos/pydata/xarray/issues/783 | MDEyOklzc3VlQ29tbWVudDE5MzU3Njg1Ng== | shoyer 1217238 | 2016-03-08T02:56:28Z | 2016-03-08T02:56:49Z | MEMBER | As expected, the following all dask.array solution triggers this: ``` python dates = pd.date_range('2001-01-01', freq='D', periods=1000) sizes = pd.Series(dates, dates).resample('1M', how='count').values chunks = (tuple(sizes), (100,)) x = da.ones((3630, 100), chunks=chunks) assert x[240:270].shape == x[240:270].compute().shape AssertionError``` |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Array size changes following loading of numpy array 138332032 | |
193527245 | https://github.com/pydata/xarray/issues/783#issuecomment-193527245 | https://api.github.com/repos/pydata/xarray/issues/783 | MDEyOklzc3VlQ29tbWVudDE5MzUyNzI0NQ== | shoyer 1217238 | 2016-03-08T00:36:14Z | 2016-03-08T00:36:14Z | MEMBER | Something like this might work to generate pathological chunks for dask.array:
(I don't have xarray or dask installed on my work computer, but I could check this later) |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Array size changes following loading of numpy array 138332032 | |
193522753 | https://github.com/pydata/xarray/issues/783#issuecomment-193522753 | https://api.github.com/repos/pydata/xarray/issues/783 | MDEyOklzc3VlQ29tbWVudDE5MzUyMjc1Mw== | mrocklin 306380 | 2016-03-08T00:20:43Z | 2016-03-08T00:20:43Z | MEMBER | @shoyer perhaps you can help to translate the code within @pwolfram 's script (in particular the lines that I've highlighted) and say how xarray would use dask.array to accomplish this. I think this is a case where we each have some necessary expertise to resolve this issue. We probably need to work together to efficiently hunt down what's going on. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Array size changes following loading of numpy array 138332032 | |
193521326 | https://github.com/pydata/xarray/issues/783#issuecomment-193521326 | https://api.github.com/repos/pydata/xarray/issues/783 | MDEyOklzc3VlQ29tbWVudDE5MzUyMTMyNg== | shoyer 1217238 | 2016-03-08T00:17:00Z | 2016-03-08T00:17:00Z | MEMBER | If you don't specify a chunksize, xarray should use each file as a full "chunk". So it would probably be useful to know the shapes of each array you are loading with To be clear, all the logic implementing the chunking and indexing code for xarray objects containing dask arrays lives inside dask.array itself, not in our xarray wrapper (which is pretty thin). This doesn't make this any less of an issue for you, but I'm pretty sure (and I think @mrocklin agrees) that the bug here in probably in the dask.array layer. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Array size changes following loading of numpy array 138332032 | |
193501447 | https://github.com/pydata/xarray/issues/783#issuecomment-193501447 | https://api.github.com/repos/pydata/xarray/issues/783 | MDEyOklzc3VlQ29tbWVudDE5MzUwMTQ0Nw== | mrocklin 306380 | 2016-03-07T23:22:46Z | 2016-03-07T23:22:46Z | MEMBER | It looks like the issue is in these lines:
I'm confused by the chunksize change from 21 to 23. In straight dask.array I'm unable to reproduce this problem, although obviously I'm doing something differently here than how xarray does things. ``` python In [1]: import dask.array as da x In [2]: x = da.ones((3630, 100), chunks=(21, 100)) In [3]: y = x[730:830, :] In [4]: y.shape Out[4]: (30, 100) In [5]: y.compute().shape Out[5]: (30, 100) In [6]: y.chunks Out[6]: ((21, 9), (100,)) ``` It would be awesome if you all could produce a failing example with just dask.array. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Array size changes following loading of numpy array 138332032 | |
192048274 | https://github.com/pydata/xarray/issues/783#issuecomment-192048274 | https://api.github.com/repos/pydata/xarray/issues/783 | MDEyOklzc3VlQ29tbWVudDE5MjA0ODI3NA== | shoyer 1217238 | 2016-03-04T01:28:23Z | 2016-03-04T01:28:23Z | MEMBER | This does look very strange. I'm guessing it's a dask.array bug (cc @mrocklin). Can you make a reproducible example? If so, we'll probably be able to figure this out. How do you make this data? Tracking this sort of thing down is a good motivation for an eager-evaluation mode in dask.array... (https://github.com/dask/dask/issues/292) |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Array size changes following loading of numpy array 138332032 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 2