issue_comments
1 row where issue = 172291585 and user = 5356122 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date)
issue 1
- align() should align chunks · 1 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
264510264 | https://github.com/pydata/xarray/issues/979#issuecomment-264510264 | https://api.github.com/repos/pydata/xarray/issues/979 | MDEyOklzc3VlQ29tbWVudDI2NDUxMDI2NA== | clarkfitzg 5356122 | 2016-12-02T17:23:46Z | 2016-12-02T17:23:46Z | MEMBER | As an end user, it would be really nice to not have to worry about chunks at all. I'd like to write the same code in xarray using Numpy and have it do the right thing in dask transparently. It seems like dask is moving in this direction (see Automatic blocksize for read_csv dask/dask#1147). Agree with @shoyer that these features belong in dask. |
{ "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
align() should align chunks 172291585 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 1