issue_comments
3 rows where author_association = "NONE", issue = 320838184 and user = 5442433 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date)
issue 1
- Avoiding duplicate time coordinates when opening multiple files · 3 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
613525795 | https://github.com/pydata/xarray/issues/2108#issuecomment-613525795 | https://api.github.com/repos/pydata/xarray/issues/2108 | MDEyOklzc3VlQ29tbWVudDYxMzUyNTc5NQ== | brey 5442433 | 2020-04-14T15:55:05Z | 2020-04-14T15:55:05Z | NONE | I am adding here a comment to keep it alive. In fact, this is more complicated than it seems because in combining files with duplicate times one has to choose how to merge i.e keep first, keep last or even a combination of the two. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Avoiding duplicate time coordinates when opening multiple files 320838184 | |
389196623 | https://github.com/pydata/xarray/issues/2108#issuecomment-389196623 | https://api.github.com/repos/pydata/xarray/issues/2108 | MDEyOklzc3VlQ29tbWVudDM4OTE5NjYyMw== | brey 5442433 | 2018-05-15T14:53:38Z | 2018-05-15T14:53:38Z | NONE | Thanks @shoyer. Your approach works better (one line) plus is consistent with the xarray-pandas shared paradigm. Unfortunately, I can't spare the time to do the PR right now. I haven't done it before for xarray and it will require some time overhead. Maybe someone with more experience can oblige. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Avoiding duplicate time coordinates when opening multiple files 320838184 | |
387343836 | https://github.com/pydata/xarray/issues/2108#issuecomment-387343836 | https://api.github.com/repos/pydata/xarray/issues/2108 | MDEyOklzc3VlQ29tbWVudDM4NzM0MzgzNg== | brey 5442433 | 2018-05-08T09:33:14Z | 2018-05-08T09:33:14Z | NONE | To partially answer my issue, I came up with the following post-processing option
Maybe this can be integrated somehow... |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Avoiding duplicate time coordinates when opening multiple files 320838184 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 1