issue_comments
8 rows where issue = 1247010680 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: reactions, created_at (date), updated_at (date)
issue 1
- Opening dataset without loading any indexes? · 8 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
1142380488 | https://github.com/pydata/xarray/issues/6633#issuecomment-1142380488 | https://api.github.com/repos/pydata/xarray/issues/6633 | IC_kwDOAMm_X85EF1fI | dcherian 2448579 | 2022-05-31T16:50:21Z | 2022-05-31T16:50:21Z | MEMBER | This would also fix #2233 |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Opening dataset without loading any indexes? 1247010680 | |
1137851771 | https://github.com/pydata/xarray/issues/6633#issuecomment-1137851771 | https://api.github.com/repos/pydata/xarray/issues/6633 | IC_kwDOAMm_X85D0j17 | rabernat 1197350 | 2022-05-25T21:10:44Z | 2022-05-25T21:10:44Z | MEMBER | Yes it is definitely a pathological example. 💣 But the fact remains that there are many cases where we just want to discover dataset contents as quickly as possible and want to avoid the cost of loading coordinates and creating indexes. |
{ "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Opening dataset without loading any indexes? 1247010680 | |
1137839614 | https://github.com/pydata/xarray/issues/6633#issuecomment-1137839614 | https://api.github.com/repos/pydata/xarray/issues/6633 | IC_kwDOAMm_X85D0g3- | shoyer 1217238 | 2022-05-25T20:55:14Z | 2022-05-25T20:55:14Z | MEMBER | Looking at this mur-sst dataset in particular, it stores time in chunks of size 5. That means fetching the 6443 time values requires 1288 separate HTTP requests -- no wonder it's so slow! If the time axis were instead stored in a single chunk of 51 KB, Xarray would only need 3 small size HTTP requests to load the lat, lon and time indexes, which would probably complete in a fraction of a second. That said, I agree that this would be nice to have in general. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Opening dataset without loading any indexes? 1247010680 | |
1137821786 | https://github.com/pydata/xarray/issues/6633#issuecomment-1137821786 | https://api.github.com/repos/pydata/xarray/issues/6633 | IC_kwDOAMm_X85D0cha | rabernat 1197350 | 2022-05-25T20:34:30Z | 2022-05-25T20:34:59Z | MEMBER | Here is an example that really highlights the performance cost of always loading dimension coordinates:
Obviously this example is pretty extreme. There are things that could be done to optimize it, etc. But it really highlights the costs of eagerly loading dimension coordinates. If I don't care about label-based indexing for this dataset, I would rather have my 17s back! :+1: to " |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Opening dataset without loading any indexes? 1247010680 | |
1137754031 | https://github.com/pydata/xarray/issues/6633#issuecomment-1137754031 | https://api.github.com/repos/pydata/xarray/issues/6633 | IC_kwDOAMm_X85D0L-v | shoyer 1217238 | 2022-05-25T19:12:40Z | 2022-05-25T19:12:40Z | MEMBER |
+1 this syntax makes sense to me! |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Opening dataset without loading any indexes? 1247010680 | |
1137748248 | https://github.com/pydata/xarray/issues/6633#issuecomment-1137748248 | https://api.github.com/repos/pydata/xarray/issues/6633 | IC_kwDOAMm_X85D0KkY | TomNicholas 35968931 | 2022-05-25T19:07:50Z | 2022-05-25T19:07:50Z | MEMBER | Thanks for replying both.
I'll have to defer to the others I tagged for the gory details. Perhaps one of them can cross-link to the specific issue they were having?
I would probably do |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Opening dataset without loading any indexes? 1247010680 | |
1137710350 | https://github.com/pydata/xarray/issues/6633#issuecomment-1137710350 | https://api.github.com/repos/pydata/xarray/issues/6633 | IC_kwDOAMm_X85D0BUO | benbovy 4160723 | 2022-05-25T18:47:14Z | 2022-05-25T18:47:14Z | MEMBER |
It might indeed be worth considering this case too in #6392. Maybe |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Opening dataset without loading any indexes? 1247010680 | |
1137572812 | https://github.com/pydata/xarray/issues/6633#issuecomment-1137572812 | https://api.github.com/repos/pydata/xarray/issues/6633 | IC_kwDOAMm_X85DzfvM | shoyer 1217238 | 2022-05-25T17:10:04Z | 2022-05-25T17:10:04Z | MEMBER | Early versions of Xarray used to have lazy loading of data for indexes, but we removed this for the sake of simplicity. In principle we could restore lazy indexes, but another option (post explicit index refactor) might be an option for opening a dataset without creating indexes for 1D coordinates along dimensions. Another way to solve this sort of challenges might be to load index data in parallel when using Dask. Right now I believe the data corresponding to indexes is always loaded eagerly, without using Dask. All that said -- Do you have a specific example where this has been problematic? In my experience it has been pretty reasonable to use xarray.Dataset objects for schema-like templates, even with index data needing to be loaded eagerly. Possibly another Zarr chunking scheme for your index data could be more efficient? |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Opening dataset without loading any indexes? 1247010680 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 5