home / github

Menu
  • GraphQL API
  • Search all tables

issue_comments

Table actions
  • GraphQL API for issue_comments

4 rows where issue = 421029352 and user = 1197350 sorted by updated_at descending

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: created_at (date), updated_at (date)

user 1

  • rabernat · 4 ✖

issue 1

  • expose zarr caching from xarray · 4 ✖

author_association 1

  • MEMBER 4
id html_url issue_url node_id user created_at updated_at ▲ author_association body reactions performed_via_github_app issue
1246005938 https://github.com/pydata/xarray/issues/2812#issuecomment-1246005938 https://api.github.com/repos/pydata/xarray/issues/2812 IC_kwDOAMm_X85KRIqy rabernat 1197350 2022-09-13T22:18:31Z 2022-09-13T22:18:31Z MEMBER

Glad you got it working! So you're saying it does not work with open_zarr and does work with open_dataset(...engine='zarr')? Weird. We should deprecate open_zarr.

However, the behavior in Dask is strange. I think it is making each worker have its own cache and blowing up memory if I ask for a large cache.

Yes, I think I experienced that as well. I think the entire cache is serialized and passed around between workers.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  expose zarr caching from xarray 421029352
1243823078 https://github.com/pydata/xarray/issues/2812#issuecomment-1243823078 https://api.github.com/repos/pydata/xarray/issues/2812 IC_kwDOAMm_X85KIzvm rabernat 1197350 2022-09-12T14:25:39Z 2022-09-12T14:25:39Z MEMBER

I have successfully used the Zarr LRU cache with Xarray. You just have to initialize the Store object outside of Xarray and then pass it to open_zarr or open_dataset(store, engine="zarr").

Have you tried that?

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  expose zarr caching from xarray 421029352
472905515 https://github.com/pydata/xarray/issues/2812#issuecomment-472905515 https://api.github.com/repos/pydata/xarray/issues/2812 MDEyOklzc3VlQ29tbWVudDQ3MjkwNTUxNQ== rabernat 1197350 2019-03-14T15:02:22Z 2019-03-14T15:02:22Z MEMBER

I have created two PRs which attempt to provide zarr caching in different ways. I would welcome some advice on which one is a better approach.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  expose zarr caching from xarray 421029352
472871184 https://github.com/pydata/xarray/issues/2812#issuecomment-472871184 https://api.github.com/repos/pydata/xarray/issues/2812 MDEyOklzc3VlQ29tbWVudDQ3Mjg3MTE4NA== rabernat 1197350 2019-03-14T14:07:03Z 2019-03-14T14:07:03Z MEMBER

Or should we use xarray's own caching mechanism?

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  expose zarr caching from xarray 421029352

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [performed_via_github_app] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
    ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
    ON [issue_comments] ([user]);
Powered by Datasette · Queries took 74.368ms · About: xarray-datasette