home / github

Menu
  • Search all tables
  • GraphQL API

issue_comments

Table actions
  • GraphQL API for issue_comments

9 rows where author_association = "MEMBER" and issue = 606683601 sorted by updated_at descending

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: reactions, created_at (date), updated_at (date)

user 6

  • shoyer 2
  • dcherian 2
  • TomNicholas 2
  • rabernat 1
  • jhamman 1
  • keewis 1

issue 1

  • xarray.open_mzar: open multiple zarr files (in parallel) · 9 ✖

author_association 1

  • MEMBER · 9 ✖
id html_url issue_url node_id user created_at updated_at ▲ author_association body reactions performed_via_github_app issue
651734845 https://github.com/pydata/xarray/pull/4003#issuecomment-651734845 https://api.github.com/repos/pydata/xarray/issues/4003 MDEyOklzc3VlQ29tbWVudDY1MTczNDg0NQ== keewis 14808389 2020-06-30T11:30:10Z 2020-06-30T11:30:10Z MEMBER

Since I am a novice at this, on my end, should I close this PR?

don't worry about that: we can close this PR when we merge #4187

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xarray.open_mzar: open multiple zarr files (in parallel) 606683601
651486921 https://github.com/pydata/xarray/pull/4003#issuecomment-651486921 https://api.github.com/repos/pydata/xarray/issues/4003 MDEyOklzc3VlQ29tbWVudDY1MTQ4NjkyMQ== shoyer 1217238 2020-06-30T02:40:28Z 2020-06-30T02:40:28Z MEMBER

Sure, I can move it, but I just wanted to make sure @Mikejmnez gets the credit for this PR

Yes, absolutely! As long as you preserve his original commits and add yours on top of them, both of you will be credited in the Git history. If you're writing a release note in whats-new.rst about the feature, please include both of your names in the credits.

{
    "total_count": 1,
    "+1": 1,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xarray.open_mzar: open multiple zarr files (in parallel) 606683601
651476633 https://github.com/pydata/xarray/pull/4003#issuecomment-651476633 https://api.github.com/repos/pydata/xarray/issues/4003 MDEyOklzc3VlQ29tbWVudDY1MTQ3NjYzMw== shoyer 1217238 2020-06-30T02:11:18Z 2020-06-30T02:11:18Z MEMBER

@weiji14 could you kindly reopen your new pull request against the main xarray repository? Your pull request is currently in Mikejmnez/xarray

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xarray.open_mzar: open multiple zarr files (in parallel) 606683601
620196044 https://github.com/pydata/xarray/pull/4003#issuecomment-620196044 https://api.github.com/repos/pydata/xarray/issues/4003 MDEyOklzc3VlQ29tbWVudDYyMDE5NjA0NA== dcherian 2448579 2020-04-27T19:47:38Z 2020-04-27T19:47:38Z MEMBER

IMO we should support zarr-store-1/ zarr-store-2/ zarr-file-store-3 but raise NotImplementedError for

zarr-store-1/ subdir/zarr-store-2

I don't know whether it would be easy to detect the glob pattern for the second example in all cases though.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xarray.open_mzar: open multiple zarr files (in parallel) 606683601
620176655 https://github.com/pydata/xarray/pull/4003#issuecomment-620176655 https://api.github.com/repos/pydata/xarray/issues/4003 MDEyOklzc3VlQ29tbWVudDYyMDE3NjY1NQ== rabernat 1197350 2020-04-27T19:09:33Z 2020-04-27T19:09:33Z MEMBER

I agree with everything Joe said. I'm fine with getting a NotImplemented error if I try to glob with zarr and open_mfdataset.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xarray.open_mzar: open multiple zarr files (in parallel) 606683601
620169860 https://github.com/pydata/xarray/pull/4003#issuecomment-620169860 https://api.github.com/repos/pydata/xarray/issues/4003 MDEyOklzc3VlQ29tbWVudDYyMDE2OTg2MA== jhamman 2443309 2020-04-27T18:56:10Z 2020-04-27T18:56:10Z MEMBER

+1 on deprecating open_zarr and moving to `open_dataset(..., engine='zarr')

I also agree that globing zarr stores is a tricky nut to crack. For the sake of simplicity, I'd like to suggest handling this functionaly as separate PRs. Given the heterogeneity in zarr storage options, I'm not sure its practical to support this behavior within Xarray but I'd be happy to discuss that in a separate issue.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xarray.open_mzar: open multiple zarr files (in parallel) 606683601
620151075 https://github.com/pydata/xarray/pull/4003#issuecomment-620151075 https://api.github.com/repos/pydata/xarray/issues/4003 MDEyOklzc3VlQ29tbWVudDYyMDE1MTA3NQ== TomNicholas 35968931 2020-04-27T18:19:41Z 2020-04-27T18:19:41Z MEMBER

@rabernat and @jhamman I expect you will want to weigh in on how best to handle this for zarr

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xarray.open_mzar: open multiple zarr files (in parallel) 606683601
619644606 https://github.com/pydata/xarray/pull/4003#issuecomment-619644606 https://api.github.com/repos/pydata/xarray/issues/4003 MDEyOklzc3VlQ29tbWVudDYxOTY0NDYwNg== TomNicholas 35968931 2020-04-26T23:53:31Z 2020-04-26T23:53:31Z MEMBER

+1 for having open_dataset and open_mfdataset as the main (ideally only) points of entry for users, which then delegate to different backend openers. That will keep the API neater, avoid duplicate code, and be easier to make into a completely general and extensible solution eventually.

{
    "total_count": 3,
    "+1": 3,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xarray.open_mzar: open multiple zarr files (in parallel) 606683601
619641890 https://github.com/pydata/xarray/pull/4003#issuecomment-619641890 https://api.github.com/repos/pydata/xarray/issues/4003 MDEyOklzc3VlQ29tbWVudDYxOTY0MTg5MA== dcherian 2448579 2020-04-26T23:30:46Z 2020-04-26T23:30:46Z MEMBER

I think the better way to do this would be to add a kwarg to open_dataset that specifies the backend to use. For e.g. xr.open_dataset(..., format="zarr").

This would then delegate to open_zarr or a new open_netcdf or open_rasterio as appropriate. Then open_mfdataset would just work for all these formats without requiring duplicate code.

cc @pydata/xarray for thoughts.

{
    "total_count": 3,
    "+1": 3,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xarray.open_mzar: open multiple zarr files (in parallel) 606683601

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [performed_via_github_app] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
    ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
    ON [issue_comments] ([user]);
Powered by Datasette · Queries took 1516.757ms · About: xarray-datasette