home / github

Menu
  • Search all tables
  • GraphQL API

issue_comments

Table actions
  • GraphQL API for issue_comments

7 rows where issue = 758606082 and user = 14808389 sorted by updated_at descending

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: reactions, created_at (date), updated_at (date)

user 1

  • keewis · 7 ✖

issue 1

  • xr.DataArray.from_dask_dataframe feature · 7 ✖

author_association 1

  • MEMBER 7
id html_url issue_url node_id user created_at updated_at ▲ author_association body reactions performed_via_github_app issue
818269258 https://github.com/pydata/xarray/pull/4659#issuecomment-818269258 https://api.github.com/repos/pydata/xarray/issues/4659 MDEyOklzc3VlQ29tbWVudDgxODI2OTI1OA== keewis 14808389 2021-04-12T21:56:59Z 2021-04-12T21:56:59Z MEMBER

this should be ready for review

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xr.DataArray.from_dask_dataframe feature 758606082
818182598 https://github.com/pydata/xarray/pull/4659#issuecomment-818182598 https://api.github.com/repos/pydata/xarray/issues/4659 MDEyOklzc3VlQ29tbWVudDgxODE4MjU5OA== keewis 14808389 2021-04-12T20:30:23Z 2021-04-12T21:56:33Z MEMBER

@AyrtonB, I took the liberty of pushing the changes I had in mind to your branch, using a adapted version of your docstring. The only thing that should be missing is to figure out if it's possible to reduce the number of computes to 2 instead of n_columns + 1.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xr.DataArray.from_dask_dataframe feature 758606082
811388476 https://github.com/pydata/xarray/pull/4659#issuecomment-811388476 https://api.github.com/repos/pydata/xarray/issues/4659 MDEyOklzc3VlQ29tbWVudDgxMTM4ODQ3Ng== keewis 14808389 2021-03-31T19:40:51Z 2021-03-31T19:40:51Z MEMBER

@pydata/xarray, any opinion on the API design?

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xr.DataArray.from_dask_dataframe feature 758606082
798989229 https://github.com/pydata/xarray/pull/4659#issuecomment-798989229 https://api.github.com/repos/pydata/xarray/issues/4659 MDEyOklzc3VlQ29tbWVudDc5ODk4OTIyOQ== keewis 14808389 2021-03-14T22:10:00Z 2021-03-14T22:10:00Z MEMBER

I don't think there is a lot left to decide: we want to keep the conversion logic in from_dask_dataframe and maybe helper functions, and I think we should mirror the pandas integration as close as possible (which means we need a Dataset.from_dask_dataframe and a DataArray.from_dask_series class method).

The only thing I think is left to figure out is how to best compute the chunk sizes with as few computations of dask (as defined by raise_if_dask_computes) as possible.

cc @dcherian

{
    "total_count": 1,
    "+1": 1,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xr.DataArray.from_dask_dataframe feature 758606082
740041249 https://github.com/pydata/xarray/pull/4659#issuecomment-740041249 https://api.github.com/repos/pydata/xarray/issues/4659 MDEyOklzc3VlQ29tbWVudDc0MDA0MTI0OQ== keewis 14808389 2020-12-07T16:50:03Z 2020-12-07T16:51:34Z MEMBER

there's a few things to fix in pycompat for this to work: first of all, import dask.dataframe before accessing dask.dataframe.core.DataFrame. We should also move the assignment to dask_dataframe_type to its own try / except block since it's possible to have dask.array but not dask.dataframe installed. And the reason for the ImportError you got is that we need a value for dask_dataframe_type if there was a ImportError. I'm thinking of something like this: ```python try: import dask.dataframe

dask_dataframe_type = (dask.dataframe.core.DataFrame,)

except ImportError: dask_dataframe_type = () ```

{
    "total_count": 1,
    "+1": 1,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xr.DataArray.from_dask_dataframe feature 758606082
740006353 https://github.com/pydata/xarray/pull/4659#issuecomment-740006353 https://api.github.com/repos/pydata/xarray/issues/4659 MDEyOklzc3VlQ29tbWVudDc0MDAwNjM1Mw== keewis 14808389 2020-12-07T15:55:12Z 2020-12-07T15:55:12Z MEMBER

sorry, it is indeed called dask_array_compat. Looking closer, you probably won't be able to use that. Instead, I'd advise to do a local import (for an example see Dataset.to_dask_dataframe). For the change in variable.py I would use the same pattern as for pycompat.dask_array_type, so if dask.dataframe is not available dask_dataframe_type should be ().

{
    "total_count": 1,
    "+1": 1,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xr.DataArray.from_dask_dataframe feature 758606082
739994871 https://github.com/pydata/xarray/pull/4659#issuecomment-739994871 https://api.github.com/repos/pydata/xarray/issues/4659 MDEyOklzc3VlQ29tbWVudDczOTk5NDg3MQ== keewis 14808389 2020-12-07T15:36:57Z 2020-12-07T15:42:22Z MEMBER

you can just decorate tests that require dask with requires_dask and they will be skipped automatically if dask is not installed

Edit: actually, you seem to import dask in some modules, which is not what we want. We usually use either the dask_compat module or pycompat.dask_array_type to work around that.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  xr.DataArray.from_dask_dataframe feature 758606082

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [performed_via_github_app] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
    ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
    ON [issue_comments] ([user]);
Powered by Datasette · Queries took 115.087ms · About: xarray-datasette