home / github

Menu
  • Search all tables
  • GraphQL API

issue_comments

Table actions
  • GraphQL API for issue_comments

6 rows where issue = 453964049 sorted by updated_at descending

✖
✖

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: reactions, created_at (date), updated_at (date)

user 2

  • shoyer 3
  • djhoese 3

author_association 2

  • CONTRIBUTOR 3
  • MEMBER 3

issue 1

  • Fix 'to_masked_array' computing dask arrays twice · 6 ✖
id html_url issue_url node_id user created_at updated_at ▲ author_association body reactions performed_via_github_app issue
505531367 https://github.com/pydata/xarray/pull/3006#issuecomment-505531367 https://api.github.com/repos/pydata/xarray/issues/3006 MDEyOklzc3VlQ29tbWVudDUwNTUzMTM2Nw== shoyer 1217238 2019-06-25T16:56:19Z 2019-06-25T16:56:19Z MEMBER

hopefully out this week! I am going to wait another day or two for someone else to look at https://github.com/pydata/xarray/pull/3040/ and then will probably be issuing a release.

{
    "total_count": 2,
    "+1": 1,
    "-1": 0,
    "laugh": 0,
    "hooray": 1,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Fix 'to_masked_array' computing dask arrays twice 453964049
505530269 https://github.com/pydata/xarray/pull/3006#issuecomment-505530269 https://api.github.com/repos/pydata/xarray/issues/3006 MDEyOklzc3VlQ29tbWVudDUwNTUzMDI2OQ== djhoese 1828519 2019-06-25T16:53:34Z 2019-06-25T16:53:34Z CONTRIBUTOR

@shoyer Any idea when there might be another release of xarray where this fix will be included? I'm teaching a tutorial at SciPy this year that is effected by this bug. Learners are starting to prepare for the tutorials and I'd like if they could have this fix before the day of the tutorial.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Fix 'to_masked_array' computing dask arrays twice 453964049
500411377 https://github.com/pydata/xarray/pull/3006#issuecomment-500411377 https://api.github.com/repos/pydata/xarray/issues/3006 MDEyOklzc3VlQ29tbWVudDUwMDQxMTM3Nw== shoyer 1217238 2019-06-10T13:12:37Z 2019-06-10T13:12:37Z MEMBER

@djhoese It looks like something broke in the development version of dask. But that shouldn't hold up merging your fix here...

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Fix 'to_masked_array' computing dask arrays twice 453964049
500400954 https://github.com/pydata/xarray/pull/3006#issuecomment-500400954 https://api.github.com/repos/pydata/xarray/issues/3006 MDEyOklzc3VlQ29tbWVudDUwMDQwMDk1NA== djhoese 1828519 2019-06-10T12:36:55Z 2019-06-10T12:36:55Z CONTRIBUTOR

@shoyer Makes sense. Any idea what's up with the travis test? It doesn't look like it is from my changes.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Fix 'to_masked_array' computing dask arrays twice 453964049
500290315 https://github.com/pydata/xarray/pull/3006#issuecomment-500290315 https://api.github.com/repos/pydata/xarray/issues/3006 MDEyOklzc3VlQ29tbWVudDUwMDI5MDMxNQ== shoyer 1217238 2019-06-10T04:46:57Z 2019-06-10T04:46:57Z MEMBER

I'd be happy to merge this fix. I think it's a vestige of when we used to always cached computations.

I'd also love to have more comprehensive test coverage, and I like the look of your custom scheduler. But I'm not sure it's worth adding that many lines of test logic for what is essentially a one line fix.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Fix 'to_masked_array' computing dask arrays twice 453964049
500264231 https://github.com/pydata/xarray/pull/3006#issuecomment-500264231 https://api.github.com/repos/pydata/xarray/issues/3006 MDEyOklzc3VlQ29tbWVudDUwMDI2NDIzMQ== djhoese 1828519 2019-06-10T01:36:00Z 2019-06-10T01:36:00Z CONTRIBUTOR

In my own tests I've been using the following custom scheduler with dask.config.set(scheduler=CustomScheduler()) to point out what code is computing the array when I don't want it to:

``` class CustomScheduler(object): """Custom dask scheduler that raises an exception if dask is computed too many times."""

def __init__(self, max_computes=1):
    """Set starting and maximum compute counts."""
    self.max_computes = max_computes
    self.total_computes = 0

def __call__(self, dsk, keys, **kwargs):
    """Compute dask task and keep track of number of times we do so."""
    import dask
    self.total_computes += 1
    if self.total_computes > self.max_computes:
        raise RuntimeError("Too many dask computations were scheduled: {}".format(self.total_computes))
    return dask.get(dsk, keys, **kwargs)

```

Does something like this exist in the xarray tests? If not, I could add it then add a dask test to the DataArray tests.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Fix 'to_masked_array' computing dask arrays twice 453964049

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [performed_via_github_app] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
    ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
    ON [issue_comments] ([user]);
Powered by Datasette · Queries took 2403.144ms · About: xarray-datasette
  • Sort ascending
  • Sort descending
  • Facet by this
  • Hide this column
  • Show all columns
  • Show not-blank rows