home / github

Menu
  • Search all tables
  • GraphQL API

issue_comments

Table actions
  • GraphQL API for issue_comments

1 row where author_association = "MEMBER", issue = 305663416 and user = 1217238 sorted by updated_at descending

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: created_at (date), updated_at (date)

user 1

  • shoyer · 1 ✖

issue 1

  • Canonical approach for new vectorized functions · 1 ✖

author_association 1

  • MEMBER · 1 ✖
id html_url issue_url node_id user created_at updated_at ▲ author_association body reactions performed_via_github_app issue
373511743 https://github.com/pydata/xarray/issues/1992#issuecomment-373511743 https://api.github.com/repos/pydata/xarray/issues/1992 MDEyOklzc3VlQ29tbWVudDM3MzUxMTc0Mw== shoyer 1217238 2018-03-15T20:26:51Z 2018-03-15T20:26:51Z MEMBER

I started numbagg with the intent of using it for xarray, but never got around to finishing it up. I think it would be perfectly fine to add as an optional dependency. Mostly we are using bottleneck instead because it's already out there and is slightly more complete, but certainly writing new aggregation functions with numbagg rather than bottleneck would be much easier.

One thing I was also excited about doing with numbagg was writing fast grouped aggregation functions, e.g., to make xarray's groupby().mean() comparable in performance to pandas. I had some promising preliminary results, but now I can't find the notebook where I wrote them!

Out of curiosity, I ran the benchmark suite from the numbagg README again. Both Numba and bottleneck have been significantly improved over the past three years, but Numba now beats bottleneck across the board: ``` %timeit numbagg.nanmean(x) 1000 loops, best of 3: 1.92 ms per loop %timeit numbagg.nanmean(x, axis=0) 100 loops, best of 3: 2.75 ms per loop %timeit numbagg.nanmean(x, axis=1) 1000 loops, best of 3: 1.86 ms per loop

%timeit bottleneck.nanmean(x) 100 loops, best of 3: 2.75 ms per loop %timeit bottleneck.nanmean(x, axis=0) 100 loops, best of 3: 3.47 ms per loop %timeit bottleneck.nanmean(x, axis=1) 100 loops, best of 3: 2.72 ms per loop ```

{
    "total_count": 1,
    "+1": 1,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Canonical approach for new vectorized functions 305663416

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [performed_via_github_app] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
    ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
    ON [issue_comments] ([user]);
Powered by Datasette · Queries took 190.203ms · About: xarray-datasette