home / github

Menu
  • GraphQL API
  • Search all tables

issue_comments

Table actions
  • GraphQL API for issue_comments

6 rows where author_association = "CONTRIBUTOR", issue = 1423312198 and user = 90008 sorted by updated_at descending

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: reactions, created_at (date), updated_at (date)

user 1

  • hmaarrfk · 6 ✖

issue 1

  • Remove debugging slow assert statement · 6 ✖

author_association 1

  • CONTRIBUTOR · 6 ✖
id html_url issue_url node_id user created_at updated_at ▲ author_association body reactions performed_via_github_app issue
1291948502 https://github.com/pydata/xarray/pull/7221#issuecomment-1291948502 https://api.github.com/repos/pydata/xarray/issues/7221 IC_kwDOAMm_X85NAZHW hmaarrfk 90008 2022-10-26T12:19:49Z 2022-10-26T12:23:46Z CONTRIBUTOR

I know it is not comparable, but I was really curious what "dictionary insertion" costs, in order to be able to understand if my comparisons were fair:

code ```python from tqdm import tqdm import xarray as xr from time import perf_counter import numpy as np N = 1000 # Everybody is lazy loading now, so lets force modules to get instantiated dummy_dataset = xr.Dataset() dummy_dataset['a'] = 1 dummy_dataset['b'] = 1 del dummy_dataset time_elapsed = np.zeros(N) # dataset = xr.Dataset() dataset = {} for i in tqdm(range(N)): # for i in range(N): time_start = perf_counter() dataset[f"var{i}"] = i time_end = perf_counter() time_elapsed[i] = time_end - time_start # %% from matplotlib import pyplot as plt plt.plot(np.arange(N), time_elapsed * 1E6, label='Time to add one variable') plt.xlabel("Number of existing variables") plt.ylabel("Time to add a variables (us)") plt.ylim([0, 10]) plt.title("Dictionary insertion") plt.grid(True) ```

I think xarray gives me 3 order of magnitude of "thinking" benefit, so I'll take it! python --version Python 3.9.13

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Remove debugging slow assert statement 1423312198
1291894024 https://github.com/pydata/xarray/pull/7221#issuecomment-1291894024 https://api.github.com/repos/pydata/xarray/issues/7221 IC_kwDOAMm_X85NAL0I hmaarrfk 90008 2022-10-26T11:32:32Z 2022-10-26T11:32:32Z CONTRIBUTOR

Ok. I'll want to rethink them.

I know it looks quadratic time, but i really would like to test n=1000 and i have an idea

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Remove debugging slow assert statement 1423312198
1291450556 https://github.com/pydata/xarray/pull/7221#issuecomment-1291450556 https://api.github.com/repos/pydata/xarray/issues/7221 IC_kwDOAMm_X85M-fi8 hmaarrfk 90008 2022-10-26T03:32:53Z 2022-10-26T03:32:53Z CONTRIBUTOR

I'm somewhat ocnfused, I can run the benchmark locally

``` [ 1.80%] ··· dataset_creation.Creation.time_dataset_creation 4.37±0s

```

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Remove debugging slow assert statement 1423312198
1291447746 https://github.com/pydata/xarray/pull/7221#issuecomment-1291447746 https://api.github.com/repos/pydata/xarray/issues/7221 IC_kwDOAMm_X85M-e3C hmaarrfk 90008 2022-10-26T03:27:36Z 2022-10-26T03:27:36Z CONTRIBUTOR

:/ not fun, the benchmark is failing. not sure why.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Remove debugging slow assert statement 1423312198
1291399714 https://github.com/pydata/xarray/pull/7221#issuecomment-1291399714 https://api.github.com/repos/pydata/xarray/issues/7221 IC_kwDOAMm_X85M-TIi hmaarrfk 90008 2022-10-26T02:14:40Z 2022-10-26T02:14:40Z CONTRIBUTOR

Would be interesting to see whether this was covered by our existing asv benchmarks.

I wasn't able to find something that really benchmarked "large" datasets.

Would be a good benchmark to add if we don't have one already.

Added one.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Remove debugging slow assert statement 1423312198
1291389702 https://github.com/pydata/xarray/pull/7221#issuecomment-1291389702 https://api.github.com/repos/pydata/xarray/issues/7221 IC_kwDOAMm_X85M-QsG hmaarrfk 90008 2022-10-26T01:59:57Z 2022-10-26T01:59:57Z CONTRIBUTOR

out of interest, how did you find this?

Spyder profiler

{
    "total_count": 2,
    "+1": 2,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Remove debugging slow assert statement 1423312198

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [performed_via_github_app] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
    ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
    ON [issue_comments] ([user]);
Powered by Datasette · Queries took 15.414ms · About: xarray-datasette