home / github

Menu
  • Search all tables
  • GraphQL API

issue_comments

Table actions
  • GraphQL API for issue_comments

6 rows where author_association = "MEMBER", issue = 166593563 and user = 1217238 sorted by updated_at descending

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: created_at (date), updated_at (date)

user 1

  • shoyer · 6 ✖

issue 1

  • Speed up operations with xarray dataset · 6 ✖

author_association 1

  • MEMBER · 6 ✖
id html_url issue_url node_id user created_at updated_at ▲ author_association body reactions performed_via_github_app issue
234056046 https://github.com/pydata/xarray/issues/912#issuecomment-234056046 https://api.github.com/repos/pydata/xarray/issues/912 MDEyOklzc3VlQ29tbWVudDIzNDA1NjA0Ng== shoyer 1217238 2016-07-20T19:29:55Z 2016-07-20T19:29:55Z MEMBER

Just looking at a task manager while a task executes can give you a sense of what's going on. Dask also has some diagnostics that may be helpful: http://dask.pydata.org/en/latest/diagnostics.html On Wed, Jul 20, 2016 at 11:44 AM Saulo Meirelles notifications@github.com wrote:

No, not really. I got no error message whatsoever. Is there any test I can do to tackle this?

Sent from Smartphone. Please forgive typos.

On Jul 20, 2016 8:41 PM, "Stephan Hoyer" notifications@github.com wrote:

I decided to wait for .load() to do the job but the kernel dies after a while.

Are you running out of memory? Can you tell what's going on? This is a little surprising to me.

— You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/pydata/xarray/issues/912#issuecomment-234042142, or mute the thread < https://github.com/notifications/unsubscribe-auth/AHKCTXaBpbA0ieSdI2I_hIUjVBxuKaNeks5qXmvPgaJpZM4JQ0_D

.

— You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/pydata/xarray/issues/912#issuecomment-234043292, or mute the thread https://github.com/notifications/unsubscribe-auth/ABKS1ujXItyYDLgA4ZtBkHEbTBTiTnrvks5qXmylgaJpZM4JQ0_D .

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Speed up operations with xarray dataset 166593563
234042142 https://github.com/pydata/xarray/issues/912#issuecomment-234042142 https://api.github.com/repos/pydata/xarray/issues/912 MDEyOklzc3VlQ29tbWVudDIzNDA0MjE0Mg== shoyer 1217238 2016-07-20T18:41:17Z 2016-07-20T18:41:17Z MEMBER

I decided to wait for .load() to do the job but the kernel dies after a while.

Are you running out of memory? Can you tell what's going on? This is a little surprising to me.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Speed up operations with xarray dataset 166593563
234026185 https://github.com/pydata/xarray/issues/912#issuecomment-234026185 https://api.github.com/repos/pydata/xarray/issues/912 MDEyOklzc3VlQ29tbWVudDIzNDAyNjE4NQ== shoyer 1217238 2016-07-20T17:47:45Z 2016-07-20T17:47:45Z MEMBER

It's worth noting that conc_avg = ds.conc_profs.chunk({'burst': 10}).mean(('z','duration')) doesn't actually do any computation -- that's why it's so fast. It just sets up the computation graph. No computation happens until you write .load().

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Speed up operations with xarray dataset 166593563
233998757 https://github.com/pydata/xarray/issues/912#issuecomment-233998757 https://api.github.com/repos/pydata/xarray/issues/912 MDEyOklzc3VlQ29tbWVudDIzMzk5ODc1Nw== shoyer 1217238 2016-07-20T16:11:27Z 2016-07-20T16:11:27Z MEMBER

When you write ds.conc_profs.chunk(2400), it sets up the data to be loaded in a giant chunk, almost the entire file at once. Even if you use .isel() afterwards, dask does not always manage to subset the data from the initial chunk. (Sometimes it does succeed, which makes this a little confusing.)

You will probably be more successful if you try something like ds.conc_profs.chunk({'burst': 10}) instead, which keeps the intermediate chunks to a reasonable size.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Speed up operations with xarray dataset 166593563
233996527 https://github.com/pydata/xarray/issues/912#issuecomment-233996527 https://api.github.com/repos/pydata/xarray/issues/912 MDEyOklzc3VlQ29tbWVudDIzMzk5NjUyNw== shoyer 1217238 2016-07-20T16:03:30Z 2016-07-20T16:03:30Z MEMBER

Thanks for describing that -- I misread your initial description and thought you were using open_mfdataset rather than open_dataset (the glob threw me off!). The source of these files shouldn't matter once you have it in a netCDF file.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Speed up operations with xarray dataset 166593563
233991357 https://github.com/pydata/xarray/issues/912#issuecomment-233991357 https://api.github.com/repos/pydata/xarray/issues/912 MDEyOklzc3VlQ29tbWVudDIzMzk5MTM1Nw== shoyer 1217238 2016-07-20T15:46:50Z 2016-07-20T15:46:50Z MEMBER

What do the original input files look like, before you join them together? This may be a case where the dask.array task scheduler does very poorly.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Speed up operations with xarray dataset 166593563

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [performed_via_github_app] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
    ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
    ON [issue_comments] ([user]);
Powered by Datasette · Queries took 166.212ms · About: xarray-datasette