home / github

Menu
  • Search all tables
  • GraphQL API

issue_comments

Table actions
  • GraphQL API for issue_comments

7 rows where user = 64621312 sorted by updated_at descending

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: issue_url, reactions, created_at (date), updated_at (date)

issue 5

  • PermissionError: [Errno 13] Permission denied 2
  • Writing netcdf after running xarray.dataset.reindex to fill gaps in a time series fails due to memory allocation error 2
  • 2 Dimension Plot Producing Discontinuous Grid 1
  • Writing a netCDF file is slow 1
  • Memory Leakage Issue When Running to_netcdf 1

user 1

  • lassiterdc · 7 ✖

author_association 1

  • NONE 7
id html_url issue_url node_id user created_at updated_at ▲ author_association body reactions performed_via_github_app issue
1248550900 https://github.com/pydata/xarray/issues/7018#issuecomment-1248550900 https://api.github.com/repos/pydata/xarray/issues/7018 IC_kwDOAMm_X85Ka1_0 lassiterdc 64621312 2022-09-15T19:59:39Z 2022-09-15T19:59:39Z NONE

I tried your suggestion and still ran into a memory allocation error, but it sounds like you're onto something. I also found this other thread about reindex causing memory allocation errors but it doesn't look like a solution was discovered there either. https://github.com/pydata/xarray/issues/2745

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Writing netcdf after running xarray.dataset.reindex to fill gaps in a time series fails due to memory allocation error 1368696980
1242793638 https://github.com/pydata/xarray/issues/7018#issuecomment-1242793638 https://api.github.com/repos/pydata/xarray/issues/7018 IC_kwDOAMm_X85KE4am lassiterdc 64621312 2022-09-10T19:34:26Z 2022-09-10T19:34:26Z NONE

I found a functional workaround is to chunk by one of the spatial dimensions instead. I'd still like to know why the code above fails though. I'm assuming there's a scheduled task with dask the occurs before to_netcdf but I haven't been able to figure out what that is.

python ds = xr.open_dataset(fldr_in_grib, engine="cfgrib", chunks={"latitude":875}, backend_kwargs={'indexpath': ''})

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Writing netcdf after running xarray.dataset.reindex to fill gaps in a time series fails due to memory allocation error 1368696980
1219672919 https://github.com/pydata/xarray/issues/6924#issuecomment-1219672919 https://api.github.com/repos/pydata/xarray/issues/6924 IC_kwDOAMm_X85IsrtX lassiterdc 64621312 2022-08-18T16:05:02Z 2022-08-18T16:05:02Z NONE

I cross posted this as a dask issue and on stack overflow. I learned that "dask will often have as many chunks in memory as twice the number of active threads" (best practices with dask arrays) and including dask.config.set(scheduler='synchronous') which forces single thread computation (dask scheduling) resulted in the behavior I expected which is memory usage fluctuations that are roughly the magnitude of the chunk size.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Memory Leakage Issue When Running to_netcdf  1340994913
1216907148 https://github.com/pydata/xarray/issues/6920#issuecomment-1216907148 https://api.github.com/repos/pydata/xarray/issues/6920 IC_kwDOAMm_X85IiIeM lassiterdc 64621312 2022-08-16T16:59:47Z 2022-08-16T16:59:47Z NONE

Thanks, @andersy005. I think that xr.save_mfdataset() could certainly be helpful in my workflow but unfortunately, I have to consolidate these data from a netcdf for each 2-minute timestep to a netcdf for each day, and it sounds like there's no way around that bottleneck. I've come across suggestions to save the dataset to a zarr group and then export as a netcdf, so I'm going to give that a shot.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Writing a netCDF file is slow 1340474484
1208388460 https://github.com/pydata/xarray/issues/6892#issuecomment-1208388460 https://api.github.com/repos/pydata/xarray/issues/6892 IC_kwDOAMm_X85IBots lassiterdc 64621312 2022-08-08T17:12:41Z 2022-08-08T17:12:41Z NONE

I realized that the reason I wasn't seeing what I was expecting was because of the dataset, not the plotting function. There shouldn't be any 0s in the rainrate variable.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  2 Dimension Plot Producing Discontinuous Grid 1332143835
1187959131 https://github.com/pydata/xarray/issues/6805#issuecomment-1187959131 https://api.github.com/repos/pydata/xarray/issues/6805 IC_kwDOAMm_X85GztFb lassiterdc 64621312 2022-07-18T17:57:35Z 2022-07-18T17:57:35Z NONE

I can't believe I forgot the asterisk!!! Thank you for catching that.

{
    "total_count": 2,
    "+1": 2,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  PermissionError: [Errno 13] Permission denied 1308176241
1187910591 https://github.com/pydata/xarray/issues/6805#issuecomment-1187910591 https://api.github.com/repos/pydata/xarray/issues/6805 IC_kwDOAMm_X85GzhO_ lassiterdc 64621312 2022-07-18T17:42:14Z 2022-07-18T17:42:14Z NONE

I just edited the original post to show that xr.open_dataset works. I hope that is adequate to show that the issue is with xarray and not netcdf4.Dataset.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  PermissionError: [Errno 13] Permission denied 1308176241

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [performed_via_github_app] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
    ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
    ON [issue_comments] ([user]);
Powered by Datasette · Queries took 18.619ms · About: xarray-datasette