home / github / issues

Menu
  • GraphQL API
  • Search all tables

issues: 225734529

This data as json

id node_id number title user state locked assignee milestone comments created_at updated_at closed_at author_association active_lock_reason draft pull_request body reactions performed_via_github_app state_reason repo type
225734529 MDU6SXNzdWUyMjU3MzQ1Mjk= 1394 autoclose with distributed doesn't seem to work 1197350 closed 0     9 2017-05-02T15:37:07Z 2019-01-13T19:35:10Z 2019-01-13T19:35:10Z MEMBER      

I am trying to analyze a very large netCDF dataset using xarray and distributed.

I open my dataset with the new autoclose option: python ds = xr.open_mfdataset(ddir + '*.nc', decode_cf=False, autoclose=True)

However, when I try some reduction operation (e.g. ds['Salt'].mean()), I can see my open file count continue to rise monotonically. Eventually the dask worker dies with OSError: [Errno 24] Too many open files: '/proc/65644/sta once I hit the system ulimit.

Am I doing something wrong here? Why are the files not being closed? cc: @pwolfram

{
    "url": "https://api.github.com/repos/pydata/xarray/issues/1394/reactions",
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  completed 13221727 issue

Links from other tables

  • 3 rows from issues_id in issues_labels
  • 9 rows from issue in issue_comments
Powered by Datasette · Queries took 0.565ms · About: xarray-datasette