issues: 218260909
This data as json
id | node_id | number | title | user | state | locked | assignee | milestone | comments | created_at | updated_at | closed_at | author_association | active_lock_reason | draft | pull_request | body | reactions | performed_via_github_app | state_reason | repo | type |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
218260909 | MDU6SXNzdWUyMTgyNjA5MDk= | 1340 | round-trip performance with save_mfdataset / open_mfdataset | 1197350 | closed | 0 | 11 | 2017-03-30T16:52:26Z | 2019-05-01T22:12:06Z | 2019-05-01T22:12:06Z | MEMBER | I have encountered some major performance bottlenecks in trying to write and then read multi-file netcdf datasets. I start with an xarray dataset created by xgcm with the following repr:
An important point to note is that there are lots of "non-dimension coordinates" corresponding to various parameters of the numerical grid. I save this dataset to a multi-file netCDF dataset as follows:
Then I try to re-load this dataset
This raises an error:
I need to specify I just thought I would document this, because 18 minutes seems way too long to load a dataset. |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1340/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | 13221727 | issue |