issue_comments: 412177726
This data as json
html_url | issue_url | id | node_id | user | created_at | updated_at | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
https://github.com/pydata/xarray/issues/2159#issuecomment-412177726 | https://api.github.com/repos/pydata/xarray/issues/2159 | 412177726 | MDEyOklzc3VlQ29tbWVudDQxMjE3NzcyNg== | 35968931 | 2018-08-10T19:08:56Z | 2018-08-11T00:09:28Z | MEMBER | I've been looking through the functions The current behaviour isn't completely explicit, and I would like to check my understanding with a few questions: 1) If you 2) Although
will only organise the datasets into groups according to the set of dimensions they have, it doesn't order the datasets within each group according to the values in the dimension coordinates? We can show this because this (new) testcase fails: ```python @requires_dask def test_auto_combine_along_coords(self): # drop the third dimension to keep things relatively understandable data = create_test_data() for k in list(data.variables): if 'dim3' in data[k].dims: del data[k]
``` with output
3) So the call to
4) Therefore what needs to be done here is the Also, ```python User specifies how they split up their domaindomain_decomposition_structure = how_was_this_parallelized('output.*.nc') Feeds this info into open_mfdatasetfull_domain = xr.open_mfdataset('output.*.nc', positions=domain_decomposition_structure) ``` This approach would be much less general but would dodge the issue of writing generalized N-D auto-concatenation logic. Final point - this common use case also has the added complexity of having ghost or guard cells around every dataset, which should be thrown away. Clearly some user input is required here ( |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
324350248 |