issue_comments: 454423937
This data as json
html_url | issue_url | id | node_id | user | created_at | updated_at | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
https://github.com/pydata/xarray/issues/2662#issuecomment-454423937 | https://api.github.com/repos/pydata/xarray/issues/2662 | 454423937 | MDEyOklzc3VlQ29tbWVudDQ1NDQyMzkzNw== | 35968931 | 2019-01-15T15:05:22Z | 2019-01-15T15:05:22Z | MEMBER | Yes thankyou @malmans2, this is very helpful!
This was very puzzling because the code is supposed to split the datasets up according to their data variables, which means merge won't be used to concatenate and this should be fast, as before. But I found the problem! In before
With this change then I get ```python No longer slow if netCDFs are stored in several folders:%timeit ds_2folders = xr.open_mfdataset('rep/.nc', concat_dim='T')
Without this pre-sorting, Whether or not groupby sorted properly depended on the order of datasets in the input to groupby, which eventually depended on the way they were loaded (as the example in this issue makes clear). The reason this mistake got past the unit tests is that |
{ "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 } |
397063221 |