issue_comments: 1034678675
This data as json
html_url | issue_url | id | node_id | user | created_at | updated_at | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
https://github.com/pydata/xarray/issues/6069#issuecomment-1034678675 | https://api.github.com/repos/pydata/xarray/issues/6069 | 1034678675 | IC_kwDOAMm_X849q_GT | 9576982 | 2022-02-10T09:18:47Z | 2022-02-10T09:18:47Z | NONE | If Xarray/zarr is to replace netcdf, appending by time step is really an important feature
Most (all?) numerical models will output results per time step onto a multidimensional grid with different variables
Said grid will also have other parameters that will help rebuild the geometry or follow standards, like CF and Ugrid (The things that you are supposed to drop). The geometry of the grid is computed at the initialisation of the model. It is a bit counter intuitive to get rid of it for incremental backups especially that each write will not concern this part of the file.
What I do at the moment is that I create a first dataset at the final dimension based on dummy dask arrays
Export it With a buffer system, I create a new dataset for each buffer with the right data at the right place meaning only the time interval concerned and I write At the end I write all the parameters before closing the main dataset. To my knowledge, that's the only method which works. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
1077079208 |