home / github / issue_comments

Menu
  • Search all tables
  • GraphQL API

issue_comments: 1034678675

This data as json

html_url issue_url id node_id user created_at updated_at author_association body reactions performed_via_github_app issue
https://github.com/pydata/xarray/issues/6069#issuecomment-1034678675 https://api.github.com/repos/pydata/xarray/issues/6069 1034678675 IC_kwDOAMm_X849q_GT 9576982 2022-02-10T09:18:47Z 2022-02-10T09:18:47Z NONE

If Xarray/zarr is to replace netcdf, appending by time step is really an important feature Most (all?) numerical models will output results per time step onto a multidimensional grid with different variables Said grid will also have other parameters that will help rebuild the geometry or follow standards, like CF and Ugrid (The things that you are supposed to drop). The geometry of the grid is computed at the initialisation of the model. It is a bit counter intuitive to get rid of it for incremental backups especially that each write will not concern this part of the file. What I do at the moment is that I create a first dataset at the final dimension based on dummy dask arrays Export it to_zarr withcompute = False

With a buffer system, I create a new dataset for each buffer with the right data at the right place meaning only the time interval concerned and I write to_zarr with the region attribute I flush the buffer dataset after it being written.

At the end I write all the parameters before closing the main dataset.

To my knowledge, that's the only method which works.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  1077079208
Powered by Datasette · Queries took 0.653ms · About: xarray-datasette