html_url,issue_url,id,node_id,user,created_at,updated_at,author_association,body,reactions,performed_via_github_app,issue https://github.com/pydata/xarray/issues/1440#issuecomment-632294837,https://api.github.com/repos/pydata/xarray/issues/1440,632294837,MDEyOklzc3VlQ29tbWVudDYzMjI5NDgzNw==,1197350,2020-05-21T19:19:50Z,2020-05-21T19:19:50Z,MEMBER,"> It seems to me that the there are lots of ""layers"" of ""chunking"", especially when you are talking about chunking an entire dataset, To simplify a little bit, here we are only talking about reading a single store, i.e. one netcdf file or one zarr group. Also out of scope is the underlying storage medium (e.g. block size). ","{""total_count"": 2, ""+1"": 2, ""-1"": 0, ""laugh"": 0, ""hooray"": 0, ""confused"": 0, ""heart"": 0, ""rocket"": 0, ""eyes"": 0}",,233350060 https://github.com/pydata/xarray/issues/1440#issuecomment-632266536,https://api.github.com/repos/pydata/xarray/issues/1440,632266536,MDEyOklzc3VlQ29tbWVudDYzMjI2NjUzNg==,1197350,2020-05-21T18:23:13Z,2020-05-21T18:23:13Z,MEMBER,"> Can we overload the `chunks` argument in `open_xxx` to do this? We are already adding support for `chunks=""auto""` ... This gets tricky, because we may want slightly different behavior depending on whether the underlying array store is chunked.","{""total_count"": 0, ""+1"": 0, ""-1"": 0, ""laugh"": 0, ""hooray"": 0, ""confused"": 0, ""heart"": 0, ""rocket"": 0, ""eyes"": 0}",,233350060 https://github.com/pydata/xarray/issues/1440#issuecomment-632183683,https://api.github.com/repos/pydata/xarray/issues/1440,632183683,MDEyOklzc3VlQ29tbWVudDYzMjE4MzY4Mw==,1197350,2020-05-21T16:13:46Z,2020-05-21T16:14:08Z,MEMBER,"We discussed this issue today in our pangeo coffee break. We think the following plan would be good: - [ ] Write a function called `auto_chunk(variable)` which examines a variable for the presence of a `chunks` attribute in encoding or within the data itself. Returns a new variable with chunked data. - [ ] Refactor `open_zarr` to call this function - [ ] Add it also to `open_dataset` to enable auto-chunking of netCDF and geotiff data Should we have an option like `chunk_size='native'`, or `chunk_size='100MB'`, with chunks chosen to align with source chunks.","{""total_count"": 2, ""+1"": 0, ""-1"": 0, ""laugh"": 0, ""hooray"": 2, ""confused"": 0, ""heart"": 0, ""rocket"": 0, ""eyes"": 0}",,233350060