html_url,issue_url,id,node_id,user,created_at,updated_at,author_association,body,reactions,performed_via_github_app,issue
https://github.com/pydata/xarray/issues/2300#issuecomment-406718847,https://api.github.com/repos/pydata/xarray/issues/2300,406718847,MDEyOklzc3VlQ29tbWVudDQwNjcxODg0Nw==,1217238,2018-07-20T20:31:42Z,2018-07-20T20:31:42Z,MEMBER,"> Curious: Is there any downside in xarray to using datasets with inconsistent chunks?
No, there's no downside here. It's just not possible to define a single dict of chunks in this case.
Can you look into the `encoding` attributes of any variables you load from disk?
It would also help to come up with a self-contained example that reproduces this using dummy data.","{""total_count"": 0, ""+1"": 0, ""-1"": 0, ""laugh"": 0, ""hooray"": 0, ""confused"": 0, ""heart"": 0, ""rocket"": 0, ""eyes"": 0}",,342531772
https://github.com/pydata/xarray/issues/2300#issuecomment-406165245,https://api.github.com/repos/pydata/xarray/issues/2300,406165245,MDEyOklzc3VlQ29tbWVudDQwNjE2NTI0NQ==,1217238,2018-07-19T06:08:26Z,2018-07-19T06:08:26Z,MEMBER,I just pushed a new xarray release (0.10.8) earlier today. We had a fix for zarr chunking in there (https://github.com/pydata/xarray/pull/2228) -- does that solve your issue?,"{""total_count"": 0, ""+1"": 0, ""-1"": 0, ""laugh"": 0, ""hooray"": 0, ""confused"": 0, ""heart"": 0, ""rocket"": 0, ""eyes"": 0}",,342531772