html_url,issue_url,id,node_id,user,created_at,updated_at,author_association,body,reactions,performed_via_github_app,issue https://github.com/pydata/xarray/issues/3486#issuecomment-909342151,https://api.github.com/repos/pydata/xarray/issues/3486,909342151,IC_kwDOAMm_X842M3XH,2448579,2021-08-31T15:27:28Z,2021-08-31T15:27:57Z,MEMBER,"What happens is that dask first constructs chunks of size specified in `open_mfdataset` and then breaks those up to new chunk sizes specified in the `.chunk()` call. A similar behaviour is present for repeated chunk calls `.chunk().chunk()`, these do not get optimized to a single chunk call yet. So yes, you should pass appropriate chunk sizes in `open_mfdataset`","{""total_count"": 0, ""+1"": 0, ""-1"": 0, ""laugh"": 0, ""hooray"": 0, ""confused"": 0, ""heart"": 0, ""rocket"": 0, ""eyes"": 0}",,517799069 https://github.com/pydata/xarray/issues/3486#issuecomment-906023525,https://api.github.com/repos/pydata/xarray/issues/3486,906023525,IC_kwDOAMm_X842ANJl,2656596,2021-08-26T02:19:50Z,2021-08-26T02:19:50Z,NONE,"This seems to be an ongoing problem ([Unexpected behaviour when chunking with multiple netcdf files in xarray/dask](https://stackoverflow.com/questions/62932044/unexpected-behaviour-when-chunking-with-multiple-netcdf-files-in-xarray-dask), [Performance of chunking in xarray / dask when opening and re-chunking a dataset](https://stackoverflow.com/questions/58838873/performance-of-chunking-in-xarray-dask-when-opening-and-re-chunking-a-dataset)) that has not been resolved nor has feedback been provided. I've been running into this problem trying to handle netcdfs that are larger than my RAM. From my testing, chunks must be passed with open_mfdataset to be of any use. The chunks method on the datatset after opening seems to do nothing in this use case.","{""total_count"": 0, ""+1"": 0, ""-1"": 0, ""laugh"": 0, ""hooray"": 0, ""confused"": 0, ""heart"": 0, ""rocket"": 0, ""eyes"": 0}",,517799069