issue_comments
28 rows where user = 6574622 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: issue_url, reactions, created_at (date), updated_at (date)
user 1
- d70-t · 28 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
1064981526 | https://github.com/pydata/xarray/issues/6329#issuecomment-1064981526 | https://api.github.com/repos/pydata/xarray/issues/6329 | IC_kwDOAMm_X84_elQW | d70-t 6574622 | 2022-03-11T10:28:35Z | 2022-03-11T10:28:35Z | CONTRIBUTOR | Thanks for pointing out |
{ "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
`to_zarr` with append or region mode and `_FillValue` doesnt work 1159923690 | |
1063977656 | https://github.com/pydata/xarray/issues/6329#issuecomment-1063977656 | https://api.github.com/repos/pydata/xarray/issues/6329 | IC_kwDOAMm_X84_awK4 | d70-t 6574622 | 2022-03-10T11:56:44Z | 2022-03-10T11:56:44Z | CONTRIBUTOR | Yes, this is kind of the behaviour I'd expect. And great that it helped clarifying things. Still, building up the metadata nicely upfront (which is required for region writes) ist quite convoluted... That's what I meant with
in the previous comment. I think, establishing and documenting good practices for this would help, but probably we also want to have better tools. In any case, this would probably be yet another issue. Note that if you care about this paricular example (e.g. appending in a single thread in increasing order of timesteps), then it should also be possible to do this much simpler using append: ```python filename='processed_dataset.zarr' ds = xr.tutorial.open_dataset('air_temperature') ds.air.encoding['dtype']=np.dtype('float32') X,Y=250, 250 #size of each final timestep for i in range(len(ds.time)): # some kind of heavy processing arr_r=some_processing(ds.isel(time=slice(i,i+1)),X,Y) del arr_r.air.attrs["_FillValue"] if os.path.exists(filename): arr_r.to_zarr(filename, append_dim='time') else: arr_r.to_zarr(filename) ``` If you find out more about the cloud case, please post a note, otherwise, we can assume that the original bug report is fine? |
{ "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
`to_zarr` with append or region mode and `_FillValue` doesnt work 1159923690 | |
1063859715 | https://github.com/pydata/xarray/issues/6329#issuecomment-1063859715 | https://api.github.com/repos/pydata/xarray/issues/6329 | IC_kwDOAMm_X84_aTYD | d70-t 6574622 | 2022-03-10T09:44:59Z | 2022-03-10T09:44:59Z | CONTRIBUTOR | Sure, no problem. I believe, this page has a good summary:
So the difference between "a" and "r+" roughly codifies the intended behaviour for sequential access (it's ok to modify everything) and parallel access to independent chunks (where modifying metadata would be bad). So probably that message was suggesting that you have to use "a" if you want to modify metadata (e.g. by expanding the shape), which is true. But to me, it's unclear how one would do that safely with (potentially) parallel region writes, so it's kind of reasonable that region writes don't like to modify metadata. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
`to_zarr` with append or region mode and `_FillValue` doesnt work 1159923690 | |
1062755678 | https://github.com/pydata/xarray/issues/6329#issuecomment-1062755678 | https://api.github.com/repos/pydata/xarray/issues/6329 | IC_kwDOAMm_X84_WF1e | d70-t 6574622 | 2022-03-09T10:06:22Z | 2022-03-09T10:06:22Z | CONTRIBUTOR | Yes, that looks like the error as described in the initial post.
Adding the described workaround (i.e.
Which is due to a mix of append-mode (
Currently, I can't really imagine how a mix of both should behave. If you can't prepare the dataset for the final shape upfront (to use |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
`to_zarr` with append or region mode and `_FillValue` doesnt work 1159923690 | |
1061711069 | https://github.com/pydata/xarray/issues/6329#issuecomment-1061711069 | https://api.github.com/repos/pydata/xarray/issues/6329 | IC_kwDOAMm_X84_SGzd | d70-t 6574622 | 2022-03-08T12:09:38Z | 2022-03-08T12:09:38Z | CONTRIBUTOR | You've got the |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
`to_zarr` with append or region mode and `_FillValue` doesnt work 1159923690 | |
1061081884 | https://github.com/pydata/xarray/issues/6329#issuecomment-1061081884 | https://api.github.com/repos/pydata/xarray/issues/6329 | IC_kwDOAMm_X84_PtMc | d70-t 6574622 | 2022-03-07T20:03:18Z | 2022-03-07T20:03:18Z | CONTRIBUTOR | { "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
`to_zarr` with append or region mode and `_FillValue` doesnt work 1159923690 | ||
1059426353 | https://github.com/pydata/xarray/issues/6329#issuecomment-1059426353 | https://api.github.com/repos/pydata/xarray/issues/6329 | IC_kwDOAMm_X84_JZAx | d70-t 6574622 | 2022-03-04T18:48:13Z | 2022-03-04T18:48:13Z | CONTRIBUTOR | If that's necessary to reproduce the problem, then yes. If it's possible to show the same thing with less "noise", then it's better to not use the tutorial dataset and to not use something like a cloud backend. But we can also try to iterate on this again, to progressively get down to a smaller example. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
`to_zarr` with append or region mode and `_FillValue` doesnt work 1159923690 | |
1059405550 | https://github.com/pydata/xarray/issues/6069#issuecomment-1059405550 | https://api.github.com/repos/pydata/xarray/issues/6069 | IC_kwDOAMm_X84_JT7u | d70-t 6574622 | 2022-03-04T18:16:57Z | 2022-03-04T18:16:57Z | CONTRIBUTOR | I'll set up a new issue. @Boorhin, I couldn't confirm the weirdness with the small example, but will put in a note to your comment. If you can reproduce the weirdness on the minimal example, would you make a comment to the new issue? |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
to_zarr: region not recognised as dataset dimensions 1077079208 | |
1059378287 | https://github.com/pydata/xarray/issues/6069#issuecomment-1059378287 | https://api.github.com/repos/pydata/xarray/issues/6069 | IC_kwDOAMm_X84_JNRv | d70-t 6574622 | 2022-03-04T17:39:24Z | 2022-03-04T17:39:24Z | CONTRIBUTOR | I've made a simpler example of the The workaround:
@dcherian, @Boorhin should we make a new (CF-related) issue out of this and try to keep focussing on append and region use-cases here, which seemed to be the initial problem in this thread (probably by going further through your example @Boorhin?). |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
to_zarr: region not recognised as dataset dimensions 1077079208 | |
1059078961 | https://github.com/pydata/xarray/issues/6069#issuecomment-1059078961 | https://api.github.com/repos/pydata/xarray/issues/6069 | IC_kwDOAMm_X84_IEMx | d70-t 6574622 | 2022-03-04T11:27:12Z | 2022-03-04T11:27:44Z | CONTRIBUTOR | btw, as a work-around it works when removing the
But still, this might call for another issue to solve. |
{ "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
to_zarr: region not recognised as dataset dimensions 1077079208 | |
1059076885 | https://github.com/pydata/xarray/issues/6069#issuecomment-1059076885 | https://api.github.com/repos/pydata/xarray/issues/6069 | IC_kwDOAMm_X84_IDsV | d70-t 6574622 | 2022-03-04T11:23:56Z | 2022-03-04T11:23:56Z | CONTRIBUTOR | Ok, I believe, I've now reproduced your error: ```python import xarray as xr from rasterio.enums import Resampling import numpy as np ds = xr.tutorial.open_dataset('air_temperature').isel(time=0) ds = ds.rio.write_crs('EPSG:4326') dst = ds.rio.reproject('EPSG:3857', shape=(250, 250), resampling=Resampling.bilinear, nodata=np.nan) dst.air.encoding = {} dst = dst.assign(air=dst.air.expand_dims("time"), time=dst.time.expand_dims("time")) m = {}
dst.to_zarr(m)
dst.to_zarr(m, append_dim="time")
This seems to be due to handling of CF-Conventions which might go wrong in the append case: the |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
to_zarr: region not recognised as dataset dimensions 1077079208 | |
1059063397 | https://github.com/pydata/xarray/issues/6069#issuecomment-1059063397 | https://api.github.com/repos/pydata/xarray/issues/6069 | IC_kwDOAMm_X84_IAZl | d70-t 6574622 | 2022-03-04T11:05:07Z | 2022-03-04T11:05:07Z | CONTRIBUTOR | This error ist unrelated to region or append writes. The dataset
but still carries encoding-information from
|
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
to_zarr: region not recognised as dataset dimensions 1077079208 | |
1059025444 | https://github.com/pydata/xarray/issues/6069#issuecomment-1059025444 | https://api.github.com/repos/pydata/xarray/issues/6069 | IC_kwDOAMm_X84_H3Ik | d70-t 6574622 | 2022-03-04T10:13:40Z | 2022-03-04T10:13:40Z | CONTRIBUTOR | 🤷 can't help any further without a minimal reproducible example here... |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
to_zarr: region not recognised as dataset dimensions 1077079208 | |
1058381922 | https://github.com/pydata/xarray/issues/6069#issuecomment-1058381922 | https://api.github.com/repos/pydata/xarray/issues/6069 | IC_kwDOAMm_X84_FaBi | d70-t 6574622 | 2022-03-03T18:56:13Z | 2022-03-03T18:56:13Z | CONTRIBUTOR | I don't yet know a proper answer, but there'd be three observations I have:
* The |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
to_zarr: region not recognised as dataset dimensions 1077079208 | |
1052252098 | https://github.com/pydata/xarray/issues/6069#issuecomment-1052252098 | https://api.github.com/repos/pydata/xarray/issues/6069 | IC_kwDOAMm_X84-uBfC | d70-t 6574622 | 2022-02-26T16:07:56Z | 2022-02-26T16:07:56Z | CONTRIBUTOR | While testing a bit further, I found another case which might potentially be dangerous: ```python ds is the same as above, but chunksize is {"time": 1, "x": 1}once on the coordinatords.to_zarr("test.zarr", compute=False, encoding={"time": {"chunks": [1]}, "x": {"chunks": [1]}}) in parallelds.isel(time=slice(0,1), x=slice(0,1)).to_zarr("test.zarr", mode="r+", region={"time": slice(0,1), "x": slice(0,1)}) ds.isel(time=slice(0,1), x=slice(1,2)).to_zarr("test.zarr", mode="r+", region={"time": slice(0,1), "x": slice(1,2)}) ds.isel(time=slice(0,1), x=slice(2,3)).to_zarr("test.zarr", mode="r+", region={"time": slice(0,1), "x": slice(2,3)}) ds.isel(time=slice(1,2), x=slice(0,1)).to_zarr("test.zarr", mode="r+", region={"time": slice(1,2), "x": slice(0,1)}) ds.isel(time=slice(1,2), x=slice(1,2)).to_zarr("test.zarr", mode="r+", region={"time": slice(1,2), "x": slice(1,2)}) ds.isel(time=slice(1,2), x=slice(2,3)).to_zarr("test.zarr", mode="r+", region={"time": slice(1,2), "x": slice(2,3)}) ``` This example doesn't produce any error, but the |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
to_zarr: region not recognised as dataset dimensions 1077079208 | |
1052240616 | https://github.com/pydata/xarray/issues/6069#issuecomment-1052240616 | https://api.github.com/repos/pydata/xarray/issues/6069 | IC_kwDOAMm_X84-t-ro | d70-t 6574622 | 2022-02-26T15:58:48Z | 2022-02-26T15:58:48Z | CONTRIBUTOR | I'm trying to picture some usage scenarios based on incrementally adding timesteps to data on store. I hope these might help to answer questions from above. In particular, I think that I'll use the following dataset for demonstration code:
|
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
to_zarr: region not recognised as dataset dimensions 1077079208 | |
1034170927 | https://github.com/pydata/xarray/pull/6260#issuecomment-1034170927 | https://api.github.com/repos/pydata/xarray/issues/6260 | IC_kwDOAMm_X849pDIv | d70-t 6574622 | 2022-02-09T20:38:27Z | 2022-02-09T20:39:07Z | CONTRIBUTOR | I'm wondering what the right option for this case would be:
|
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
allow coordinates to be independent of `region` selection in to_zarr 1128821318 | |
1034098265 | https://github.com/pydata/xarray/issues/6259#issuecomment-1034098265 | https://api.github.com/repos/pydata/xarray/issues/6259 | IC_kwDOAMm_X849oxZZ | d70-t 6574622 | 2022-02-09T19:05:44Z | 2022-02-09T19:05:44Z | CONTRIBUTOR | This sounds like it could theoretically be handled using intake derived datasets. To be fair, derived datasets are probably still in their early stages. But the basic idea would be to apply arbitrary transformations to a dataset after it has been opened (e.g. with |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Be able to override calendar in `open_dataset`/`open_mfdataset`/etc OR include another calendar name 1128759050 | |
1033803014 | https://github.com/pydata/xarray/pull/6258#issuecomment-1033803014 | https://api.github.com/repos/pydata/xarray/issues/6258 | IC_kwDOAMm_X849npUG | d70-t 6574622 | 2022-02-09T14:12:21Z | 2022-02-09T14:12:21Z | CONTRIBUTOR | Indeed, those variable names have been quite unfortunate! I've changed them to |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
removed check for last dask chunk size in to_zarr 1128485610 | |
1033781323 | https://github.com/pydata/xarray/pull/6258#issuecomment-1033781323 | https://api.github.com/repos/pydata/xarray/issues/6258 | IC_kwDOAMm_X849nkBL | d70-t 6574622 | 2022-02-09T13:50:13Z | 2022-02-09T13:50:13Z | CONTRIBUTOR | Thanks Ryan for having a look into this. Accidentally I didn't run enough of the tests locally before submitting the PR. I've now checked the failing tests and came to the conclusion that the previously existing tests had been overly restrictive and rewrote them to reflect more closely what I believe that we actually want. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
removed check for last dask chunk size in to_zarr 1128485610 | |
863972083 | https://github.com/pydata/xarray/issues/5490#issuecomment-863972083 | https://api.github.com/repos/pydata/xarray/issues/5490 | MDEyOklzc3VlQ29tbWVudDg2Mzk3MjA4Mw== | d70-t 6574622 | 2021-06-18T11:32:38Z | 2021-06-18T11:33:14Z | CONTRIBUTOR | I've checked your example files. This is mostly related to the fact, that the original data is encoded as Probably the scaling and adding is carried out in Possibly related issues are #4826 and #3020 |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Nan/ changed values in output when only reading data, saving and reading again 924676925 | |
863945975 | https://github.com/pydata/xarray/issues/5490#issuecomment-863945975 | https://api.github.com/repos/pydata/xarray/issues/5490 | MDEyOklzc3VlQ29tbWVudDg2Mzk0NTk3NQ== | d70-t 6574622 | 2021-06-18T10:44:38Z | 2021-06-18T10:44:38Z | CONTRIBUTOR | Are your input files on (exactly) the same grid? If not, combining the files might introduce In [2]: np.nan == np.nan Out[2]: False ``` Which is as it should be per IEEE 754. When writing out the files to netCDF, do you accidentally convert from 64bit float to 32bit float? |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Nan/ changed values in output when only reading data, saving and reading again 924676925 | |
863939178 | https://github.com/pydata/xarray/issues/5489#issuecomment-863939178 | https://api.github.com/repos/pydata/xarray/issues/5489 | MDEyOklzc3VlQ29tbWVudDg2MzkzOTE3OA== | d70-t 6574622 | 2021-06-18T10:32:10Z | 2021-06-18T10:32:10Z | CONTRIBUTOR | I think there's more to think of then the suggested solution. For example when opening remote datasets (e.g. OPeNDAP resources), the supplied path will be a string which does not refer to a local path. The decision if a supplied "path" is valid might thus require to find an appropriate IO backend and then ask the backend if the supplied "path" is a valid one. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Misleading error when opening file that does not exist 924559401 | |
863098508 | https://github.com/pydata/xarray/issues/5189#issuecomment-863098508 | https://api.github.com/repos/pydata/xarray/issues/5189 | MDEyOklzc3VlQ29tbWVudDg2MzA5ODUwOA== | d70-t 6574622 | 2021-06-17T09:49:48Z | 2021-06-17T09:49:48Z | CONTRIBUTOR | Pydap has several important fixes which have been merged into |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
KeyError pulling from Nasa server with Pydap 861684673 | |
824207037 | https://github.com/pydata/xarray/issues/1650#issuecomment-824207037 | https://api.github.com/repos/pydata/xarray/issues/1650 | MDEyOklzc3VlQ29tbWVudDgyNDIwNzAzNw== | d70-t 6574622 | 2021-04-21T16:46:54Z | 2021-06-15T16:18:54Z | CONTRIBUTOR | I'd be interested in this kind of thing as well. :+1: We are having long time series data, which we would like to access via opendap or zarr over HTTP. Currently, the |
{ "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Low memory/out-of-core index? 267628781 | |
797151374 | https://github.com/pydata/xarray/pull/4966#issuecomment-797151374 | https://api.github.com/repos/pydata/xarray/issues/4966 | MDEyOklzc3VlQ29tbWVudDc5NzE1MTM3NA== | d70-t 6574622 | 2021-03-12T00:38:47Z | 2021-03-12T00:38:47Z | CONTRIBUTOR | I don't know if this qualifies as "documentation", but according to this merged PR on the netcdf-c sources, this is how the thredds OPeNDAP server behaves, from which they conclude that netCDF should behave accordingly. I confirmed myself that this also is how my currently installed netCDF-C behaves. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
conventions: decode unsigned integers to signed if _Unsigned=false 817302678 | |
786615778 | https://github.com/pydata/xarray/issues/4954#issuecomment-786615778 | https://api.github.com/repos/pydata/xarray/issues/4954 | MDEyOklzc3VlQ29tbWVudDc4NjYxNTc3OA== | d70-t 6574622 | 2021-02-26T12:22:52Z | 2021-02-26T12:22:52Z | CONTRIBUTOR | Thanks @dcherian. I added a PR #4966 |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Handling of signed bytes from OPeNDAP via pydap 815858485 | |
676193395 | https://github.com/pydata/xarray/pull/4312#issuecomment-676193395 | https://api.github.com/repos/pydata/xarray/issues/4312 | MDEyOklzc3VlQ29tbWVudDY3NjE5MzM5NQ== | d70-t 6574622 | 2020-08-19T11:32:37Z | 2020-08-19T11:32:37Z | CONTRIBUTOR | Do you know why Read the Docs complains? And if this is related to the PR? |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
allow manual zarr encoding on unchunked dask dimensions 673513695 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
issue 12