issue_comments
4 rows where issue = 1373352524 and user = 60435591 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: reactions, created_at (date), updated_at (date)
issue 1
- Encoding error when saving netcdf · 4 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
1460160756 | https://github.com/pydata/xarray/issues/7039#issuecomment-1460160756 | https://api.github.com/repos/pydata/xarray/issues/7039 | IC_kwDOAMm_X85XCEj0 | veenstrajelmer 60435591 | 2023-03-08T13:32:02Z | 2023-03-08T13:32:02Z | CONTRIBUTOR | Hi @etsmith14. The suggestion I did loses accuracy and depending on the variable this is not acceptable. However, recomputing |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Encoding error when saving netcdf 1373352524 | |
1458582799 | https://github.com/pydata/xarray/issues/7039#issuecomment-1458582799 | https://api.github.com/repos/pydata/xarray/issues/7039 | IC_kwDOAMm_X85W8DUP | veenstrajelmer 60435591 | 2023-03-07T17:47:20Z | 2023-03-07T22:48:36Z | CONTRIBUTOR | @etsmith14: another workaround is removing the |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Encoding error when saving netcdf 1373352524 | |
1438244290 | https://github.com/pydata/xarray/issues/7039#issuecomment-1438244290 | https://api.github.com/repos/pydata/xarray/issues/7039 | IC_kwDOAMm_X85Vud3C | veenstrajelmer 60435591 | 2023-02-21T10:36:48Z | 2023-02-21T10:50:20Z | CONTRIBUTOR | I have been thinking about a desireable solution, but I have a bit of trouble with it. Besides removing dtype from encoding (resulting in floats being written), one could also change the scale_factor to a higher value (e.g. 0.5). Writing this to int does take half the disksize than releasing the int restriction and writing it to float32. Whatever you do, the data is altered at least slightly. Apparently, the data cannot be properly written to integers after reading it. This is a bit odd I would say, would that mean that the scaling+offset of ERA5 data is that thightly chosen that when applying it to another dataset/month, the data would fall out of the integer reach? Would be great if this would "just work". At the moment, apparently reading and writing ERA5 data with xarray results in incorrect netcdf files. I expected xarray would work off the shelf with these type of data, it feels like xarray is designed for doing exactly these type of things. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Encoding error when saving netcdf 1373352524 | |
1432943981 | https://github.com/pydata/xarray/issues/7039#issuecomment-1432943981 | https://api.github.com/repos/pydata/xarray/issues/7039 | IC_kwDOAMm_X85VaP1t | veenstrajelmer 60435591 | 2023-02-16T11:29:45Z | 2023-02-16T11:29:45Z | CONTRIBUTOR | I have also encountered an issue with reading of ERA5 data with open_mfdatset, writing it to_netcdf() and reading it again (https://github.com/Deltares/dfm_tools/issues/239). I was actually looking for a place to land this, and found your issue. My expectation is that this is because the ERA5 data is saved as ints, but all files have different offsets/scalingfactors. Upon opening it with open_mfdataset(), the data is converted to floats and to the offset/scalingfactor of the first file. This is fine, but the issue occurs I think (and what you also mention) since {'dtype': 'int16'} is in the encoding. The file is written as ints and this seems to mess up the data. (all a theory) A workaround is to remove the dtype from the encoding for all variables in the file (or update to float32), but that seems cumbersome. |
{ "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Encoding error when saving netcdf 1373352524 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 1