issue_comments
7 rows where issue = 631085856 and user = 6042212 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: reactions, created_at (date), updated_at (date)
issue 1
- Document writing netcdf from xarray directly to S3 · 7 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
1453911083 | https://github.com/pydata/xarray/issues/4122#issuecomment-1453911083 | https://api.github.com/repos/pydata/xarray/issues/4122 | IC_kwDOAMm_X85WqOwr | martindurant 6042212 | 2023-03-03T18:12:01Z | 2023-03-03T18:12:01Z | CONTRIBUTOR |
No compression, encoding or chunking except for the one "append" dimension. |
{ "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Document writing netcdf from xarray directly to S3 631085856 | |
1453902381 | https://github.com/pydata/xarray/issues/4122#issuecomment-1453902381 | https://api.github.com/repos/pydata/xarray/issues/4122 | IC_kwDOAMm_X85WqMot | martindurant 6042212 | 2023-03-03T18:04:29Z | 2023-03-03T18:04:29Z | CONTRIBUTOR | scipy only reads/writes netcdf2/3 ( https://docs.scipy.org/doc/scipy/reference/generated/scipy.io.netcdf_file.html ), which is a very different and simpler format than netcdf4. The latter uses HDF5 as a container, and h5netcdf as the xarray engine. I guess "to_netcdf" is ambiguous. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Document writing netcdf from xarray directly to S3 631085856 | |
1453898602 | https://github.com/pydata/xarray/issues/4122#issuecomment-1453898602 | https://api.github.com/repos/pydata/xarray/issues/4122 | IC_kwDOAMm_X85WqLtq | martindurant 6042212 | 2023-03-03T18:01:30Z | 2023-03-03T18:01:30Z | CONTRIBUTOR |
This is netCDF3, in that case. If that's fine for you, no problem. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Document writing netcdf from xarray directly to S3 631085856 | |
1453558039 | https://github.com/pydata/xarray/issues/4122#issuecomment-1453558039 | https://api.github.com/repos/pydata/xarray/issues/4122 | IC_kwDOAMm_X85Wo4kX | martindurant 6042212 | 2023-03-03T13:48:09Z | 2023-03-03T13:48:09Z | CONTRIBUTOR | Maybe it is netCDF3? xarray is supposed to be able to determine the file type
|
{ "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Document writing netcdf from xarray directly to S3 631085856 | |
1400583499 | https://github.com/pydata/xarray/issues/4122#issuecomment-1400583499 | https://api.github.com/repos/pydata/xarray/issues/4122 | IC_kwDOAMm_X85TezVL | martindurant 6042212 | 2023-01-23T15:57:24Z | 2023-01-23T15:57:24Z | CONTRIBUTOR | Would you mind writing out long-hand the version that worked and the version that didn't? |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Document writing netcdf from xarray directly to S3 631085856 | |
1400545067 | https://github.com/pydata/xarray/issues/4122#issuecomment-1400545067 | https://api.github.com/repos/pydata/xarray/issues/4122 | IC_kwDOAMm_X85Tep8r | martindurant 6042212 | 2023-01-23T15:31:16Z | 2023-01-23T15:31:16Z | CONTRIBUTOR | I can confirm that something like the following does work, basically automating the "write local and then push" workflow:
Unfortunately, directly writing to the remote file without a local cached file is not supported, because HDF5 does not write in a linear way. |
{ "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Document writing netcdf from xarray directly to S3 631085856 | |
639777701 | https://github.com/pydata/xarray/issues/4122#issuecomment-639777701 | https://api.github.com/repos/pydata/xarray/issues/4122 | MDEyOklzc3VlQ29tbWVudDYzOTc3NzcwMQ== | martindurant 6042212 | 2020-06-05T20:17:38Z | 2020-06-05T20:17:38Z | CONTRIBUTOR | The write feature for simplecache isn't released yet, of course. It would be interesting if someone could subclass file and write locally with h5netcdf to see what kind of seeks it does. Is it popping back to some file header to update array sizes? Presumably it would need a fixed-size header to do that. Parquet and other cloud formats have the metadata at the footer exactly for this reason, so you only write once you know everything and you only ever move forward in the fie. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Document writing netcdf from xarray directly to S3 631085856 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 1