issue_comments
4 rows where author_association = "MEMBER", issue = 576337745 and user = 1197350 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date)
issue 1
- Errors using to_zarr for an s3 store · 4 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
595732104 | https://github.com/pydata/xarray/issues/3831#issuecomment-595732104 | https://api.github.com/repos/pydata/xarray/issues/3831 | MDEyOklzc3VlQ29tbWVudDU5NTczMjEwNA== | rabernat 1197350 | 2020-03-06T11:44:02Z | 2020-03-06T11:44:02Z | MEMBER |
This is basically an integration problem. What we are lacking is a comprehensive set of integration tests for this ecosystem (xarray + dask + zarr + fsspec and all its implementations). Pangeo has served as a de facto point for this discussion, since we are using the whole stack. Some similar issues there are: - https://github.com/pangeo-data/pangeo/issues/767 (xarray + opendap) - https://github.com/pangeo-data/pangeo/issues/765 (xarray + zarr + s3fs) - https://github.com/pangeo-data/pangeo/issues/741 (xarray + zarr + gcsfs) - https://github.com/pangeo-data/pangeo/issues/691 etc... All of these libraries understandably want to push the issues somewhere else, since they tend to be complex and hard to reduce to a MCVE. But there are fundamental issues related to integration that have to be addressed somewhere.
Yes and no. The details of how xarray is talking to these stores may matter. Continuing our planned refactor of the backened classes to use entry points, and formalizing the interface for backends, should help surface problems. The way we do consolidated metadata, for example, is pretty ad hoc:
Better diagnostics and more informative errors is always good. But we don't want to do this randomly. I think it should be part of the backend refactor. When is that happening btw? 🙃 We can't hope to test every possible permutation of xarray / zarr store / fsspec implementation within xarray. One idea I have thought about is an "integration bot", who watches all of these libraries and runs its own integration tests. This bot could even be configured to watch the repos and comment on PRs. That would be a cool project! Would appreciate @jhamman's thoughts here. |
{ "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Errors using to_zarr for an s3 store 576337745 | |
595383819 | https://github.com/pydata/xarray/issues/3831#issuecomment-595383819 | https://api.github.com/repos/pydata/xarray/issues/3831 | MDEyOklzc3VlQ29tbWVudDU5NTM4MzgxOQ== | rabernat 1197350 | 2020-03-05T18:41:32Z | 2020-03-05T18:41:32Z | MEMBER |
Key question: did you restart your kernel or call The goal here is to drill down into the stack and find the point where s3fs is failing to update its cache correctly. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Errors using to_zarr for an s3 store 576337745 | |
595377795 | https://github.com/pydata/xarray/issues/3831#issuecomment-595377795 | https://api.github.com/repos/pydata/xarray/issues/3831 | MDEyOklzc3VlQ29tbWVudDU5NTM3Nzc5NQ== | rabernat 1197350 | 2020-03-05T18:27:53Z | 2020-03-05T18:27:53Z | MEMBER | In that case, I'm fairly certain it is https://github.com/dask/s3fs/issues/285. There is a bug in s3fs where it caches the directory listing and then doesn't update it again, even if you delete files. This would potential cause problems when trying to overwrite, since s3fs would think the objects are already there, even if they are deleted. The same bug means consolidated metadata usually doesn't work. Perhaps @martindurant can weigh in. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Errors using to_zarr for an s3 store 576337745 | |
595298921 | https://github.com/pydata/xarray/issues/3831#issuecomment-595298921 | https://api.github.com/repos/pydata/xarray/issues/3831 | MDEyOklzc3VlQ29tbWVudDU5NTI5ODkyMQ== | rabernat 1197350 | 2020-03-05T15:47:12Z | 2020-03-05T15:47:12Z | MEMBER | These are tricky issues because they involve the integration of at least three libraries (xarray, zarr, s3fs, and possibly dask as well). Are you using dask? There could be some issues with s3fs caching (see https://github.com/dask/s3fs/issues/285). If you start fresh on a new path with nothing in it (so you don't need |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Errors using to_zarr for an s3 store 576337745 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 1