issue_comments
7 rows where issue = 180451196 and user = 1217238 sorted by updated_at descending
This data as json, CSV (advanced)
issue 1
- Disable automatic cache with dask · 7 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
261046945 | https://github.com/pydata/xarray/pull/1024#issuecomment-261046945 | https://api.github.com/repos/pydata/xarray/issues/1024 | MDEyOklzc3VlQ29tbWVudDI2MTA0Njk0NQ== | shoyer 1217238 | 2016-11-16T19:30:53Z | 2016-11-16T19:30:53Z | MEMBER | @kynan I think this is fixed in #1128, which has a slightly more robust solution. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Disable automatic cache with dask 180451196 | |
260393416 | https://github.com/pydata/xarray/pull/1024#issuecomment-260393416 | https://api.github.com/repos/pydata/xarray/issues/1024 | MDEyOklzc3VlQ29tbWVudDI2MDM5MzQxNg== | shoyer 1217238 | 2016-11-14T16:57:59Z | 2016-11-14T16:57:59Z | MEMBER | Thanks for your patience! This is a nice improvement. I have an idea for a variation that might make for a cleaner (less dask specific) way to handle the remaining caching logic -- I'll add you a reviewer on that PR. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Disable automatic cache with dask 180451196 | |
258679497 | https://github.com/pydata/xarray/pull/1024#issuecomment-258679497 | https://api.github.com/repos/pydata/xarray/issues/1024 | MDEyOklzc3VlQ29tbWVudDI1ODY3OTQ5Nw== | shoyer 1217238 | 2016-11-06T13:01:50Z | 2016-11-06T13:01:50Z | MEMBER | Awesome, thanks for your help! On Sat, Nov 5, 2016 at 6:56 PM crusaderky notifications@github.com wrote:
|
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Disable automatic cache with dask 180451196 | |
258620328 | https://github.com/pydata/xarray/pull/1024#issuecomment-258620328 | https://api.github.com/repos/pydata/xarray/issues/1024 | MDEyOklzc3VlQ29tbWVudDI1ODYyMDMyOA== | shoyer 1217238 | 2016-11-05T15:53:06Z | 2016-11-05T15:53:06Z | MEMBER |
Yes, please do! @crusaderky I think we are OK going ahead here with Option D. If we do eventually extend xarray with out of core indexes, that will be done with a separate layer (not in |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Disable automatic cache with dask 180451196 | |
258524115 | https://github.com/pydata/xarray/pull/1024#issuecomment-258524115 | https://api.github.com/repos/pydata/xarray/issues/1024 | MDEyOklzc3VlQ29tbWVudDI1ODUyNDExNQ== | shoyer 1217238 | 2016-11-04T19:19:00Z | 2016-11-04T19:19:00Z | MEMBER |
Right now, xarray is not going to be great fit for such cases, because we already cache an index in memory for any labeled indexing operations. So at best, you could do something like I doubt very many people are relying on this, though indeed, this would include some users of an array database we wrote at my former employer, which did not support different chunking schemes for different variables (which could make coordinate lookup very slow). CC @ToddSmall in case he has opinions here. For out-of-core operations with labels on big unstructured meshes, you really need a generalization of the |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Disable automatic cache with dask 180451196 | |
256125722 | https://github.com/pydata/xarray/pull/1024#issuecomment-256125722 | https://api.github.com/repos/pydata/xarray/issues/1024 | MDEyOklzc3VlQ29tbWVudDI1NjEyNTcyMg== | shoyer 1217238 | 2016-10-25T18:25:30Z | 2016-10-25T18:25:30Z | MEMBER | I'm going to ping the mailing list for input, but I think it would be pretty safe. On Tue, Oct 25, 2016 at 11:11 AM, crusaderky notifications@github.com wrote:
|
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Disable automatic cache with dask 180451196 | |
255286001 | https://github.com/pydata/xarray/pull/1024#issuecomment-255286001 | https://api.github.com/repos/pydata/xarray/issues/1024 | MDEyOklzc3VlQ29tbWVudDI1NTI4NjAwMQ== | shoyer 1217238 | 2016-10-21T03:36:01Z | 2016-10-21T03:36:01Z | MEMBER |
I'm nervous about eager loading, especially for non-index coordinates. They can have more than one dimension, and thus can contain a lot of data. So potentially eagerly loading non-index coordinates could break existing use cases. On the other hand, non-index coordinates indeed checked for equality in most xarray operations (e.g., for the coordinate merge in align). So it is indeed useful not to have to recompute them all the time. Even eagerly loading indexes is potentially problematic, if loading the index values is expensive. So I'm conflicted:
- I like the current caching behavior for I'm going to start throwing out ideas for how to deal with this: Option AAdd two new (public?) methods, something like Hypothetically, we could even have options for turning this caching systematically on/off (e.g., Your proposal is basically an extreme version of this, where we call Advantages:
- It's fairly predictable when caching happens (especially if we opt for calling Downsides: - Caching is more aggressive than necessary -- we cache indexes even if that coord isn't actually indexed. Option BLike Option A, but someone infer the full set of variables that need to be cached (e.g., in a This solves the downside of A, but diminishes the predictability. We're basically back to how things work now. Option CCache dask.array in Advantages: - Much simpler and easier to implement than the alternatives. - Implicit conversions are greatly diminished. Downsides:
- Non-index coordinates get thrown away after being evaluated once. If you're doing lots of operations of the form Option DLoad the contents of an This has the most predictable performance, but might cause trouble for some edge use cases? I need to think about this a little more, but right now I am leaning towards Option C or D. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Disable automatic cache with dask 180451196 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 1