issue_comments
3 rows where author_association = "MEMBER", issue = 230566456 and user = 1217238 sorted by updated_at descending
This data as json, CSV (advanced)
These facets timed out: author_association, issue
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
303606683 | https://github.com/pydata/xarray/pull/1421#issuecomment-303606683 | https://api.github.com/repos/pydata/xarray/issues/1421 | MDEyOklzc3VlQ29tbWVudDMwMzYwNjY4Mw== | shoyer 1217238 | 2017-05-24T03:18:16Z | 2017-05-24T03:18:16Z | MEMBER | How about something like the following: In def maybe_encode_pickle(var):
if var.dtype == object:
attrs = var.attrs.copy()
safe_setitem('_FileFormat', 'python-pickle')
protocol = var.encoding.pop('pickle_protocol', 2)
data = utils.encode_pickle(var.values, protocol=protocol)
var = Variable(var.dims, data, attrs, var.encoding)
return var
In the netCDF backends, add a check for variable with For decoding, reverse the process. Convert custom vlen dtypes to For bonus points, generalize handling of vlen types with |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Adding arbitrary object serialization 230566456 | |
303583082 | https://github.com/pydata/xarray/pull/1421#issuecomment-303583082 | https://api.github.com/repos/pydata/xarray/issues/1421 | MDEyOklzc3VlQ29tbWVudDMwMzU4MzA4Mg== | shoyer 1217238 | 2017-05-24T00:52:20Z | 2017-05-24T00:52:20Z | MEMBER |
I think we do want some sort of marker attribute, but I agree that it doesn't need to include the pickle version. Maybe the attribute
I think netCDF actually maps Certainly handling opaque types in netCDF4-python would be nice, though I don't think it should be a blocker for this. I suspect the reason this isn't done is that NumPy maps |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Adding arbitrary object serialization 230566456 | |
303541907 | https://github.com/pydata/xarray/pull/1421#issuecomment-303541907 | https://api.github.com/repos/pydata/xarray/issues/1421 | MDEyOklzc3VlQ29tbWVudDMwMzU0MTkwNw== | shoyer 1217238 | 2017-05-23T21:46:39Z | 2017-05-23T21:48:18Z | MEMBER | Thanks for giving this a shot!
I'm having a hard time imagining any other serialization formats for serializing arbitrary Python objects. One addition reason for favoring
Yes, this is a little tricky. The current design is not great here. Ideally, though, we would still keep all of the encoding/decoding logic separate from the datastores. I need to think about this a little more. One other concern is how to represent this data on disk in netCDF/HDF5 variables. Ideally, we would have a format that could work -- at least in principle -- with Annoyingly, these libraries currently have incompatible dtype support:
So if we want something that works with both, we'll need to add some additional metadata field in the form of an attribute to indicate how do decoding. Maybe something like I have some inline comments I'll add below. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Adding arbitrary object serialization 230566456 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 1