issue_comments
2 rows where author_association = "MEMBER" and issue = 503578688 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date)
issue 1
- implement normalize_token · 2 ✖
 
| id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue | 
|---|---|---|---|---|---|---|---|---|---|---|---|
| 541315926 | https://github.com/pydata/xarray/issues/3378#issuecomment-541315926 | https://api.github.com/repos/pydata/xarray/issues/3378 | MDEyOklzc3VlQ29tbWVudDU0MTMxNTkyNg== | crusaderky 6213168 | 2019-10-12T11:27:52Z | 2019-10-12T11:38:11Z | MEMBER | https://docs.dask.org/en/latest/custom-collections.html#implementing-deterministic-hashing ```python @normalize_token.register(Dataset) def tokenize_dataset(ds): return Dataset, ds._variables, ds._coord_names, ds._attrs @normalize_token.register(DataArray) def tokenize_dataarray(da): return DataArray, ds._variable, ds._coords, ds._name Note: the @singledispatch for IndexVariable must be defined before the one for Variable@normalize_token.register(IndexVariable) def tokenize_indexvariable(v): # Don't waste time converting pd.Index to np.ndarray return IndexVariable, v._dims, v._data.array, v._attrs @normalize_token.register(Variable) def tokenize_variable(v): # Note: it's v.data, not v._data, in order to cope with the # wrappers around NetCDF and the like return Variable, v._dims, v.data, v._attrs ``` You'll need to write a dummy normalize_token for when dask is not installed. Unit tests: - running tokenize() twice on the same object returns the same result - changing the content of a data_var (or the variable, for DataArray) changes the output - changing the content of a coord changes the output - changing attrs, name, or dimension names change the output - whether a variable is a data_var or a coord changes the output - dask arrays aren't computed - non-numpy, non-dask NEP18 data is not converted to numpy - works with xarray's fancy wrappers around NetCDF and the like  | 
                
                    {
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
} | 
                
                    implement normalize_token 503578688 | |
| 541121326 | https://github.com/pydata/xarray/issues/3378#issuecomment-541121326 | https://api.github.com/repos/pydata/xarray/issues/3378 | MDEyOklzc3VlQ29tbWVudDU0MTEyMTMyNg== | dcherian 2448579 | 2019-10-11T15:55:23Z | 2019-10-11T15:55:23Z | MEMBER | How should this be implemented?  | 
                
                    {
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
} | 
                
                    implement normalize_token 503578688 | 
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [performed_via_github_app] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
    ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
    ON [issue_comments] ([user]);
user 2