issue_comments
4 rows where issue = 287223508 and user = 8881170 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: reactions, created_at (date), updated_at (date)
issue 1
- apply_ufunc(dask='parallelized') with multiple outputs · 4 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
628135082 | https://github.com/pydata/xarray/issues/1815#issuecomment-628135082 | https://api.github.com/repos/pydata/xarray/issues/1815 | MDEyOklzc3VlQ29tbWVudDYyODEzNTA4Mg== | bradyrx 8881170 | 2020-05-13T17:27:06Z | 2020-05-13T17:27:06Z | CONTRIBUTOR |
Good call. I figured there was a workaround. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
apply_ufunc(dask='parallelized') with multiple outputs 287223508 | |
628070696 | https://github.com/pydata/xarray/issues/1815#issuecomment-628070696 | https://api.github.com/repos/pydata/xarray/issues/1815 | MDEyOklzc3VlQ29tbWVudDYyODA3MDY5Ng== | bradyrx 8881170 | 2020-05-13T15:33:56Z | 2020-05-13T15:33:56Z | CONTRIBUTOR | One issue I see is that this would return multiple dask objects, correct? So to get the results from them, you'd have to run The earlier mentioned code snippets provide a nice path forward, since you can just run compute on one object, and then split its |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
apply_ufunc(dask='parallelized') with multiple outputs 287223508 | |
614244205 | https://github.com/pydata/xarray/issues/1815#issuecomment-614244205 | https://api.github.com/repos/pydata/xarray/issues/1815 | MDEyOklzc3VlQ29tbWVudDYxNDI0NDIwNQ== | bradyrx 8881170 | 2020-04-15T19:45:50Z | 2020-04-15T19:45:50Z | CONTRIBUTOR | I think ideally it would be nice to return multiple DataArrays or a Dataset of variables. But I'm really happy with this solution. I'm using it on a 600GB dataset of particle trajectories and was able to write a ufunc to go through and return each particle's x, y, z location when it met a certain condition. I think having something simple like the stackoverflow snippet I posted would be great for the docs as an |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
apply_ufunc(dask='parallelized') with multiple outputs 287223508 | |
614216243 | https://github.com/pydata/xarray/issues/1815#issuecomment-614216243 | https://api.github.com/repos/pydata/xarray/issues/1815 | MDEyOklzc3VlQ29tbWVudDYxNDIxNjI0Mw== | bradyrx 8881170 | 2020-04-15T18:49:51Z | 2020-04-15T18:49:51Z | CONTRIBUTOR | This looks essentially the same to @stefraynaud's answer, but I came across this stackoverflow response here: https://stackoverflow.com/questions/52094320/with-xarray-how-to-parallelize-1d-operations-on-a-multidimensional-dataset. @andersy005, I imagine you're far past this now. And this might have been related to discussions with Genevieve and I anyways. ```python def new_linregress(x, y): # Wrapper around scipy linregress to use in apply_ufunc slope, intercept, r_value, p_value, std_err = stats.linregress(x, y) return np.array([slope, intercept, r_value, p_value, std_err]) return a new DataArraystats = xr.apply_ufunc(new_linregress, ds[x], ds[y], input_core_dims=[['year'], ['year']], output_core_dims=[["parameter"]], vectorize=True, dask="parallelized", output_dtypes=['float64'], output_sizes={"parameter": 5}, ) ``` |
{ "total_count": 3, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 3, "rocket": 0, "eyes": 0 } |
apply_ufunc(dask='parallelized') with multiple outputs 287223508 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 1