issue_comments
5 rows where author_association = "MEMBER" and issue = 361016974 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date)
issue 1
- Limiting threads/cores used by xarray(/dask?) · 5 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
460393715 | https://github.com/pydata/xarray/issues/2417#issuecomment-460393715 | https://api.github.com/repos/pydata/xarray/issues/2417 | MDEyOklzc3VlQ29tbWVudDQ2MDM5MzcxNQ== | jhamman 2443309 | 2019-02-04T20:07:56Z | 2019-02-04T20:07:56Z | MEMBER | @Zeitsperre - are you still having problems in this area? If not, is okay if we close this issue? |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Limiting threads/cores used by xarray(/dask?) 361016974 | |
460298993 | https://github.com/pydata/xarray/issues/2417#issuecomment-460298993 | https://api.github.com/repos/pydata/xarray/issues/2417 | MDEyOklzc3VlQ29tbWVudDQ2MDI5ODk5Mw== | jhamman 2443309 | 2019-02-04T15:50:09Z | 2019-02-04T15:51:43Z | MEMBER | On a few systems, I've noticed that I need to set the environment variable xref: https://stackoverflow.com/questions/39422092/error-with-omp-num-threads-when-using-dask-distributed |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Limiting threads/cores used by xarray(/dask?) 361016974 | |
460020879 | https://github.com/pydata/xarray/issues/2417#issuecomment-460020879 | https://api.github.com/repos/pydata/xarray/issues/2417 | MDEyOklzc3VlQ29tbWVudDQ2MDAyMDg3OQ== | jhamman 2443309 | 2019-02-03T03:54:59Z | 2019-02-03T03:54:59Z | MEMBER | @Zeitsperre - this issue has been inactive for a while. Did you find a solution to y our problem? |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Limiting threads/cores used by xarray(/dask?) 361016974 | |
422461245 | https://github.com/pydata/xarray/issues/2417#issuecomment-422461245 | https://api.github.com/repos/pydata/xarray/issues/2417 | MDEyOklzc3VlQ29tbWVudDQyMjQ2MTI0NQ== | shoyer 1217238 | 2018-09-18T16:31:03Z | 2018-09-18T16:31:03Z | MEMBER | If your data using in-file HDF5 chunks/compression it's possible that HDF5 is uncompressing the data is parallel, though I haven't seen that before personally. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Limiting threads/cores used by xarray(/dask?) 361016974 | |
422206083 | https://github.com/pydata/xarray/issues/2417#issuecomment-422206083 | https://api.github.com/repos/pydata/xarray/issues/2417 | MDEyOklzc3VlQ29tbWVudDQyMjIwNjA4Mw== | shoyer 1217238 | 2018-09-17T23:40:52Z | 2018-09-17T23:40:52Z | MEMBER | Step 1 would be making sure that you're actually using dask :). Xarray only uses dask with That said, xarray's only built-in support for parallelism is through Dask, so I'm not sure what is using all your CPU. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Limiting threads/cores used by xarray(/dask?) 361016974 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 2