issue_comments
5 rows where author_association = "MEMBER" and issue = 361016974 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date)
issue 1
- Limiting threads/cores used by xarray(/dask?) · 5 ✖
| id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
|---|---|---|---|---|---|---|---|---|---|---|---|
| 460393715 | https://github.com/pydata/xarray/issues/2417#issuecomment-460393715 | https://api.github.com/repos/pydata/xarray/issues/2417 | MDEyOklzc3VlQ29tbWVudDQ2MDM5MzcxNQ== | jhamman 2443309 | 2019-02-04T20:07:56Z | 2019-02-04T20:07:56Z | MEMBER | @Zeitsperre - are you still having problems in this area? If not, is okay if we close this issue? |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Limiting threads/cores used by xarray(/dask?) 361016974 | |
| 460298993 | https://github.com/pydata/xarray/issues/2417#issuecomment-460298993 | https://api.github.com/repos/pydata/xarray/issues/2417 | MDEyOklzc3VlQ29tbWVudDQ2MDI5ODk5Mw== | jhamman 2443309 | 2019-02-04T15:50:09Z | 2019-02-04T15:51:43Z | MEMBER | On a few systems, I've noticed that I need to set the environment variable xref: https://stackoverflow.com/questions/39422092/error-with-omp-num-threads-when-using-dask-distributed |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Limiting threads/cores used by xarray(/dask?) 361016974 | |
| 460020879 | https://github.com/pydata/xarray/issues/2417#issuecomment-460020879 | https://api.github.com/repos/pydata/xarray/issues/2417 | MDEyOklzc3VlQ29tbWVudDQ2MDAyMDg3OQ== | jhamman 2443309 | 2019-02-03T03:54:59Z | 2019-02-03T03:54:59Z | MEMBER | @Zeitsperre - this issue has been inactive for a while. Did you find a solution to y our problem? |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Limiting threads/cores used by xarray(/dask?) 361016974 | |
| 422461245 | https://github.com/pydata/xarray/issues/2417#issuecomment-422461245 | https://api.github.com/repos/pydata/xarray/issues/2417 | MDEyOklzc3VlQ29tbWVudDQyMjQ2MTI0NQ== | shoyer 1217238 | 2018-09-18T16:31:03Z | 2018-09-18T16:31:03Z | MEMBER | If your data using in-file HDF5 chunks/compression it's possible that HDF5 is uncompressing the data is parallel, though I haven't seen that before personally. |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Limiting threads/cores used by xarray(/dask?) 361016974 | |
| 422206083 | https://github.com/pydata/xarray/issues/2417#issuecomment-422206083 | https://api.github.com/repos/pydata/xarray/issues/2417 | MDEyOklzc3VlQ29tbWVudDQyMjIwNjA4Mw== | shoyer 1217238 | 2018-09-17T23:40:52Z | 2018-09-17T23:40:52Z | MEMBER | Step 1 would be making sure that you're actually using dask :). Xarray only uses dask with That said, xarray's only built-in support for parallelism is through Dask, so I'm not sure what is using all your CPU. |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Limiting threads/cores used by xarray(/dask?) 361016974 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] (
[html_url] TEXT,
[issue_url] TEXT,
[id] INTEGER PRIMARY KEY,
[node_id] TEXT,
[user] INTEGER REFERENCES [users]([id]),
[created_at] TEXT,
[updated_at] TEXT,
[author_association] TEXT,
[body] TEXT,
[reactions] TEXT,
[performed_via_github_app] TEXT,
[issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
ON [issue_comments] ([user]);
user 2