issue_comments
8 rows where issue = 304624171 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date)
issue 1
- Load a small subset of data from a big dataset takes forever · 8 ✖
| id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
|---|---|---|---|---|---|---|---|---|---|---|---|
| 373694632 | https://github.com/pydata/xarray/issues/1985#issuecomment-373694632 | https://api.github.com/repos/pydata/xarray/issues/1985 | MDEyOklzc3VlQ29tbWVudDM3MzY5NDYzMg== | malmans2 22245117 | 2018-03-16T12:09:50Z | 2018-03-16T12:09:50Z | CONTRIBUTOR | Alright, I found the problem. I'm loading several variables from different files. All the variables have 1464 snapshots. However, one of the 3D variables has just one snapshot at a different time (I found a bag in my bash script to re-organize the raw data). When I load my dataset using .open_mfdataset, the time dimension has an extra snapshot (length is 1465). However, xarray doesn't like it and when I run functions such as to_netcdf it takes forever (no error). Thanks @fujiisoup for the help! |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Load a small subset of data from a big dataset takes forever 304624171 | |
| 372852604 | https://github.com/pydata/xarray/issues/1985#issuecomment-372852604 | https://api.github.com/repos/pydata/xarray/issues/1985 | MDEyOklzc3VlQ29tbWVudDM3Mjg1MjYwNA== | fujiisoup 6815844 | 2018-03-13T23:24:37Z | 2018-03-13T23:24:37Z | MEMBER | I see no problem with your code... Can you try updating xarray to 0.10.2 (released today)? We updated some logic of lazy indexing. |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Load a small subset of data from a big dataset takes forever 304624171 | |
| 372570107 | https://github.com/pydata/xarray/issues/1985#issuecomment-372570107 | https://api.github.com/repos/pydata/xarray/issues/1985 | MDEyOklzc3VlQ29tbWVudDM3MjU3MDEwNw== | malmans2 22245117 | 2018-03-13T07:21:10Z | 2018-03-13T07:21:10Z | CONTRIBUTOR | I forgot to mention that I'm getting this warning: /home/idies/anaconda3/lib/python3.5/site-packages/dask/core.py:306: FutureWarning: elementwise comparison failed; returning scalar instead, but in the future will perform elementwise comparison elif type_arg is type(key) and arg == key: However, I don't think it is relevant since I get the same warning when I'm able to run .to_netcdf() on the 3D variable. |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Load a small subset of data from a big dataset takes forever 304624171 | |
| 372566304 | https://github.com/pydata/xarray/issues/1985#issuecomment-372566304 | https://api.github.com/repos/pydata/xarray/issues/1985 | MDEyOklzc3VlQ29tbWVudDM3MjU2NjMwNA== | malmans2 22245117 | 2018-03-13T07:01:51Z | 2018-03-13T07:01:51Z | CONTRIBUTOR | The problem occurs when I run the very last line, which is to_netcdf().
Right before, the dataset looks like this:
|
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Load a small subset of data from a big dataset takes forever 304624171 | |
| 372563938 | https://github.com/pydata/xarray/issues/1985#issuecomment-372563938 | https://api.github.com/repos/pydata/xarray/issues/1985 | MDEyOklzc3VlQ29tbWVudDM3MjU2MzkzOA== | fujiisoup 6815844 | 2018-03-13T06:48:23Z | 2018-03-13T06:48:23Z | MEMBER | Umm. I could not find what is wrong with your code.
Can you find which line loads the data into memory?
If your data is still a dask array, it does not print the entries of the array but instead, it shows something like this,
|
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Load a small subset of data from a big dataset takes forever 304624171 | |
| 372558850 | https://github.com/pydata/xarray/issues/1985#issuecomment-372558850 | https://api.github.com/repos/pydata/xarray/issues/1985 | MDEyOklzc3VlQ29tbWVudDM3MjU1ODg1MA== | malmans2 22245117 | 2018-03-13T06:19:47Z | 2018-03-13T06:23:00Z | CONTRIBUTOR | I have the same issue if I don't copy the dataset. Here are the coordinates of my dataset:
``` I think somewhere I trigger the loading of the whole dataset. Otherwise, I don't understand why it works when I open just one month instead of the whole year. |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Load a small subset of data from a big dataset takes forever 304624171 | |
| 372545491 | https://github.com/pydata/xarray/issues/1985#issuecomment-372545491 | https://api.github.com/repos/pydata/xarray/issues/1985 | MDEyOklzc3VlQ29tbWVudDM3MjU0NTQ5MQ== | fujiisoup 6815844 | 2018-03-13T04:44:52Z | 2018-03-13T04:48:56Z | MEMBER | I notice this line
|
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Load a small subset of data from a big dataset takes forever 304624171 | |
| 372544809 | https://github.com/pydata/xarray/issues/1985#issuecomment-372544809 | https://api.github.com/repos/pydata/xarray/issues/1985 | MDEyOklzc3VlQ29tbWVudDM3MjU0NDgwOQ== | fujiisoup 6815844 | 2018-03-13T04:39:47Z | 2018-03-13T04:39:47Z | MEMBER |
I don't think so. We support lazy indexing for any dimensional arrays (but not coordinate variables).
What does your data (especially '4Dvariable.nc') look like?
Is |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Load a small subset of data from a big dataset takes forever 304624171 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] (
[html_url] TEXT,
[issue_url] TEXT,
[id] INTEGER PRIMARY KEY,
[node_id] TEXT,
[user] INTEGER REFERENCES [users]([id]),
[created_at] TEXT,
[updated_at] TEXT,
[author_association] TEXT,
[body] TEXT,
[reactions] TEXT,
[performed_via_github_app] TEXT,
[issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
ON [issue_comments] ([user]);
user 2