issue_comments
14 rows where user = 2560426 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: issue_url, reactions, created_at (date), updated_at (date)
user 1
- heerad · 14 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
778841149 | https://github.com/pydata/xarray/issues/3961#issuecomment-778841149 | https://api.github.com/repos/pydata/xarray/issues/3961 | MDEyOklzc3VlQ29tbWVudDc3ODg0MTE0OQ== | heerad 2560426 | 2021-02-14T21:01:21Z | 2021-02-14T21:01:21Z | NONE |
To clarify, do you mean adding a sleep of e.g. 1 second prior to your Is this solution only addressing the issue of opening the same ds multiple times within a python process, or would it also address multiple processes opening the same ds? |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Hangs while saving netcdf file opened using xr.open_mfdataset with lock=None 597657663 | |
778838527 | https://github.com/pydata/xarray/issues/3961#issuecomment-778838527 | https://api.github.com/repos/pydata/xarray/issues/3961 | MDEyOklzc3VlQ29tbWVudDc3ODgzODUyNw== | heerad 2560426 | 2021-02-14T20:40:38Z | 2021-02-14T20:40:38Z | NONE | Also seeing this as of version 0.16.1. In some cases, I need In other cases, I need Is the current recommended solution to set |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Hangs while saving netcdf file opened using xr.open_mfdataset with lock=None 597657663 | |
713172015 | https://github.com/pydata/xarray/issues/4482#issuecomment-713172015 | https://api.github.com/repos/pydata/xarray/issues/4482 | MDEyOklzc3VlQ29tbWVudDcxMzE3MjAxNQ== | heerad 2560426 | 2020-10-20T22:17:08Z | 2020-10-20T22:21:14Z | NONE | On the topic of fillna(), I'm seeing an odd unrelated issue that I don't have an explanation for. I have a dataarray When I do
Stack trace shows it's failing on a I have no idea how to reproduce this simply... If it helps narrow things down, |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Allow skipna in .dot() 713834297 | |
708474940 | https://github.com/pydata/xarray/issues/4482#issuecomment-708474940 | https://api.github.com/repos/pydata/xarray/issues/4482 | MDEyOklzc3VlQ29tbWVudDcwODQ3NDk0MA== | heerad 2560426 | 2020-10-14T15:21:29Z | 2020-10-14T15:21:55Z | NONE | Adding on, whatever the solution is that avoids blowing up memory, especially when using with |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Allow skipna in .dot() 713834297 | |
707331260 | https://github.com/pydata/xarray/issues/4482#issuecomment-707331260 | https://api.github.com/repos/pydata/xarray/issues/4482 | MDEyOklzc3VlQ29tbWVudDcwNzMzMTI2MA== | heerad 2560426 | 2020-10-12T20:31:26Z | 2020-10-12T21:05:24Z | NONE | See below. I temporarily write some files to netcdf then recombine them lazily using The issue seems to present itself more consistently when my I used the ``` import numpy as np import xarray as xr import os N = 1000 N_per_file = 10 M = 100 K = 10 window_size = 150 tmp_dir = 'tmp' os.mkdir(tmp_dir) save many netcdf files, later to be concatted into a dask.delayed datasetfor i in range(0, N, N_per_file):
open lazilyx = xr.open_mfdataset('{}/*.nc'.format(tmp_dir), parallel=True, concat_dim='d1').vals a rolling window along a stacked dimensionx_windows = x.stack(d13=['d1', 'd3']).rolling(d13=window_size).construct('window') we'll dot x_windows with y along the window dimensiony = xr.DataArray([1]*window_size, dims='window') incremental memory: 1.94 MiBx_windows.dot(y).compute() incremental memory: 20.00 MiBx_windows.notnull().dot(y).compute() incremental memory: 182.13 MiBx_windows.fillna(0.).dot(y).compute() incremental memory: 211.52 MiBx_windows.weighted(y).mean('window', skipna=True).compute() ``` |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Allow skipna in .dot() 713834297 | |
707238146 | https://github.com/pydata/xarray/issues/4482#issuecomment-707238146 | https://api.github.com/repos/pydata/xarray/issues/4482 | MDEyOklzc3VlQ29tbWVudDcwNzIzODE0Ng== | heerad 2560426 | 2020-10-12T17:01:54Z | 2020-10-12T17:16:07Z | NONE | Adding on here, even if This is happening with More evidence in favor: if I do
I'm happy to live with a memory copy for now with |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Allow skipna in .dot() 713834297 | |
702939943 | https://github.com/pydata/xarray/issues/4482#issuecomment-702939943 | https://api.github.com/repos/pydata/xarray/issues/4482 | MDEyOklzc3VlQ29tbWVudDcwMjkzOTk0Mw== | heerad 2560426 | 2020-10-02T20:20:53Z | 2020-10-02T20:32:32Z | NONE | Great, looks like I missed that option. Thanks. For reference, |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Allow skipna in .dot() 713834297 | |
702346076 | https://github.com/pydata/xarray/issues/4474#issuecomment-702346076 | https://api.github.com/repos/pydata/xarray/issues/4474 | MDEyOklzc3VlQ29tbWVudDcwMjM0NjA3Ng== | heerad 2560426 | 2020-10-01T19:20:50Z | 2020-10-01T19:23:31Z | NONE | { "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Implement rolling_exp for dask arrays 712052219 | ||
702331156 | https://github.com/pydata/xarray/issues/4474#issuecomment-702331156 | https://api.github.com/repos/pydata/xarray/issues/4474 | MDEyOklzc3VlQ29tbWVudDcwMjMzMTE1Ng== | heerad 2560426 | 2020-10-01T18:52:18Z | 2020-10-01T18:52:18Z | NONE | Yes, see http://xarray.pydata.org/en/stable/computation.html#rolling-window-operations.
|
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Implement rolling_exp for dask arrays 712052219 | |
702307334 | https://github.com/pydata/xarray/issues/4475#issuecomment-702307334 | https://api.github.com/repos/pydata/xarray/issues/4475 | MDEyOklzc3VlQ29tbWVudDcwMjMwNzMzNA== | heerad 2560426 | 2020-10-01T18:07:55Z | 2020-10-01T18:07:55Z | NONE | Sounds good, I'll do this in the meantime. Still quite interested in |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Preprocess function for save_mfdataset 712189206 | |
702265883 | https://github.com/pydata/xarray/issues/4475#issuecomment-702265883 | https://api.github.com/repos/pydata/xarray/issues/4475 | MDEyOklzc3VlQ29tbWVudDcwMjI2NTg4Mw== | heerad 2560426 | 2020-10-01T16:52:59Z | 2020-10-01T16:52:59Z | NONE | Multiple threads (the default), because it's recommended "for numeric code that releases the GIL (like NumPy, Pandas, Scikit-Learn, Numba, …)" according to the dask docs. I guess I could do multi-threaded for the compute part (everything up to the definition of |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Preprocess function for save_mfdataset 712189206 | |
702181324 | https://github.com/pydata/xarray/issues/4474#issuecomment-702181324 | https://api.github.com/repos/pydata/xarray/issues/4474 | MDEyOklzc3VlQ29tbWVudDcwMjE4MTMyNA== | heerad 2560426 | 2020-10-01T14:39:01Z | 2020-10-01T14:39:01Z | NONE | Great! This will be a common use-case for me, and I imagine others who are doing any sort of time series computation on large datasets. |
{ "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Implement rolling_exp for dask arrays 712052219 | |
702178407 | https://github.com/pydata/xarray/issues/4475#issuecomment-702178407 | https://api.github.com/repos/pydata/xarray/issues/4475 | MDEyOklzc3VlQ29tbWVudDcwMjE3ODQwNw== | heerad 2560426 | 2020-10-01T14:34:28Z | 2020-10-01T14:34:28Z | NONE | Thank you, this works for me. However, it's quite slow and seems to scale faster than linearly as the length of Could it be connected to https://github.com/pydata/xarray/issues/2912#issuecomment-485497398 where they suggest to use Appreciate the help! |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Preprocess function for save_mfdataset 712189206 | |
701676076 | https://github.com/pydata/xarray/issues/4475#issuecomment-701676076 | https://api.github.com/repos/pydata/xarray/issues/4475 | MDEyOklzc3VlQ29tbWVudDcwMTY3NjA3Ng== | heerad 2560426 | 2020-09-30T22:17:24Z | 2020-09-30T22:17:24Z | NONE | Unfortunately that doesn't work:
|
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Preprocess function for save_mfdataset 712189206 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
issue 4