issue_comments
4 rows where issue = 402413097 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: reactions, created_at (date), updated_at (date)
issue 1
- bfill behavior dask arrays with small chunk size · 4 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
611687777 | https://github.com/pydata/xarray/issues/2699#issuecomment-611687777 | https://api.github.com/repos/pydata/xarray/issues/2699 | MDEyOklzc3VlQ29tbWVudDYxMTY4Nzc3Nw== | kefirbandi 1277781 | 2020-04-09T18:36:36Z | 2020-04-09T18:36:36Z | CONTRIBUTOR | I encountered this bug a few days ago. I understand it isn't trivial to fix, but would it be possible to check and throw an exception? Still better than having it go unnoticed. Thanks |
{ "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
bfill behavior dask arrays with small chunk size 402413097 | |
537018647 | https://github.com/pydata/xarray/issues/2699#issuecomment-537018647 | https://api.github.com/repos/pydata/xarray/issues/2699 | MDEyOklzc3VlQ29tbWVudDUzNzAxODY0Nw== | gmaze 1956032 | 2019-10-01T12:43:46Z | 2019-10-01T12:43:46Z | CONTRIBUTOR | I also recently encountered this bug and without user warnings it took me a while to identify its origin. I'll use this temporary fix. Thanks |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
bfill behavior dask arrays with small chunk size 402413097 | |
456999707 | https://github.com/pydata/xarray/issues/2699#issuecomment-456999707 | https://api.github.com/repos/pydata/xarray/issues/2699 | MDEyOklzc3VlQ29tbWVudDQ1Njk5OTcwNw== | shoyer 1217238 | 2019-01-23T22:58:50Z | 2019-01-23T23:13:59Z | MEMBER | I think this will work (though it needs more tests): ```python import bottleneck import dask.array as da import numpy as np def _last_element(array, axis): slices = [slice(None)] * array.ndim slices[axis] = slice(-1, None) return array[tuple(slices)] def _concat_push_slice(last_elements, array, axis): concatenated = np.concatenate([last_elements, array], axis=axis) pushed = bottleneck.push(concatenated, axis=axis) slices = [slice(None)] * array.ndim slices[axis] = slice(1, None) sliced = pushed[tuple(slices)] return sliced def push(array, axis): if axis < 0: axis += array.ndim pushed = array.map_blocks(bottleneck.push, dtype=array.dtype, axis=axis) new_chunks = list(array.chunks) new_chunks[axis] = tuple(1 for _ in array.chunks[axis]) last_elements = pushed.map_blocks( _last_element, dtype=array.dtype, chunks=tuple(new_chunks), axis=axis) pushed_last_elements = ( last_elements.rechunk({axis: -1}) .map_blocks(bottleneck.push, dtype=array.dtype, axis=axis) .rechunk({axis: 1}) ) nan_shape = tuple(1 if axis == a else s for a, s in enumerate(array.shape)) nan_chunks = tuple((1,) if axis == a else c for a, c in enumerate(array.chunks)) shifted_pushed_last_elements = da.concatenate( [da.full(np.nan, shape=nan_shape, chunks=nan_chunks), pushed_last_elements[(slice(None),) * axis + (slice(None, -1),)]], axis=axis) return da.map_blocks( _concat_push_slice, shifted_pushed_last_elements, pushed, dtype=array.dtype, chunks=array.chunks, axis=axis, ) testsarray = np.array([np.nan, np.nan, np.nan, 1, 2, 3, np.nan, np.nan, 4, 5, np.nan, 6]) expected = bottleneck.push(array, axis=0) for c in range(1, 11): actual = push(da.from_array(array, chunks=c), axis=0).compute() np.testing.assert_equal(actual, expected) ``` |
{ "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
bfill behavior dask arrays with small chunk size 402413097 | |
456988170 | https://github.com/pydata/xarray/issues/2699#issuecomment-456988170 | https://api.github.com/repos/pydata/xarray/issues/2699 | MDEyOklzc3VlQ29tbWVudDQ1Njk4ODE3MA== | shoyer 1217238 | 2019-01-23T22:17:21Z | 2019-01-23T22:17:38Z | MEMBER | Thanks for the clear report. Indeed, this looks like a bug.
Instead, I think we need a multi-step process for parallelizing |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
bfill behavior dask arrays with small chunk size 402413097 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 3