issues
19 rows where comments = 6 and user = 1217238 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date), closed_at (date)
id | node_id | number | title | user | state | locked | assignee | milestone | comments | created_at | updated_at ▲ | closed_at | author_association | active_lock_reason | draft | pull_request | body | reactions | performed_via_github_app | state_reason | repo | type |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
842436143 | MDU6SXNzdWU4NDI0MzYxNDM= | 5081 | Lazy indexing arrays as a stand-alone package | shoyer 1217238 | open | 0 | 6 | 2021-03-27T07:06:03Z | 2023-12-15T13:20:03Z | MEMBER | From @rabernat on Twitter:
The idea here is create a first-class "duck array" library for lazy indexing that could replace xarray's internal classes for lazy indexing. This would be in some ways similar to dask.array, but much simpler, because it doesn't have to worry about parallel computing. Desired features:
A common feature of these operations is they can (and almost always should) be fused with indexing: if N elements are selected via indexing, only O(N) compute and memory is required to produce them, regards of the size of the original arrays as long as the number of applied operations can be treated as a constant. Memory access is significantly slower than compute on modern hardware, so recomputing these operations on the fly is almost always a good idea. Out of scope: lazy computation when indexing could require access to many more elements to compute the desired value than are returned. For example, This is valuable functionality for Xarray for two reasons:
Related issues:
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/5081/reactions", "total_count": 6, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 6, "rocket": 0, "eyes": 0 } |
xarray 13221727 | issue | ||||||||
253395960 | MDU6SXNzdWUyNTMzOTU5NjA= | 1533 | Index variables loaded from dask can be computed twice | shoyer 1217238 | closed | 0 | 6 | 2017-08-28T17:18:27Z | 2023-04-06T04:15:46Z | 2023-04-06T04:15:46Z | MEMBER | as reported by @crusaderky in #1522 |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1533/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
948890466 | MDExOlB1bGxSZXF1ZXN0NjkzNjY1NDEy | 5624 | Make typing-extensions optional | shoyer 1217238 | closed | 0 | 6 | 2021-07-20T17:43:22Z | 2021-07-22T23:30:49Z | 2021-07-22T23:02:03Z | MEMBER | 0 | pydata/xarray/pulls/5624 | Type checking may be a little worse if typing-extensions are not installed, but I don't think it's worth the trouble of adding another hard dependency just for one use for TypeGuard. Note: sadly this doesn't work yet. Mypy (and pylance) don't like the type alias defined with try/except. Any ideas? In the worst case, we could revert the TypeGuard entirely, but that would be a shame...
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/5624/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | pull | |||||
715374721 | MDU6SXNzdWU3MTUzNzQ3MjE= | 4490 | Group together decoding options into a single argument | shoyer 1217238 | open | 0 | 6 | 2020-10-06T06:15:18Z | 2020-10-29T04:07:46Z | MEMBER | Is your feature request related to a problem? Please describe.
Describe the solution you'd like To simple the interface, I propose to group together all the decoding options into a new @dataclass(frozen=True) class DecodingOptions: mask: Optional[bool] = None scale: Optional[bool] = None datetime: Optional[bool] = None timedelta: Optional[bool] = None use_cftime: Optional[bool] = None concat_characters: Optional[bool] = None coords: Optional[bool] = None drop_variables: Optional[List[str]] = None
``` The signature of Question: are Note: the current signature is Usage with the new interface would look like This requires a little bit more typing than what we currently have, but it has a few advantages:
Describe alternatives you've considered For the overall approach:
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/4490/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | issue | ||||||||
702372014 | MDExOlB1bGxSZXF1ZXN0NDg3NjYxMzIz | 4426 | Fix for h5py deepcopy issues | shoyer 1217238 | closed | 0 | 6 | 2020-09-16T01:11:00Z | 2020-09-18T22:31:13Z | 2020-09-18T22:31:09Z | MEMBER | 0 | pydata/xarray/pulls/4426 |
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/4426/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | pull | |||||
417542619 | MDU6SXNzdWU0MTc1NDI2MTk= | 2803 | Test failure with TestValidateAttrs.test_validating_attrs | shoyer 1217238 | closed | 0 | 6 | 2019-03-05T23:03:02Z | 2020-08-25T14:29:19Z | 2019-03-14T15:59:13Z | MEMBER | This is due to setting multi-dimensional attributes being an error, as of the latest netCDF4-Python release: https://github.com/Unidata/netcdf4-python/blob/master/Changelog E.g., as seen on Appveyor: https://ci.appveyor.com/project/shoyer/xray/builds/22834250/job/9q0ip6i3cchlbkw2 ``` ================================== FAILURES =================================== ___ TestValidateAttrs.test_validating_attrs _____ self = <xarray.tests.test_backends.TestValidateAttrs object at 0x00000096BE5FAFD0> def test_validating_attrs(self): def new_dataset(): return Dataset({'data': ('y', np.arange(10.0))}, {'y': np.arange(10)})
xarray\core\dataset.py:1323: in to_netcdf compute=compute) xarray\backends\api.py:767: in to_netcdf unlimited_dims=unlimited_dims) xarray\backends\api.py:810: in dump_to_store unlimited_dims=unlimited_dims) xarray\backends\common.py:262: in store self.set_attributes(attributes) xarray\backends\common.py:278: in set_attributes self.set_attribute(k, v) xarray\backends\netCDF4_.py:418: in set_attribute set_nc_attribute(self.ds, key, value) xarray\backends\netCDF4.py:294: in _set_nc_attribute obj.setncattr(key, value) netCDF4_netCDF4.pyx:2781: in netCDF4._netCDF4.Dataset.setncattr ???
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/2803/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
398107776 | MDU6SXNzdWUzOTgxMDc3NzY= | 2666 | Dataset.from_dataframe will produce a FutureWarning for DatetimeTZ data | shoyer 1217238 | open | 0 | 6 | 2019-01-11T02:45:49Z | 2019-12-30T22:58:23Z | MEMBER | This appears with the development version of pandas; see https://github.com/pandas-dev/pandas/issues/24716 for details. Example: ``` In [16]: df = pd.DataFrame({"A": pd.date_range('2000', periods=12, tz='US/Central')}) In [17]: df.to_xarray() /Users/taugspurger/Envs/pandas-dev/lib/python3.7/site-packages/xarray/core/dataset.py:3111: FutureWarning: Converting timezone-aware DatetimeArray to timezone-naive ndarray with 'datetime64[ns]' dtype. In the future, this will return an ndarray with 'object' dtype where each element is a 'pandas.Timestamp' with the correct 'tz'. To accept the future behavior, pass 'dtype=object'. To keep the old behavior, pass 'dtype="datetime64[ns]"'. data = np.asarray(series).reshape(shape) Out[17]: <xarray.Dataset> Dimensions: (index: 12) Coordinates: * index (index) int64 0 1 2 3 4 5 6 7 8 9 10 11 Data variables: A (index) datetime64[ns] 2000-01-01T06:00:00 ... 2000-01-12T06:00:00 ``` |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/2666/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | issue | ||||||||
269348789 | MDU6SXNzdWUyNjkzNDg3ODk= | 1668 | Remove use of allow_cleanup_failure in test_backends.py | shoyer 1217238 | open | 0 | 6 | 2017-10-28T20:47:31Z | 2019-09-29T20:07:03Z | MEMBER | This exists for the benefit of Windows, on which trying to delete an open file results in an error. But really, it would be nice to have a test suite that doesn't leave any temporary files hanging around. The main culprit is tests like this, where opening a file triggers an error:
The way to fix this is to use mocking of some sort, to intercept calls to backend file objects and close them afterwards. |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1668/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | issue | ||||||||
278713328 | MDU6SXNzdWUyNzg3MTMzMjg= | 1756 | Deprecate inplace methods | shoyer 1217238 | closed | 0 | 0.11 2856429 | 6 | 2017-12-02T20:09:00Z | 2019-03-25T19:19:10Z | 2018-11-03T21:24:13Z | MEMBER | The following methods have an As proposed in https://github.com/pydata/xarray/issues/1755#issuecomment-348682403, let's deprecate all of these at the next major release (v0.11). They add unnecessary complexity to methods and promote confusing about xarray's data model. Practically, we would change all of the default values to |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1756/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | |||||
395332265 | MDExOlB1bGxSZXF1ZXN0MjQxODExMjc4 | 2642 | Use pycodestyle for lint checks. | shoyer 1217238 | closed | 0 | 6 | 2019-01-02T18:11:38Z | 2019-03-14T06:27:20Z | 2019-01-03T18:10:13Z | MEMBER | 0 | pydata/xarray/pulls/2642 | flake8 includes a few more useful checks, but it's annoying to only see it's output in Travis-CI results. This keeps Travis-CI and pep8speaks in sync. |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/2642/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | pull | |||||
369310993 | MDU6SXNzdWUzNjkzMTA5OTM= | 2480 | test_apply_dask_new_output_dimension is broken on master with dask-dev | shoyer 1217238 | closed | 0 | 6 | 2018-10-11T21:24:33Z | 2018-10-12T16:26:17Z | 2018-10-12T16:26:17Z | MEMBER | Example build failure: https://travis-ci.org/pydata/xarray/jobs/439949937 ``` =================================== FAILURES =================================== ___ test_apply_dask_new_output_dimension ___ @requires_dask def test_apply_dask_new_output_dimension(): import dask.array as da
xarray/tests/test_computation.py:24: in assert_identical assert a.identical(b), msg xarray/core/dataarray.py:1923: in identical self._all_compat(other, 'identical')) xarray/core/dataarray.py:1875: in _all_compat compat(self, other)) xarray/core/dataarray.py:1872: in compat return getattr(x.variable, compat_str)(y.variable) xarray/core/variable.py:1461: in identical self.equals(other)) xarray/core/variable.py:1439: in equals equiv(self.data, other.data))) xarray/core/duck_array_ops.py:144: in array_equiv arr1, arr2 = as_like_arrays(arr1, arr2) xarray/core/duck_array_ops.py:128: in as_like_arrays return tuple(np.asarray(d) for d in data) xarray/core/duck_array_ops.py:128: in <genexpr> return tuple(np.asarray(d) for d in data) ../../../miniconda/envs/test_env/lib/python3.6/site-packages/numpy/core/numeric.py:501: in asarray return array(a, dtype, copy=False, order=order) ../../../miniconda/envs/test_env/lib/python3.6/site-packages/dask/array/core.py:1118: in array x = self.compute() ../../../miniconda/envs/test_env/lib/python3.6/site-packages/dask/base.py:156: in compute (result,) = compute(self, traverse=False, kwargs) ../../../miniconda/envs/test_env/lib/python3.6/site-packages/dask/base.py:390: in compute dsk = collections_to_dsk(collections, optimize_graph, kwargs) ../../../miniconda/envs/test_env/lib/python3.6/site-packages/dask/base.py:194: in collections_to_dsk for opt, (dsk, keys) in groups.items()])) ../../../miniconda/envs/test_env/lib/python3.6/site-packages/dask/base.py:194: in <listcomp> for opt, (dsk, keys) in groups.items()])) ../../../miniconda/envs/test_env/lib/python3.6/site-packages/dask/array/optimization.py:41: in optimize dsk = ensure_dict(dsk) ../../../miniconda/envs/test_env/lib/python3.6/site-packages/dask/utils.py:830: in ensure_dict result.update(dd) ../../../miniconda/envs/test_env/lib/python3.6/_collections_abc.py:720: in iter yield from self._mapping ../../../miniconda/envs/test_env/lib/python3.6/site-packages/dask/array/top.py:168: in iter return iter(self._dict) ../../../miniconda/envs/test_env/lib/python3.6/site-packages/dask/array/top.py:160: in _dict concatenate=self.concatenate ../../../miniconda/envs/test_env/lib/python3.6/site-packages/dask/array/top.py:305: in top keytups = list(itertools.product(*[range(dims[i]) for i in out_indices])) .0 = <tuple_iterator object at 0x7f606ba84fd0>
My guess is that this is somehow related to @mrocklin's recent refactor of dask.array.atop: https://github.com/dask/dask/pull/3998 If the cause isn't obvious, I'll try to come up with a simple dask only example that reproduces it. |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/2480/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
361915770 | MDU6SXNzdWUzNjE5MTU3NzA= | 2424 | 0.10.9 release | shoyer 1217238 | closed | 0 | 6 | 2018-09-19T20:31:29Z | 2018-09-26T01:05:09Z | 2018-09-22T15:14:48Z | MEMBER | It's now been two months since the 0.10.8 release, so we really ought to issue a new minor release. I was initially thinking of skipping straight to 0.11.0 if we include https://github.com/pydata/xarray/pull/2261 (xarray.backends refactor), but it seems that will take a bit longer to review/test so it's probably worth issuing a 0.10.9 release first. @pydata/xarray -- are there any PRs / bug-fixes in particular we should wait for before issuing the release? I suppose it would be good to sort out https://github.com/pydata/xarray/issues/2422 (Plot2D no longer sorts coordinates before plotting) |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/2424/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
302153432 | MDExOlB1bGxSZXF1ZXN0MTcyNzYxNTAw | 1962 | Support __array_ufunc__ for xarray objects. | shoyer 1217238 | closed | 0 | 6 | 2018-03-05T02:36:20Z | 2018-03-12T20:31:07Z | 2018-03-12T20:31:07Z | MEMBER | 0 | pydata/xarray/pulls/1962 | This means NumPy ufuncs are now supported directly on xarray.Dataset objects, and opens the door to supporting computation on new data types, such as sparse arrays or arrays with units.
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1962/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | pull | |||||
195579837 | MDU6SXNzdWUxOTU1Nzk4Mzc= | 1164 | Don't warn when doing comparisons or arithmetic with NaN | shoyer 1217238 | closed | 0 | 6 | 2016-12-14T16:33:05Z | 2018-02-27T19:35:25Z | 2018-02-27T16:03:43Z | MEMBER | Pandas used to unilaterally disable NumPy's warnings for doing comparisons with NaN, but now it doesn't: https://github.com/pandas-dev/pandas/issues/13109 |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1164/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
171828347 | MDU6SXNzdWUxNzE4MjgzNDc= | 974 | Indexing with alignment and broadcasting | shoyer 1217238 | closed | 0 | 1.0 741199 | 6 | 2016-08-18T06:39:27Z | 2018-02-04T23:30:12Z | 2018-02-04T23:30:11Z | MEMBER | I think we can bring all of NumPy's advanced indexing to xarray in a very consistent way, with only very minor breaks in backwards compatibility. For boolean indexing:
- For vectorized indexing (by integer or index value):
- All of these methods should also work for indexing on This framework neatly subsumes most of the major limitations with xarray's existing indexing:
- Boolean indexing on multi-dimensional arrays works in an intuitive way, for both selection and assignment.
- No more need for specialized methods ( Questions to consider:
- How does this interact with @benbovy's enhancements for MultiIndex indexing? (#802 and #947)
- How do we handle mixed slice and array indexing? In NumPy, this is a major source of confusion, because slicing is done before broadcasting and the order of slices in the result is handled separately from broadcast indices. I think we may be able to resolve this by mapping slices in this case to 1D arrays along their respective axes, and using our normal broadcasting rules.
- Should we deprecate non-boolean indexing with xref these comments from @MaximilianR and myself Note: I would certainly welcome help making this happen from a contributor other than myself, though you should probably wait until I finish #964, first, which lays important groundwork. |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/974/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | |||||
274308380 | MDU6SXNzdWUyNzQzMDgzODA= | 1720 | Possible regression with PyNIO data not being lazily loaded | shoyer 1217238 | closed | 0 | 6 | 2017-11-15T21:20:41Z | 2017-11-17T17:33:13Z | 2017-11-17T16:44:40Z | MEMBER | @weathergod reports on the mailing list:
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1720/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
207587161 | MDU6SXNzdWUyMDc1ODcxNjE= | 1269 | GroupBy like API for resample | shoyer 1217238 | closed | 0 | 6 | 2017-02-14T17:46:02Z | 2017-09-22T16:27:35Z | 2017-09-22T16:27:35Z | MEMBER | Since we wrote It would be nice to redo the xarray resample API to match, e.g., |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1269/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
88075523 | MDU6SXNzdWU4ODA3NTUyMw== | 432 | Tools for converting between xray.Dataset and nested dictionaries/JSON | shoyer 1217238 | closed | 0 | 6 | 2015-06-13T22:25:28Z | 2016-08-11T21:54:51Z | 2016-08-11T21:54:51Z | MEMBER | This came up in discussion with @freeman-lab -- xray does not have direct support for converting datasets to or from nested dictionaries (i.e., as could be serialized in JSON). This is quite straightforward to implement oneself, of course, but there's something to be said for making this more obvious. I'm thinking of a serialization format that looks something like this:
The solution here would be to either:
1. add a few examples to the IO documentation of how to roll this one-self, or
2. create a few helper methods/functions to make this even easier: |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/432/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
104768781 | MDExOlB1bGxSZXF1ZXN0NDQxNDkyOTQ= | 559 | Fix pcolormesh plots with cartopy | shoyer 1217238 | closed | 0 | 0.6.1 1307323 | 6 | 2015-09-03T19:50:22Z | 2015-11-15T21:49:11Z | 2015-09-14T20:33:36Z | MEMBER | 0 | pydata/xarray/pulls/559 |
Before:
After:
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/559/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | pull |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issues] ( [id] INTEGER PRIMARY KEY, [node_id] TEXT, [number] INTEGER, [title] TEXT, [user] INTEGER REFERENCES [users]([id]), [state] TEXT, [locked] INTEGER, [assignee] INTEGER REFERENCES [users]([id]), [milestone] INTEGER REFERENCES [milestones]([id]), [comments] INTEGER, [created_at] TEXT, [updated_at] TEXT, [closed_at] TEXT, [author_association] TEXT, [active_lock_reason] TEXT, [draft] INTEGER, [pull_request] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [state_reason] TEXT, [repo] INTEGER REFERENCES [repos]([id]), [type] TEXT ); CREATE INDEX [idx_issues_repo] ON [issues] ([repo]); CREATE INDEX [idx_issues_milestone] ON [issues] ([milestone]); CREATE INDEX [idx_issues_assignee] ON [issues] ([assignee]); CREATE INDEX [idx_issues_user] ON [issues] ([user]);