issues
22 rows where comments = 3, repo = 13221727 and user = 35968931 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: draft, created_at (date), updated_at (date), closed_at (date)
| id | node_id | number | title | user | state | locked | assignee | milestone | comments | created_at | updated_at ▲ | closed_at | author_association | active_lock_reason | draft | pull_request | body | reactions | performed_via_github_app | state_reason | repo | type |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 2120340151 | PR_kwDOAMm_X85mHqI0 | 8714 | Avoid coercing to numpy in `as_shared_dtypes` | TomNicholas 35968931 | open | 0 | 3 | 2024-02-06T09:35:22Z | 2024-03-28T18:31:50Z | MEMBER | 0 | pydata/xarray/pulls/8714 |
|
{
"url": "https://api.github.com/repos/pydata/xarray/issues/8714/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | pull | ||||||
| 2098882374 | I_kwDOAMm_X859GmdG | 8660 | dtype encoding ignored during IO? | TomNicholas 35968931 | closed | 0 | 3 | 2024-01-24T18:50:47Z | 2024-02-05T17:35:03Z | 2024-02-05T17:35:02Z | MEMBER | What happened?When I set the What did you expect to happen?I expected that setting Minimal Complete Verifiable Example```Python air = xr.tutorial.open_dataset('air_temperature') air['air'].dtype # returns dtype('float32') air['air'].encoding['dtype'] # returns dtype('int16'), which already seems weird air.to_zarr('air.zarr') # I would assume here that the encoding actually does something during IO now if I check the zarr
|
{
"url": "https://api.github.com/repos/pydata/xarray/issues/8660/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
completed | xarray 13221727 | issue | ||||||
| 2116695961 | I_kwDOAMm_X85-KjeZ | 8699 | Wrapping a `kerchunk.Array` object directly with xarray | TomNicholas 35968931 | open | 0 | 3 | 2024-02-03T22:15:07Z | 2024-02-04T21:15:14Z | MEMBER | What is your issue?In https://github.com/fsspec/kerchunk/issues/377 the idea came up of using the xarray API to concatenate arrays which represent parts of a zarr store - i.e. using xarray to kerchunk a large set of netCDF files instead of using The idea is to make something like this work for kerchunking sets of netCDF files into zarr stores ```python ds = xr.open_mfdataset( '/my/files*.nc' engine='kerchunk', # kerchunk registers an xarray IO backend that returns zarr.Array objects combine='nested', # 'by_coords' would require actually reading coordinate data parallel=True, # would use dask.delayed to generate reference dicts for each file in parallel ) ds # now wraps a bunch of zarr.Array / kerchunk.Array objects, no need for dask arrays ds.kerchunk.to_zarr(store='out.zarr') # kerchunk defines an xarray accessor that extracts the zarr arrays and serializes them (which could also be done in parallel if writing to parquet) ``` I had a go at doing this in this notebook, and in doing so discovered a few potential issues with xarray's internals. For this to work xarray has to:
- Wrap a It's an interesting exercise in using xarray as an abstraction, with no access to real numerical values at all. |
{
"url": "https://api.github.com/repos/pydata/xarray/issues/8699/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} |
xarray 13221727 | issue | ||||||||
| 2027528985 | PR_kwDOAMm_X85hQBHP | 8525 | Remove PR labeler bot | TomNicholas 35968931 | closed | 0 | 3 | 2023-12-06T02:31:56Z | 2023-12-06T02:45:46Z | 2023-12-06T02:45:41Z | MEMBER | 0 | pydata/xarray/pulls/8525 | RIP
|
{
"url": "https://api.github.com/repos/pydata/xarray/issues/8525/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | pull | |||||
| 1806973709 | PR_kwDOAMm_X85VoNVM | 7992 | Docs page on interoperability | TomNicholas 35968931 | closed | 0 | 3 | 2023-07-17T05:02:29Z | 2023-10-26T16:08:56Z | 2023-10-26T16:04:33Z | MEMBER | 0 | pydata/xarray/pulls/7992 | Builds upon #7991 by adding a page to the internals enumerating all the different ways in which xarray is interoperable. Would be nice if https://github.com/pydata/xarray/pull/6975 were merged so that I could link to it from this new page.
|
{
"url": "https://api.github.com/repos/pydata/xarray/issues/7992/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | pull | |||||
| 1812811751 | I_kwDOAMm_X85sDU_n | 8008 | "Deep linking" disparate documentation resources together | TomNicholas 35968931 | open | 0 | 3 | 2023-07-19T22:18:55Z | 2023-10-12T18:36:52Z | MEMBER | What is your issue?Our docs have a general issue with having lots of related resources that are not necessarily linked together in a useful way. This results in users (including myself!) getting "stuck" in one part of the docs and being unaware of material that would help them solve their specific issue. To give a concrete example, if a user wants to know about
Different types of material are great, but only some of these resources are linked to others. The biggest missed opportunity here is the way all the great content on the tutorial.xarray.dev repository is not linked from anywhere on the main documentation site (I believe). To address that we could either (a) integrate the Identifying sections that could be linked and adding links would be a great task for new contributors. |
{
"url": "https://api.github.com/repos/pydata/xarray/issues/8008/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | issue | ||||||||
| 663235664 | MDU6SXNzdWU2NjMyMzU2NjQ= | 4243 | Manually drop DataArray from memory? | TomNicholas 35968931 | closed | 0 | 3 | 2020-07-21T18:54:40Z | 2023-09-12T16:17:12Z | 2023-09-12T16:17:12Z | MEMBER | Is it possible to deliberately drop data associated with a particular DataArray from memory? Obviously Also does calling python's built-in garbage collector (i.e. The context of this question is that I'm trying to resave some massive variables (~65GB each) that were loaded from thousands of files into just a few files for each variable. I would love to use @rabernat 's new rechunker package but I'm not sure how easily I can convert my current netCDF data to Zarr, and I'm interested in this question no matter how I end up solving the problem. I don't currently have a particularly good understanding of file I/O and memory management in xarray, but would like to improve it. Can anyone recommend a tool I can use to answer this kind of question myself on my own machine? I suppose it would need to be able to tell me the current memory usage of specific objects, not just the total memory usage. (@johnomotani you might be interested) |
{
"url": "https://api.github.com/repos/pydata/xarray/issues/4243/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
completed | xarray 13221727 | issue | ||||||
| 1807782455 | I_kwDOAMm_X85rwJI3 | 7996 | Stable docs build not showing latest changes after release | TomNicholas 35968931 | closed | 0 | 3 | 2023-07-17T13:24:58Z | 2023-07-17T20:48:19Z | 2023-07-17T20:48:19Z | MEMBER | What happened?I released xarray version v2023.07.0 last night, but I'm not seeing changes to the documentation reflected in the What did you expect to happen?No response Minimal Complete Verifiable ExampleNo response MVCE confirmation
Relevant log outputNo response Anything else we need to know?No response Environment |
{
"url": "https://api.github.com/repos/pydata/xarray/issues/7996/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
completed | xarray 13221727 | issue | ||||||
| 1779880070 | PR_kwDOAMm_X85UMTE7 | 7951 | Chunked array docs | TomNicholas 35968931 | closed | 0 | 3 | 2023-06-28T23:01:42Z | 2023-07-05T20:33:33Z | 2023-07-05T20:08:19Z | MEMBER | 0 | pydata/xarray/pulls/7951 | Builds upon #7911
|
{
"url": "https://api.github.com/repos/pydata/xarray/issues/7951/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | pull | |||||
| 1694956396 | I_kwDOAMm_X85lBvts | 7813 | Task naming for general chunkmanagers | TomNicholas 35968931 | open | 0 | 3 | 2023-05-03T22:56:46Z | 2023-05-05T10:30:39Z | MEMBER | What is your issue?(Follow-up to #7019) When you create a dask graph of xarray operations, the tasks in the graph get useful names according the name of the DataArray they operate on, or whether they represent an Currently for cubed this doesn't work, for example this graph from https://github.com/pangeo-data/distributed-array-examples/issues/2#issuecomment-1533852877:
cc @tomwhite @dcherian |
{
"url": "https://api.github.com/repos/pydata/xarray/issues/7813/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | issue | ||||||||
| 1512290017 | I_kwDOAMm_X85aI7bh | 7403 | Zarr error when trying to overwrite part of existing store | TomNicholas 35968931 | open | 0 | 3 | 2022-12-28T00:40:16Z | 2023-01-11T21:26:10Z | MEMBER | What happened?
What did you expect to happen?With mode I expected that because that's what the docstring of
The default mode is "w", so I was expecting it to overwrite. Minimal Complete Verifiable Example```Python import xarray as xr import numpy as np np.random.seed(0) ds = xr.Dataset() ds["data"] = (['x', 'y'], np.random.random((100,100))) ds.to_zarr("test.zarr") print(ds["data"].mean().compute()) returns array(0.49645889) as expectedds = xr.open_dataset("test.zarr", engine='zarr', chunks={}) ds["data"].mean().compute() print(ds["data"].mean().compute()) still returns array(0.49645889) as expectedds.to_zarr("test.zarr", mode="a") ```
MVCE confirmation
Relevant log outputNo response Anything else we need to know?I would like to know what the intended result is supposed to be here, so that I can make sure datatree behaves the same way, see https://github.com/xarray-contrib/datatree/issues/168. EnvironmentMain branch of xarray, zarr v2.13.3 |
{
"url": "https://api.github.com/repos/pydata/xarray/issues/7403/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | issue | ||||||||
| 906023492 | MDExOlB1bGxSZXF1ZXN0NjU3MDYxODI5 | 5400 | Multidimensional histogram | TomNicholas 35968931 | open | 0 | 3 | 2021-05-28T20:38:53Z | 2022-11-21T22:41:01Z | MEMBER | 0 | pydata/xarray/pulls/5400 | Initial work on integrating the multi-dimensional dask-powered histogram functionality from xhistogram into xarray. Just working on the skeleton to fit around the histogram algorithm for now, to be filled in later.
Question: |
{
"url": "https://api.github.com/repos/pydata/xarray/issues/5400/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | pull | ||||||
| 1417378270 | PR_kwDOAMm_X85BPGqR | 7192 | Example using Coarsen.construct to split map into regions | TomNicholas 35968931 | closed | 0 | 3 | 2022-10-20T22:14:31Z | 2022-10-21T18:14:59Z | 2022-10-21T18:14:56Z | MEMBER | 0 | pydata/xarray/pulls/7192 | I realised there is very little documentation on Unsure whether it should instead live in the page on reshaping and reorganising data though, as it is essentially a reshape operation. EDIT: Now on the reshape page
cc @jbusecke @paigem |
{
"url": "https://api.github.com/repos/pydata/xarray/issues/7192/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 1,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | pull | |||||
| 1370416843 | PR_kwDOAMm_X84-z6DG | 7023 | Remove dask_array_type checks | TomNicholas 35968931 | closed | 0 | 3 | 2022-09-12T19:31:04Z | 2022-09-13T00:35:25Z | 2022-09-13T00:35:22Z | MEMBER | 0 | pydata/xarray/pulls/7023 |
|
{
"url": "https://api.github.com/repos/pydata/xarray/issues/7023/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | pull | |||||
| 400289716 | MDU6SXNzdWU0MDAyODk3MTY= | 2686 | Is `create_test_data()` public API? | TomNicholas 35968931 | open | 0 | 3 | 2019-01-17T14:00:20Z | 2022-04-09T01:48:14Z | MEMBER | We want to encourage people to use and extend xarray, and we already provide testing functions as public API to help with this. One function I keep using when writing code which uses xarray is Is there any reason why it shouldn't be public API? Is there something I should use instead? |
{
"url": "https://api.github.com/repos/pydata/xarray/issues/2686/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | issue | ||||||||
| 1012428149 | PR_kwDOAMm_X84shL9H | 5834 | Combine by coords dataarray bugfix | TomNicholas 35968931 | closed | 0 | 3 | 2021-09-30T17:17:00Z | 2021-10-29T19:57:36Z | 2021-10-29T19:57:36Z | MEMBER | 0 | pydata/xarray/pulls/5834 | Also reorganised the logic that deals with combining mixed sets of objects (i.e. named dataarrays, unnamed dataarrays, datasets) that was added in #4696. TODO - same reorganisation / testing but for
|
{
"url": "https://api.github.com/repos/pydata/xarray/issues/5834/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | pull | |||||
| 1020555552 | PR_kwDOAMm_X84s6zAH | 5846 | Change return type of DataArray.chunks and Dataset.chunks to a dict | TomNicholas 35968931 | closed | 0 | 3 | 2021-10-08T00:02:20Z | 2021-10-26T15:52:00Z | 2021-10-26T15:51:59Z | MEMBER | 1 | pydata/xarray/pulls/5846 | Rectifies the the issue in #5843 by making Currently a WIP - I changed the behaviour but this obviously broke quite a few tests and I haven't looked at them yet.
|
{
"url": "https://api.github.com/repos/pydata/xarray/issues/5846/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | pull | |||||
| 877944829 | MDExOlB1bGxSZXF1ZXN0NjMxODI1Nzky | 5274 | Update release guide | TomNicholas 35968931 | closed | 0 | 3 | 2021-05-06T19:50:53Z | 2021-05-13T17:44:47Z | 2021-05-13T17:44:47Z | MEMBER | 0 | pydata/xarray/pulls/5274 | Updated the release guide to account for what is now automated via github actions, and any other bits I felt could be clearer. Now only 16 easy steps!
|
{
"url": "https://api.github.com/repos/pydata/xarray/issues/5274/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 1,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | pull | |||||
| 474247717 | MDU6SXNzdWU0NzQyNDc3MTc= | 3168 | apply_ufunc erroneously operating on an empty array when dask used | TomNicholas 35968931 | closed | 0 | 3 | 2019-07-29T20:44:23Z | 2020-03-30T15:08:16Z | 2020-03-30T15:08:15Z | MEMBER | Problem description
Minimum working example```python import numpy as np import xarray as xr def example_ufunc(x): print(x.shape) return np.mean(x, axis=-1) def new_mean(da, dim): result = xr.apply_ufunc(example_ufunc, da, input_core_dims=[[dim]], dask='parallelized', output_dtypes=[da.dtype]) return result shape = {'t': 2, 'x':3} data = xr.DataArray(data=np.random.rand(*shape.values()), dims=shape.keys()) unchunked = data chunked = data.chunk(shape) actual = new_mean(chunked, dim='x') # raises the warning print(actual) print(actual.compute()) # does the computation correctly ``` Result
Expected resultSame thing without the Output of
|
{
"url": "https://api.github.com/repos/pydata/xarray/issues/3168/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
completed | xarray 13221727 | issue | ||||||
| 497184021 | MDU6SXNzdWU0OTcxODQwMjE= | 3334 | plot.line fails when plot axis is a 1D coordinate | TomNicholas 35968931 | closed | 0 | 3 | 2019-09-23T15:52:48Z | 2019-09-26T08:51:59Z | 2019-09-26T08:51:59Z | MEMBER | MCVE Code Sample```python import xarray as xr import numpy as np x_coord = xr.DataArray(data=[0.1, 0.2], dims=['x']) t_coord = xr.DataArray(data=[10, 20], dims=['t']) da = xr.DataArray(data=np.array([[0, 1], [5, 9]]), dims=['x', 't'], coords={'x': x_coord, 'time': t_coord}) print(da) da.transpose('time', 'x')
Traceback (most recent call last): File "mwe.py", line 22, in <module> da.transpose('time', 'x') File "/home/tegn500/Documents/Work/Code/xarray/xarray/core/dataarray.py", line 1877, in transpose "permuted array dimensions (%s)" % (dims, tuple(self.dims)) ValueError: arguments to transpose (('time', 'x')) must be permuted array dimensions (('x', 't')) ``` As This causes bug in other parts of the code - for example I found this by trying to plot this type of dataarray:
(You can get a similar error also with If the code which explicitly checks that the arguments to transpose are dims and not just coordinate dimensions is removed, then both of these examples work as expected. I would like to generalise the transpose function to also accept dimension coordinates, is there any reason not to do this? |
{
"url": "https://api.github.com/repos/pydata/xarray/issues/3334/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
completed | xarray 13221727 | issue | ||||||
| 399389293 | MDExOlB1bGxSZXF1ZXN0MjQ0ODI4NTY3 | 2678 | Hotfix for #2662 | TomNicholas 35968931 | closed | 0 | 3 | 2019-01-15T15:11:48Z | 2019-02-02T23:50:40Z | 2019-01-17T13:05:43Z | MEMBER | 0 | pydata/xarray/pulls/2678 |
Explained in #2662. Also renamed some variables slightly for clarity. Not sure how to add a test without refactoring the groupby into a separate function, as in it's current form the problem only manifests as a (huge) slowdown. |
{
"url": "https://api.github.com/repos/pydata/xarray/issues/2678/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
xarray 13221727 | pull | |||||
| 404383025 | MDU6SXNzdWU0MDQzODMwMjU= | 2725 | Line plot with x=coord putting wrong variables on axes | TomNicholas 35968931 | closed | 0 | 3 | 2019-01-29T16:43:18Z | 2019-01-30T02:02:22Z | 2019-01-30T02:02:22Z | MEMBER | When I try to plot the values in a 1D DataArray against the values in one of its coordinates, it does not behave at all as expected: ```python import numpy as np import matplotlib.pyplot as plt from xarray import DataArray current = DataArray(name='current', data=np.array([5, 8, 14, 22, 30]), dims=['time'], coords={'time': (['time'], np.array([0.1, 0.2, 0.3, 0.4, 0.5])), 'voltage': (['time'], np.array([100, 200, 300, 400, 500]))}) print(current) Try to plot current against voltagecurrent.plot.line(x='voltage') plt.show() ``` Output:
Problem descriptionNot only is Expected OutputBased on the documentation (and common sense) I would have expected it to plot voltage on the x axis and current on the y axis. (using a branch of xarray which is up-to-date with master) |
{
"url": "https://api.github.com/repos/pydata/xarray/issues/2725/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
completed | xarray 13221727 | issue |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issues] (
[id] INTEGER PRIMARY KEY,
[node_id] TEXT,
[number] INTEGER,
[title] TEXT,
[user] INTEGER REFERENCES [users]([id]),
[state] TEXT,
[locked] INTEGER,
[assignee] INTEGER REFERENCES [users]([id]),
[milestone] INTEGER REFERENCES [milestones]([id]),
[comments] INTEGER,
[created_at] TEXT,
[updated_at] TEXT,
[closed_at] TEXT,
[author_association] TEXT,
[active_lock_reason] TEXT,
[draft] INTEGER,
[pull_request] TEXT,
[body] TEXT,
[reactions] TEXT,
[performed_via_github_app] TEXT,
[state_reason] TEXT,
[repo] INTEGER REFERENCES [repos]([id]),
[type] TEXT
);
CREATE INDEX [idx_issues_repo]
ON [issues] ([repo]);
CREATE INDEX [idx_issues_milestone]
ON [issues] ([milestone]);
CREATE INDEX [idx_issues_assignee]
ON [issues] ([assignee]);
CREATE INDEX [idx_issues_user]
ON [issues] ([user]);

