issues
21 rows where state = "closed" and user = 306380 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: comments, closed_at, created_at (date), updated_at (date), closed_at (date)
id | node_id | number | title | user | state | locked | assignee | milestone | comments | created_at | updated_at ▲ | closed_at | author_association | active_lock_reason | draft | pull_request | body | reactions | performed_via_github_app | state_reason | repo | type |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
295270362 | MDU6SXNzdWUyOTUyNzAzNjI= | 1895 | Avoid Adapters in task graphs? | mrocklin 306380 | closed | 0 | 13 | 2018-02-07T19:52:02Z | 2022-05-11T20:26:42Z | 2022-05-11T20:26:42Z | MEMBER | Looking at an ```python
This object has many dependents, and so will presumably have to float around the network to all of the workers ```python
In principle this is fine, especially if this object is cheap to serialize, move, and deserialize. It does introduce a bit of friction though. I'm curious how hard it would be to build task graphs that generated these objects on the fly, or else removed them altogether. It is slightly more convenient from a task scheduling perspective for data access tasks to not have any dependencies. |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1895/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
336371511 | MDExOlB1bGxSZXF1ZXN0MTk3ODQwODgw | 2255 | Add automatic chunking to open_rasterio | mrocklin 306380 | closed | 0 | 10 | 2018-06-27T20:15:07Z | 2022-04-07T20:21:24Z | 2022-04-07T20:21:24Z | MEMBER | 0 | pydata/xarray/pulls/2255 | This uses the automatic chunking in dask 0.18+ to chunk rasterio datasets in a nicely aligned way. Currently this doesn't implement tests due to a difficulty in creating chunked tiff images. This also uncovered some inefficiencies in how Dask doesn't align rechunking to existing chunk schemes.
I could use help on how the following:
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/2255/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | pull | |||||
318950038 | MDU6SXNzdWUzMTg5NTAwMzg= | 2093 | Default chunking in GeoTIFF images | mrocklin 306380 | closed | 0 | 10 | 2018-04-30T16:21:30Z | 2020-06-18T06:27:07Z | 2020-06-18T06:27:07Z | MEMBER | Given a tiled GeoTIFF image I'm looking for the best practice in reading it as a chunked dataset. I did this in this notebook by first opening the file with rasterio, looking at the block sizes, and then using those to inform the argument to In dask.array every time this has come up we've always shot it down, automatic chunking is error prone and hard to do well. However in these cases the object we're being given usually also conveys its chunking in a way that matches how dask.array thinks about it, so the extra cognitive load on the user has been somewhat low. Rasterio's model and API feel much more foreign to me though than a project like NetCDF or H5Py. I find myself wanting a Thoughts on this? Is this in-scope? If so then what is the right API and what is the right policy for how to make xarray/dask.array chunks larger than GeoTIFF chunks? |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/2093/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
400948664 | MDU6SXNzdWU0MDA5NDg2NjQ= | 2692 | Xarray tutorial at SciPy 2019? | mrocklin 306380 | closed | 0 | 10 | 2019-01-19T01:56:38Z | 2020-03-25T04:34:27Z | 2019-02-17T05:07:45Z | MEMBER | Is anyone interested in submitting a tutorial to SciPy 2019? I think that it would be useful to have an official Xarray tutorial out there somewhere on the internet. This could be good motivation to create one. |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/2692/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
287969295 | MDU6SXNzdWUyODc5NjkyOTU= | 1822 | Use apply_ufunc in xESMF regridding package | mrocklin 306380 | closed | 0 | 4 | 2018-01-12T00:17:04Z | 2020-01-15T00:01:49Z | 2020-01-15T00:01:49Z | MEMBER | I would like to call attention to https://github.com/JiaweiZhuang/xESMF/issues/3#issuecomment-354668897 . It seems like the xESMF package does regridding in a way that at least some XArray users find sensible. It should probably make use of, but does not currently use apply_ufunc, and is not particularly parallelizable (or at least that is my understanding). It could be that some modest development by someone more familiar with XArray could have a large impact by properly using apply_ufunc within that codebase. I apologize for posting an issue about another package in this issue tracker. Feel free to close. cc @JiaweiZhuang |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1822/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
221858543 | MDU6SXNzdWUyMjE4NTg1NDM= | 1375 | Sparse arrays | mrocklin 306380 | closed | 0 | 25 | 2017-04-14T18:00:14Z | 2019-08-30T02:36:12Z | 2019-08-13T03:31:14Z | MEMBER | I would like to have an XArray that has scipy.sparse arrays rather than numpy arrays. Is this in scope? What would need to happen within XArray to support this? |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1375/reactions", "total_count": 8, "+1": 8, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
323785231 | MDU6SXNzdWUzMjM3ODUyMzE= | 2143 | Upstream changes in Dask | mrocklin 306380 | closed | 0 | 1 | 2018-05-16T21:01:21Z | 2019-08-15T15:16:54Z | 2019-08-15T15:16:54Z | MEMBER | Hi All, There are a couple changes coming in dask that might affect XArray code:
Both of the old systems will still work, at least for a version or two, but we plan to remove them in the future. I thought I'd bring these changes up here so that we can plan a clean deprecation within XArray. These are also both not yet released, so both features are still up for discussion if this community has additional constraints. |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/2143/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
355308699 | MDU6SXNzdWUzNTUzMDg2OTk= | 2390 | Why are there two compute calls for plot? | mrocklin 306380 | closed | 0 | 3 | 2018-08-29T19:53:45Z | 2019-08-04T23:00:59Z | 2019-08-04T23:00:59Z | MEMBER | Anecdotally I find that when I call |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/2390/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
390443869 | MDExOlB1bGxSZXF1ZXN0MjM4MjA5ODI1 | 2603 | Support HighLevelGraphs | mrocklin 306380 | closed | 0 | 2 | 2018-12-12T22:52:28Z | 2018-12-13T17:13:10Z | 2018-12-13T17:13:00Z | MEMBER | 0 | pydata/xarray/pulls/2603 | Fixes https://github.com/dask/dask/issues/4291
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/2603/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | pull | |||||
372640063 | MDExOlB1bGxSZXF1ZXN0MjI0NzY5Mjg1 | 2500 | Avoid use of deprecated get= parameter in tests | mrocklin 306380 | closed | 0 | 7 | 2018-10-22T18:25:58Z | 2018-10-23T10:31:37Z | 2018-10-23T00:22:51Z | MEMBER | 0 | pydata/xarray/pulls/2500 |
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/2500/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | pull | |||||
282178751 | MDU6SXNzdWUyODIxNzg3NTE= | 1784 | Add compute=False keywords to `to_foo` functions | mrocklin 306380 | closed | 0 | 9 | 2017-12-14T17:25:19Z | 2018-05-16T15:05:03Z | 2018-05-16T15:05:03Z | MEMBER | When working with @jhamman profiling the cc @jhamman @rabernat @jakirkham (who has been looking at similar questions within |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1784/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
295146502 | MDU6SXNzdWUyOTUxNDY1MDI= | 1894 | Zarr keys include variable name | mrocklin 306380 | closed | 0 | 1 | 2018-02-07T13:56:32Z | 2018-02-17T04:40:15Z | 2018-02-17T04:40:15Z | MEMBER | When using open_zarr on a dataset with many variables the keynames include the variable name, like
In the distributed scheduler these keynames get shortened to prefixes like We may want to avoid including the variable name into the keyname here in order to avoid breaking these out into several groups. Instead you might consider putting the variable name within the key as another member of the tuple like the following:
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1894/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
296434660 | MDExOlB1bGxSZXF1ZXN0MTY4NjIyOTM3 | 1904 | Replace task_state with tasks in dask test | mrocklin 306380 | closed | 0 | 4 | 2018-02-12T16:19:14Z | 2018-02-12T21:08:06Z | 2018-02-12T21:08:06Z | MEMBER | 0 | pydata/xarray/pulls/1904 | This internal state was changed in the latest release
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1904/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | pull | |||||
276448264 | MDExOlB1bGxSZXF1ZXN0MTU0NDI5ODYz | 1741 | Auto flake | mrocklin 306380 | closed | 0 | 2 | 2017-11-23T18:00:47Z | 2018-01-14T20:49:20Z | 2018-01-14T20:49:20Z | MEMBER | 0 | pydata/xarray/pulls/1741 |
I had a free half hour so I decided to run autoflake and autopep8 tools on the codebase. |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1741/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | pull | |||||
286448591 | MDU6SXNzdWUyODY0NDg1OTE= | 1810 | data_array.<tab> reads data | mrocklin 306380 | closed | 0 | 4 | 2018-01-06T01:34:55Z | 2018-01-06T14:26:36Z | 2018-01-06T14:26:36Z | MEMBER | Code Sample, a copy-pastable example if possible
Problem descriptionThis starts reading data. I don't know why. I'm using XArray against a FUSE system that is both expensive (it's targetting Google Cloud Storage) and also has logging. I can see that auto-completion immediately starts a lot of file reading on the file system. Output of
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1810/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
279198672 | MDExOlB1bGxSZXF1ZXN0MTU2MzQwNDk3 | 1760 | Fix DataArray.__dask_scheduler__ to point to dask.threaded.get | mrocklin 306380 | closed | 0 | 8 | 2017-12-05T00:12:21Z | 2017-12-07T22:13:42Z | 2017-12-07T22:09:18Z | MEMBER | 0 | pydata/xarray/pulls/1760 | Previously this erroneously pointed to an optimize function, likely a copy-paste error. For testing this also redirects the .compute methods to use the dask.compute function directly if dask.version >= '0.16.0'. Closes #1759
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1760/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | pull | |||||
269886091 | MDExOlB1bGxSZXF1ZXN0MTQ5NzMxMDM5 | 1674 | Support Dask interface | mrocklin 306380 | closed | 0 | 12 | 2017-10-31T09:15:52Z | 2017-11-07T18:37:06Z | 2017-11-07T18:31:45Z | MEMBER | 0 | pydata/xarray/pulls/1674 | This integrates the new dask interface methods into XArray. This will place XArray as a first-class dask collection and help in particular with newer dask.distributed features.
Builds on work from @jcrist here: https://github.com/dask/dask/pull/2748 Depends on https://github.com/dask/dask/pull/2847 |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1674/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | pull | |||||
218314868 | MDU6SXNzdWUyMTgzMTQ4Njg= | 1343 | Some XArray key names don't group nicely | mrocklin 306380 | closed | 0 | 2 | 2017-03-30T20:15:44Z | 2017-05-22T20:38:56Z | 2017-05-22T20:38:56Z | MEMBER | Some XArray loading functions provide keys that don't adhere to dask conventions used for naming. We can solve this in XArray by using names like |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1343/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
218942553 | MDExOlB1bGxSZXF1ZXN0MTEzOTM2OTk3 | 1349 | Add persist method to DataSet | mrocklin 306380 | closed | 0 | 10 | 2017-04-03T13:59:02Z | 2017-04-04T16:19:20Z | 2017-04-04T16:14:17Z | MEMBER | 0 | pydata/xarray/pulls/1349 | Fixes https://github.com/pydata/xarray/issues/1344
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1349/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
xarray 13221727 | pull | |||||
218315793 | MDU6SXNzdWUyMTgzMTU3OTM= | 1344 | Dask Persist | mrocklin 306380 | closed | 0 | 5 | 2017-03-30T20:19:17Z | 2017-04-04T16:14:17Z | 2017-04-04T16:14:17Z | MEMBER | It would be convenient to load constituent dask.arrays into memory as dask.arrays rather than as numpy arrays. This would help with distributed computations where we want to load a large amount of data into distributed memory once and then iterate on the full xarray dataset repeatedly without reloading from disk every time. We can probably solve this from either side:
```python import dask dset.x, dset.y, dset.z = dask.persist(dset.x, dset.y, dset.z) ```
cc @shoyer @jcrist @rabernat @pwolfram |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1344/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue | ||||||
187594293 | MDU6SXNzdWUxODc1OTQyOTM= | 1085 | Always use absolute paths | mrocklin 306380 | closed | 0 | 3 | 2016-11-06T22:25:08Z | 2016-12-01T16:47:40Z | 2016-12-01T16:47:40Z | MEMBER | This would avoid a mismatch between clients and workers when using dask.distributed
|
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1085/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | xarray 13221727 | issue |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issues] ( [id] INTEGER PRIMARY KEY, [node_id] TEXT, [number] INTEGER, [title] TEXT, [user] INTEGER REFERENCES [users]([id]), [state] TEXT, [locked] INTEGER, [assignee] INTEGER REFERENCES [users]([id]), [milestone] INTEGER REFERENCES [milestones]([id]), [comments] INTEGER, [created_at] TEXT, [updated_at] TEXT, [closed_at] TEXT, [author_association] TEXT, [active_lock_reason] TEXT, [draft] INTEGER, [pull_request] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [state_reason] TEXT, [repo] INTEGER REFERENCES [repos]([id]), [type] TEXT ); CREATE INDEX [idx_issues_repo] ON [issues] ([repo]); CREATE INDEX [idx_issues_milestone] ON [issues] ([milestone]); CREATE INDEX [idx_issues_assignee] ON [issues] ([assignee]); CREATE INDEX [idx_issues_user] ON [issues] ([user]);