home / github

Menu
  • GraphQL API
  • Search all tables

issues

Table actions
  • GraphQL API for issues

4 rows where repo = 13221727 and user = 327925 sorted by updated_at descending

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: created_at (date), updated_at (date), closed_at (date)

type 2

  • pull 3
  • issue 1

state 2

  • closed 2
  • open 2

repo 1

  • xarray · 4 ✖
id node_id number title user state locked assignee milestone comments created_at updated_at ▲ closed_at author_association active_lock_reason draft pull_request body reactions performed_via_github_app state_reason repo type
490228661 MDExOlB1bGxSZXF1ZXN0MzE0ODc0MDI5 3283 Add hypothesis test for netCDF4 roundtrip takluyver 327925 open 0     5 2019-09-06T09:33:48Z 2022-11-21T22:45:13Z   MEMBER   0 pydata/xarray/pulls/3283

Part of #1846: add a property-based test for reading & writing netCDF4 files.

This is the first time I've played with Hypothesis, but it seems to be working - e.g. I got an error with float16, and the netCDF docs show that 16-bit floats are not a supported data type.

However:

  • This currently only tests a dataset with a single variable - it could be extended to multiple variables if that's useful.
  • It looks like netCDF4 should support unicode characters, but it failed when I didn't have max_codepoint=255 in there. I don't know if that's an expected limitation I'm not aware of, or a bug somewhere. But I thought I'd make the test pass for now.
{
    "url": "https://api.github.com/repos/pydata/xarray/issues/3283/reactions",
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
    xarray 13221727 pull
568968607 MDU6SXNzdWU1Njg5Njg2MDc= 3786 DataArray.unstack() leaving dimensions 'in order' takluyver 327925 open 0     9 2020-02-21T13:53:40Z 2020-02-23T04:03:06Z   MEMBER      

It's documented that when you .unstack() a DataArray, "New dimensions will be added at the end."

I'm using .unstack() kind of like np.reshape(), to split one dimension into two. From this perspective, it's surprising that new dimensions go at the end, rather than in place of the old dimension. This also means that either the new array is no longer C-contiguous, or the .unstack() operation has had to copy all the data to rearrange it. It appears I'm not the only one to think this: the code sample in #2419 (about adding a reshape method) does the same workaround that I do: unstack and then carefully reorder the dimensions.

First, I'd like to understand: why the end? It seems like there's an obvious place to put unstacked dimensions (ab, c -> a, b, c), but presumably there's some reason it doesn't work that way.

Then, is there anything one can do with xarray today to 'unstack' a dimension while keeping the order, short of using .transpose() afterwards to recreate the desired order? And if not, is there any potential to add a feature for that?

I guess this might touch on missing data: unstack() can create and handle 'gaps', whereas reshape() can't. If so, is there some more limited operation that could only unstack a perfectly regular multiindex - where it's guaranteed to separate into 2 dimensions without creating gaps?

{
    "url": "https://api.github.com/repos/pydata/xarray/issues/3786/reactions",
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
    xarray 13221727 issue
490316894 MDExOlB1bGxSZXF1ZXN0MzE0OTQ2MDQw 3285 Hypothesis tests for roundtrip to & from pandas takluyver 327925 closed 0     14 2019-09-06T13:05:13Z 2020-01-10T16:25:12Z 2019-10-30T14:28:52Z MEMBER   0 pydata/xarray/pulls/3285

Part of #1846: test roundtripping between xarray DataArray & Dataset and pandas Series & DataFrame.

I haven't particularly tried to hunt down corner cases (e.g. dataframes with 0 columns), in favour of adding tests that currently pass. But these tests probably form a useful platform if you do want to ensure corner cases like that behave nicely - just modify the limits and see what fails.

{
    "url": "https://api.github.com/repos/pydata/xarray/issues/3285/reactions",
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
    xarray 13221727 pull
32919692 MDExOlB1bGxSZXF1ZXN0MTU1NTY5NTY= 113 Most of Python 3 support takluyver 327925 closed 0   0.1.1 664063 6 2014-05-06T18:31:56Z 2014-07-15T20:36:05Z 2014-05-09T01:39:01Z MEMBER   0 pydata/xarray/pulls/113

This isn't entirely finished, but I need to stop working on it for a bit, and I think enough of it is ready to be reviewed. The core code is passing its tests; the remaining failures are all in talking to the Scipy and netCDF4 backends. I also have PRs open against Scipy (scipy/scipy#3617) and netCDF4 (Unidata/netcdf4-python#252) to fix bugs I've encountered there.

Particular issues that came up: - There were quite a few circular imports. For now, I've fudged these to work rather than trying to reorganise the code. - isinstance(x, int) doesn't reliably catch numpy integer types - see e.g. numpy/numpy#2951. I changed several such cases to isinstance(x, (int, np.integer)).

{
    "url": "https://api.github.com/repos/pydata/xarray/issues/113/reactions",
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
    xarray 13221727 pull

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issues] (
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [number] INTEGER,
   [title] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [state] TEXT,
   [locked] INTEGER,
   [assignee] INTEGER REFERENCES [users]([id]),
   [milestone] INTEGER REFERENCES [milestones]([id]),
   [comments] INTEGER,
   [created_at] TEXT,
   [updated_at] TEXT,
   [closed_at] TEXT,
   [author_association] TEXT,
   [active_lock_reason] TEXT,
   [draft] INTEGER,
   [pull_request] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [performed_via_github_app] TEXT,
   [state_reason] TEXT,
   [repo] INTEGER REFERENCES [repos]([id]),
   [type] TEXT
);
CREATE INDEX [idx_issues_repo]
    ON [issues] ([repo]);
CREATE INDEX [idx_issues_milestone]
    ON [issues] ([milestone]);
CREATE INDEX [idx_issues_assignee]
    ON [issues] ([assignee]);
CREATE INDEX [idx_issues_user]
    ON [issues] ([user]);
Powered by Datasette · Queries took 26.053ms · About: xarray-datasette