home / github

Menu
  • Search all tables
  • GraphQL API

issue_comments

Table actions
  • GraphQL API for issue_comments

3 rows where author_association = "NONE" and issue = 600268506 sorted by updated_at descending

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: created_at (date), updated_at (date)

user 2

  • pnsaevik 2
  • stale[bot] 1

issue 1

  • load_dataset fails when filename is unicode · 3 ✖

author_association 1

  • NONE · 3 ✖
id html_url issue_url node_id user created_at updated_at ▲ author_association body reactions performed_via_github_app issue
1119366649 https://github.com/pydata/xarray/issues/3974#issuecomment-1119366649 https://api.github.com/repos/pydata/xarray/issues/3974 IC_kwDOAMm_X85CuC35 pnsaevik 12728107 2022-05-06T08:13:04Z 2022-05-06T08:13:04Z NONE

After investigating the issue, it seems to be related to an error in the netcdf library, which in turn is related to an error in the hdf5 library. The relevant issues are

netcdf4: https://github.com/Unidata/netcdf4-python/issues/941 hdf5: https://forum.hdfgroup.org/t/non-english-characters-in-hdf5-file-name/4627/8

Quoting the dev responding in the hdf5 forum:

I would be thrilled to fix this issue. The problem is that it’s a huge amount of effort with no obvious funding source. Everyone wants this problem to be fixed but nobody wants it fixed so badly that they are willing to pay for an engineer to spend the better part of a year fixing it properly. A lot of people seem to think that we just need to tweak the “open file” code, but that isn’t true. So much stuff in the library is affected by Unicode file names on Windows and doing a hasty job will risk dramatically increasing our technical debt and bug count.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  load_dataset fails when filename is unicode 600268506
1114293903 https://github.com/pydata/xarray/issues/3974#issuecomment-1114293903 https://api.github.com/repos/pydata/xarray/issues/3974 IC_kwDOAMm_X85CasaP stale[bot] 26384082 2022-05-01T17:37:48Z 2022-05-01T17:37:48Z NONE

In order to maintain a list of currently relevant issues, we mark issues as stale after a period of inactivity

If this issue remains relevant, please comment here or remove the stale label; otherwise it will be marked as closed automatically

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  load_dataset fails when filename is unicode 600268506
614017215 https://github.com/pydata/xarray/issues/3974#issuecomment-614017215 https://api.github.com/repos/pydata/xarray/issues/3974 MDEyOklzc3VlQ29tbWVudDYxNDAxNzIxNQ== pnsaevik 12728107 2020-04-15T12:46:24Z 2020-04-15T12:46:24Z NONE

Update: The error does not seem to be present on linux:

INSTALLED VERSIONS

commit: None python: 3.7.6 (default, Jan 8 2020, 19:59:22) [GCC 7.3.0] python-bits: 64 OS: Linux OS-release: 3.10.0-1062.18.1.el7.x86_64 machine: x86_64 processor: x86_64 byteorder: little LC_ALL: None LANG: C.UTF-8 LOCALE: None.None libhdf5: 1.10.4 libnetcdf: 4.6.1

xarray: 0.15.0 pandas: 1.0.1 numpy: 1.18.1 scipy: 1.4.1 netCDF4: 1.4.2 pydap: None h5netcdf: None h5py: 2.10.0 Nio: None zarr: None cftime: 1.0.4.2 nc_time_axis: None PseudoNetCDF: None rasterio: None cfgrib: None iris: None bottleneck: 1.3.1 dask: 2.10.1 distributed: 2.10.0 matplotlib: 3.1.3 cartopy: 0.17.0 seaborn: 0.10.0 numbagg: None setuptools: 45.2.0.post20200210 pip: 20.0.2 conda: None pytest: 5.3.5 IPython: 7.12.0 sphinx: 2.4.0

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  load_dataset fails when filename is unicode 600268506

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [performed_via_github_app] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
    ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
    ON [issue_comments] ([user]);
Powered by Datasette · Queries took 13.363ms · About: xarray-datasette