home / github

Menu
  • GraphQL API
  • Search all tables

issue_comments

Table actions
  • GraphQL API for issue_comments

2 rows where issue = 350899839 and user = 64892520 sorted by updated_at descending

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: created_at (date), updated_at (date)

user 1

  • ronygolderku · 2 ✖

issue 1

  • Let's list all the netCDF files that xarray can't open · 2 ✖

author_association 1

  • NONE 2
id html_url issue_url node_id user created_at updated_at ▲ author_association body reactions performed_via_github_app issue
1382724561 https://github.com/pydata/xarray/issues/2368#issuecomment-1382724561 https://api.github.com/repos/pydata/xarray/issues/2368 IC_kwDOAMm_X85SarPR ronygolderku 64892520 2023-01-14T12:11:03Z 2023-01-14T12:11:03Z NONE

@ronygolderku thanks for your example. Looks like it fails for the same reason as was mentioned for some of the other examples above.

Is there any solution?

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Let's list all the netCDF files that xarray can't open 350899839
1382669848 https://github.com/pydata/xarray/issues/2368#issuecomment-1382669848 https://api.github.com/repos/pydata/xarray/issues/2368 IC_kwDOAMm_X85Sad4Y ronygolderku 64892520 2023-01-14T05:56:44Z 2023-01-14T05:56:44Z NONE

found this one, The dataset was given based on request. That's why... Anyway, anybody want to check, you can find this polar front

data = xr.open_dataset("C:/Users/admin/Downloads/CTOH_PolarFront_weekly_1993_2019.nc")

Output.

``` MissingDimensionsError Traceback (most recent call last) ~\AppData\Local\Temp\ipykernel_9796\2161474679.py in <module> ----> 1 data = xr.open_dataset("C:/Users/admin/Downloads/CTOH_PolarFront_weekly_1993_2019.nc")

~\anaconda3\lib\site-packages\xarray\backends\api.py in open_dataset(filename_or_obj, engine, chunks, cache, decode_cf, mask_and_scale, decode_times, decode_timedelta, use_cftime, concat_characters, decode_coords, drop_variables, backend_kwargs, args, *kwargs) 493 494 overwrite_encoded_chunks = kwargs.pop("overwrite_encoded_chunks", None) --> 495 backend_ds = backend.open_dataset( 496 filename_or_obj, 497 drop_variables=drop_variables,

~\anaconda3\lib\site-packages\xarray\backends\netCDF4_.py in open_dataset(self, filename_or_obj, mask_and_scale, decode_times, concat_characters, decode_coords, drop_variables, use_cftime, decode_timedelta, group, mode, format, clobber, diskless, persist, lock, autoclose) 562 store_entrypoint = StoreBackendEntrypoint() 563 with close_on_error(store): --> 564 ds = store_entrypoint.open_dataset( 565 store, 566 mask_and_scale=mask_and_scale,

~\anaconda3\lib\site-packages\xarray\backends\store.py in open_dataset(self, store, mask_and_scale, decode_times, concat_characters, decode_coords, drop_variables, use_cftime, decode_timedelta) 37 ) 38 ---> 39 ds = Dataset(vars, attrs=attrs) 40 ds = ds.set_coords(coord_names.intersection(vars)) 41 ds.set_close(store.close)

~\anaconda3\lib\site-packages\xarray\core\dataset.py in init(self, data_vars, coords, attrs) 749 coords = coords.variables 750 --> 751 variables, coord_names, dims, indexes, _ = merge_data_and_coords( 752 data_vars, coords, compat="broadcast_equals" 753 )

~\anaconda3\lib\site-packages\xarray\core\merge.py in merge_data_and_coords(data, coords, compat, join) 486 explicit_coords = coords.keys() 487 indexes = dict(_extract_indexes_from_coords(coords)) --> 488 return merge_core( 489 objects, compat, join, explicit_coords=explicit_coords, indexes=indexes 490 )

~\anaconda3\lib\site-packages\xarray\core\merge.py in merge_core(objects, compat, join, combine_attrs, priority_arg, explicit_coords, indexes, fill_value) 635 coerced, join=join, copy=False, indexes=indexes, fill_value=fill_value 636 ) --> 637 collected = collect_variables_and_indexes(aligned) 638 639 prioritized = _get_priority_vars_and_indexes(aligned, priority_arg, compat=compat)

~\anaconda3\lib\site-packages\xarray\core\merge.py in collect_variables_and_indexes(list_of_mappings) 294 append_all(coords, indexes) 295 --> 296 variable = as_variable(variable, name=name) 297 298 if variable.dims == (name,):

~\anaconda3\lib\site-packages\xarray\core\variable.py in as_variable(obj, name) 156 # convert the Variable into an Index 157 if obj.ndim != 1: --> 158 raise MissingDimensionsError( 159 f"{name!r} has more than 1-dimension and the same name as one of its " 160 f"dimensions {obj.dims!r}. xarray disallows such variables because they "

MissingDimensionsError: 'longitude' has more than 1-dimension and the same name as one of its dimensions ('time', 'longitude'). xarray disallows such variables because they conflict with the coordinates used to label dimensions. ```

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  Let's list all the netCDF files that xarray can't open 350899839

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [performed_via_github_app] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
    ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
    ON [issue_comments] ([user]);
Powered by Datasette · Queries took 12.562ms · About: xarray-datasette