issue_comments
2 rows where author_association = "NONE" and user = 59902324 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date)
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
800374879 | https://github.com/pydata/xarray/issues/2368#issuecomment-800374879 | https://api.github.com/repos/pydata/xarray/issues/2368 | MDEyOklzc3VlQ29tbWVudDgwMDM3NDg3OQ== | ognancy4life 59902324 | 2021-03-16T15:42:25Z | 2021-03-16T15:42:25Z | NONE | @dcherian Thanks for your reply. I think I understand the issue. What, specifically, do you suggest to fix this issue in my own code considering this is not a dataset I generated? |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Let's list all the netCDF files that xarray can't open 350899839 | |
785298202 | https://github.com/pydata/xarray/issues/2368#issuecomment-785298202 | https://api.github.com/repos/pydata/xarray/issues/2368 | MDEyOklzc3VlQ29tbWVudDc4NTI5ODIwMg== | ognancy4life 59902324 | 2021-02-24T18:54:35Z | 2021-02-24T18:56:11Z | NONE | Found one! https://www.ncei.noaa.gov/data/oceans/ncei/ocads/data/0191304/ The dataset published in Bushinsky et al. (2019), which is basically the Landshutzer et al. (2014) climatology plus SOCCOM Float-based pCO2 data, and updated through 2018. I've only tried the first file in the list (https://www.ncei.noaa.gov/data/oceans/ncei/ocads/data/0191304/MPI-SOM_FFN_SOCCOMv2018.nc), but suspect the others will have the same issue. Here's the error (sounds like you all have discussed before, but I can't see an easy answer): ```MissingDimensionsError Traceback (most recent call last) <ipython-input-4-9e0af51f1c05> in <module> ----> 1 SOMFFN = xr.open_dataset('MPI-SOM_FFN_SOCCOMv2018.nc') ~/opt/anaconda3/lib/python3.8/site-packages/xarray/backends/api.py in open_dataset(filename_or_obj, group, decode_cf, mask_and_scale, decode_times, autoclose, concat_characters, decode_coords, engine, chunks, lock, cache, drop_variables, backend_kwargs, use_cftime, decode_timedelta) 573 574 with close_on_error(store): --> 575 ds = maybe_decode_store(store, chunks) 576 577 # Ensure source filename always stored in dataset object (GH issue #2550) ~/opt/anaconda3/lib/python3.8/site-packages/xarray/backends/api.py in maybe_decode_store(store, chunks) 469 470 def maybe_decode_store(store, chunks): --> 471 ds = conventions.decode_cf( 472 store, 473 mask_and_scale=mask_and_scale, ~/opt/anaconda3/lib/python3.8/site-packages/xarray/conventions.py in decode_cf(obj, concat_characters, mask_and_scale, decode_times, decode_coords, drop_variables, use_cftime, decode_timedelta) 598 decode_timedelta=decode_timedelta, 599 ) --> 600 ds = Dataset(vars, attrs=attrs) 601 ds = ds.set_coords(coord_names.union(extra_coords).intersection(vars)) 602 ds._file_obj = file_obj ~/opt/anaconda3/lib/python3.8/site-packages/xarray/core/dataset.py in init(self, data_vars, coords, attrs) 628 coords = coords.variables 629 --> 630 variables, coord_names, dims, indexes, _ = merge_data_and_coords( 631 data_vars, coords, compat="broadcast_equals" 632 ) ~/opt/anaconda3/lib/python3.8/site-packages/xarray/core/merge.py in merge_data_and_coords(data, coords, compat, join) 465 explicit_coords = coords.keys() 466 indexes = dict(_extract_indexes_from_coords(coords)) --> 467 return merge_core( 468 objects, compat, join, explicit_coords=explicit_coords, indexes=indexes 469 ) ~/opt/anaconda3/lib/python3.8/site-packages/xarray/core/merge.py in merge_core(objects, compat, join, combine_attrs, priority_arg, explicit_coords, indexes, fill_value) 592 coerced, join=join, copy=False, indexes=indexes, fill_value=fill_value 593 ) --> 594 collected = collect_variables_and_indexes(aligned) 595 596 prioritized = _get_priority_vars_and_indexes(aligned, priority_arg, compat=compat) ~/opt/anaconda3/lib/python3.8/site-packages/xarray/core/merge.py in collect_variables_and_indexes(list_of_mappings) 276 append_all(coords, indexes) 277 --> 278 variable = as_variable(variable, name=name) 279 if variable.dims == (name,): 280 variable = variable.to_index_variable() ~/opt/anaconda3/lib/python3.8/site-packages/xarray/core/variable.py in as_variable(obj, name) 152 # convert the Variable into an Index 153 if obj.ndim != 1: --> 154 raise MissingDimensionsError( 155 "%r has more than 1-dimension and the same name as one of its " 156 "dimensions %r. xarray disallows such variables because they " MissingDimensionsError: 'date' has more than 1-dimension and the same name as one of its dimensions ('time', 'date'). xarray disallows such variables because they conflict with the coordinates used to label dimensions. ``` |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Let's list all the netCDF files that xarray can't open 350899839 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 1