issues: 180080354
This data as json
id | node_id | number | title | user | state | locked | assignee | milestone | comments | created_at | updated_at | closed_at | author_association | active_lock_reason | draft | pull_request | body | reactions | performed_via_github_app | state_reason | repo | type |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
180080354 | MDU6SXNzdWUxODAwODAzNTQ= | 1020 | Memory error when converting dataset to dataframe | 1961038 | closed | 0 | 4 | 2016-09-29T15:21:58Z | 2021-05-04T18:03:32Z | 2016-09-30T15:38:01Z | NONE | When working with NOAA's High Resolution Rapid Refresh (HRRR) model output data (GRIB2, automatically converted to NetCDF via either GrADS Data Server or THREDDS), xrd.to_dataframe() throws a MemoryError. Here's a sample URL that I use to assign to an xarray dataset object: 'http://nomads.ncep.noaa.gov:9090/dods/hrrr/hrrr'+model_day+'/hrrr_sfc_'+model_hour+'z' where model_day is of the form YYYYMMDD and model_hour is of the form HH These datasets are quite large (1155x2503 latxlon) ... is there a limit as to how large an xarray dataset can be for it to be converted to a dataframe? |
{ "url": "https://api.github.com/repos/pydata/xarray/issues/1020/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
completed | 13221727 | issue |