home / github / issues

Menu
  • Search all tables
  • GraphQL API

issues: 293293632

This data as json

id node_id number title user state locked assignee milestone comments created_at updated_at closed_at author_association active_lock_reason draft pull_request body reactions performed_via_github_app state_reason repo type
293293632 MDU6SXNzdWUyOTMyOTM2MzI= 1874 running out of memory trying to write SQL 1794116 closed 0     3 2018-01-31T20:05:39Z 2019-02-04T04:29:09Z 2019-02-04T04:29:08Z NONE      

Python version:3 xarray version: 0.9.6

I am using xarray to read very large NetCDF files (~47G).

Then I need to write the data to a postgres DB. I have tried parsing the array and using an INSERT for every row, but this is taking a very long time (weeks).

I have read that bulk insert would be a lot faster, so I am looking for a solution along those lines.

I also saw that Pandas has a DataFrame.to_sql() function and xarray has Dataset.to_dataframe() function, so I was trying out this approach. However, when trying to convert my xarray Dataset to a Pandas Dataframe, I ran out of memory quickly. Is this expected behavior? If so can you suggest another solution to this problem?

{
    "url": "https://api.github.com/repos/pydata/xarray/issues/1874/reactions",
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  completed 13221727 issue

Links from other tables

  • 1 row from issues_id in issues_labels
  • 3 rows from issue in issue_comments
Powered by Datasette · Queries took 0.654ms · About: xarray-datasette