home / github

Menu
  • Search all tables
  • GraphQL API

issue_comments

Table actions
  • GraphQL API for issue_comments

4 rows where issue = 621968474 sorted by updated_at descending

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: created_at (date), updated_at (date)

user 3

  • JavierRuano 2
  • dcherian 1
  • hetland 1

author_association 2

  • NONE 3
  • MEMBER 1

issue 1

  • lazy evaluation of large arrays fails · 4 ✖
id html_url issue_url node_id user created_at updated_at ▲ author_association body reactions performed_via_github_app issue
631670146 https://github.com/pydata/xarray/issues/4085#issuecomment-631670146 https://api.github.com/repos/pydata/xarray/issues/4085 MDEyOklzc3VlQ29tbWVudDYzMTY3MDE0Ng== hetland 2599958 2020-05-20T19:11:56Z 2020-05-20T19:11:56Z NONE

The problem was only with very large arrays, so difficult to reproduce here.

Fortunately, when I just now updated to the very latest xarray/dask, the problem seems to have gone away.

I was puzzled because things worked fine -- as expected -- but only for smaller arrays. I will close this issue.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  lazy evaluation of large arrays fails 621968474
631658737 https://github.com/pydata/xarray/issues/4085#issuecomment-631658737 https://api.github.com/repos/pydata/xarray/issues/4085 MDEyOklzc3VlQ29tbWVudDYzMTY1ODczNw== dcherian 2448579 2020-05-20T18:50:50Z 2020-05-20T18:50:50Z MEMBER

Something is very wrong here. is this reproducible?

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  lazy evaluation of large arrays fails 621968474
631636660 https://github.com/pydata/xarray/issues/4085#issuecomment-631636660 https://api.github.com/repos/pydata/xarray/issues/4085 MDEyOklzc3VlQ29tbWVudDYzMTYzNjY2MA== JavierRuano 34353851 2020-05-20T18:07:22Z 2020-05-20T18:07:22Z NONE

I use http://xarray.pydata.org/en/stable/generated/xarray.apply_ufunc.html because it is faster.

El mié., 20 may. 2020 a las 20:01, Javier Ruano (javier.ruanno@gmail.com) escribió:

if you append compute() it should not be a lazy operation. But my advice is like user only.

El mié., 20 may. 2020 a las 19:51, Rob Hetland (notifications@github.com) escribió:

I have a large DataSet, including these DataArrays:

<xarray.DataArray 'temp' (ocean_time: 1325, s_rho: 30, eta_rho: 602, xi_rho: 677)> dask.array<concatenate, shape=(1325, 30, 602, 677), dtype=float32, chunksize=(1, 1, 602, 677)>

and

<xarray.DataArray 'zeta' (ocean_time: 1325, eta_rho: 602, xi_rho: 677)> dask.array<concatenate, shape=(1325, 602, 677), dtype=float32, chunksize=(1, 602, 677)>

(The coordinates and attributes excluded for brevity, but they match in the right ways.)

When I do math operations with the 4D DataArray (temp) and 3D DataArray (zeta), no problem:

ds.zeta * ds.temp

<xarray.DataArray (ocean_time: 1325, eta_rho: 602, xi_rho: 677, s_rho: 30)> dask.array<mul, shape=(1325, 602, 677, 30), dtype=float32, chunksize=(1, 602, 677, 1)>

This returns an object instantly, and the result is lazily evaluated. However, if I just try to add temp to itself,

ds.temp + ds.temp

this fails (eventually) as my medium sized computer runs out of memory, since it starts to evaluate the numbers as if I did a compute() or asked for the values. Note 2ds.temp or ds.temp*2 is lazily evaluated, and returns an object instantly. Chunk size does not seem to be an issue, as I have tried a number of reasonable choices without success.

Why can't such simple math operations between two large arrays also be lazily evaluated?

— You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/pydata/xarray/issues/4085, or unsubscribe https://github.com/notifications/unsubscribe-auth/AIGDFO2ZUJJOVNCN4JRJYDTRSQKBTANCNFSM4NGFVQRQ .

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  lazy evaluation of large arrays fails 621968474
631633988 https://github.com/pydata/xarray/issues/4085#issuecomment-631633988 https://api.github.com/repos/pydata/xarray/issues/4085 MDEyOklzc3VlQ29tbWVudDYzMTYzMzk4OA== JavierRuano 34353851 2020-05-20T18:01:42Z 2020-05-20T18:01:42Z NONE

if you append compute() it should not be a lazy operation. But my advice is like user only.

El mié., 20 may. 2020 a las 19:51, Rob Hetland (notifications@github.com) escribió:

I have a large DataSet, including these DataArrays:

<xarray.DataArray 'temp' (ocean_time: 1325, s_rho: 30, eta_rho: 602, xi_rho: 677)> dask.array<concatenate, shape=(1325, 30, 602, 677), dtype=float32, chunksize=(1, 1, 602, 677)>

and

<xarray.DataArray 'zeta' (ocean_time: 1325, eta_rho: 602, xi_rho: 677)> dask.array<concatenate, shape=(1325, 602, 677), dtype=float32, chunksize=(1, 602, 677)>

(The coordinates and attributes excluded for brevity, but they match in the right ways.)

When I do math operations with the 4D DataArray (temp) and 3D DataArray (zeta), no problem:

ds.zeta * ds.temp

<xarray.DataArray (ocean_time: 1325, eta_rho: 602, xi_rho: 677, s_rho: 30)> dask.array<mul, shape=(1325, 602, 677, 30), dtype=float32, chunksize=(1, 602, 677, 1)>

This returns an object instantly, and the result is lazily evaluated. However, if I just try to add temp to itself,

ds.temp + ds.temp

this fails (eventually) as my medium sized computer runs out of memory, since it starts to evaluate the numbers as if I did a compute() or asked for the values. Note 2ds.temp or ds.temp*2 is lazily evaluated, and returns an object instantly. Chunk size does not seem to be an issue, as I have tried a number of reasonable choices without success.

Why can't such simple math operations between two large arrays also be lazily evaluated?

— You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/pydata/xarray/issues/4085, or unsubscribe https://github.com/notifications/unsubscribe-auth/AIGDFO2ZUJJOVNCN4JRJYDTRSQKBTANCNFSM4NGFVQRQ .

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
  lazy evaluation of large arrays fails 621968474

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [performed_via_github_app] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
    ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
    ON [issue_comments] ([user]);
Powered by Datasette · Queries took 14.096ms · About: xarray-datasette