issue_comments
4 rows where issue = 1517575123 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date)
issue 1
- Implement `DataArray.to_dask_dataframe()` · 4 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
1494027739 | https://github.com/pydata/xarray/issues/7409#issuecomment-1494027739 | https://api.github.com/repos/pydata/xarray/issues/7409 | IC_kwDOAMm_X85ZDQ3b | akanshajais 85181086 | 2023-04-03T09:58:10Z | 2023-04-03T09:58:10Z | NONE | @gcaria , Your solution looks like a reasonable approach to convert a 1D or 2D chunked DataArray to a dask DataFrame or Series, respectively. this solution will only work if the DataArray is chunked along one or both dimensions. If the DataArray is not chunked, then calling to_dask() will return an equivalent in-memory pandas DataFrame or Series. One potential improvement you could make is to add a check to ensure that the chunking is valid for conversion to a dask DataFrame or Series. For example, if the chunk sizes are too small, the overhead of parallelism may outweigh the benefits. Here's an updated version of your code that includes this check: ``` import dask.dataframe as dkd import xarray as xr from typing import Union def to_dask(da: xr.DataArray) -> Union[dkd.Series, dkd.DataFrame]:
``` This code adds a check to ensure that the chunk sizes are not too small (in this case, we've set the minimum chunk size to 100,000). If any of the chunks have a size smaller than the minimum, then the function raises a ValueError. You can adjust the minimum chunk size as needed for your specific use case. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Implement `DataArray.to_dask_dataframe()` 1517575123 | |
1461877047 | https://github.com/pydata/xarray/issues/7409#issuecomment-1461877047 | https://api.github.com/repos/pydata/xarray/issues/7409 | IC_kwDOAMm_X85XInk3 | dsgreen2 106010465 | 2023-03-09T11:44:51Z | 2023-03-09T11:44:51Z | CONTRIBUTOR | Hi. I am interested in attempting to work on this issue to add the feature. Thank you. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Implement `DataArray.to_dask_dataframe()` 1517575123 | |
1398072586 | https://github.com/pydata/xarray/issues/7409#issuecomment-1398072586 | https://api.github.com/repos/pydata/xarray/issues/7409 | IC_kwDOAMm_X85TVOUK | gcaria 44147817 | 2023-01-20T08:39:31Z | 2023-01-20T08:39:31Z | CONTRIBUTOR | Yes I did, but unfortunately I didn't think about the trick of converting to Dataset. My only thought then is that it'd be nice to also get a dask Series, so something like Feel free to close this. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Implement `DataArray.to_dask_dataframe()` 1517575123 | |
1381050335 | https://github.com/pydata/xarray/issues/7409#issuecomment-1381050335 | https://api.github.com/repos/pydata/xarray/issues/7409 | IC_kwDOAMm_X85SUSff | dcherian 2448579 | 2023-01-12T22:11:25Z | 2023-01-12T22:11:25Z | MEMBER | Have you seen https://docs.xarray.dev/en/stable/generated/xarray.Dataset.to_dask_dataframe.html? I thjink we could easily add a version for DataArray that creates a temporary dataset and then calls |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Implement `DataArray.to_dask_dataframe()` 1517575123 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 4