issue_comments
8 rows where user = 11075246 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: reactions, created_at (date), updated_at (date)
| id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
|---|---|---|---|---|---|---|---|---|---|---|---|
| 1272558504 | https://github.com/pydata/xarray/issues/7146#issuecomment-1272558504 | https://api.github.com/repos/pydata/xarray/issues/7146 | IC_kwDOAMm_X85L2bOo | d1mach 11075246 | 2022-10-09T14:49:33Z | 2022-10-09T14:49:33Z | NONE | I had to change ints and floats to doubles to reproduce the issue. ```python import h5py N_TIMES = 48 with h5py.File("/my_s3_fs/test.nc", mode="w") as f: time = f.create_dataset("time", (N_TIMES,), dtype="d") time[:] = 0
``` |
{
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Segfault writing large netcdf files to s3fs 1402002645 | |
| 1272553921 | https://github.com/pydata/xarray/issues/7146#issuecomment-1272553921 | https://api.github.com/repos/pydata/xarray/issues/7146 | IC_kwDOAMm_X85L2aHB | d1mach 11075246 | 2022-10-09T14:27:06Z | 2022-10-09T14:27:06Z | NONE | datatype seems to be not important. But the two variables are required to get a segfault. The following with just floats produces a segfault ``` import numpy as np import xarray as xr N_TIMES=48 ds = xr.Dataset({"time": ("T", np.zeros((N_TIMES))), 'd1': (["T", "x", "y"], np.zeros((N_TIMES, 201,201)))}) ds.to_netcdf(path="/my_s3_fs/test_netcdf.nc", format="NETCDF4", mode="w") ``` |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Segfault writing large netcdf files to s3fs 1402002645 | |
| 1272544819 | https://github.com/pydata/xarray/issues/7146#issuecomment-1272544819 | https://api.github.com/repos/pydata/xarray/issues/7146 | IC_kwDOAMm_X85L2X4z | d1mach 11075246 | 2022-10-09T13:37:57Z | 2022-10-09T14:25:51Z | NONE | It seems that we need the time variable to reproduce the problem. The following code does not fail: ``` import numpy as np import xarray as xr import pandas as pd N_TIMES=64 ds = xr.Dataset({'d1': (["T", "x", "y"], np.zeros((N_TIMES, 201,201)))}) ds.to_netcdf(path="/my_s3_fs/test_netcdf.nc", format="NETCDF4", mode="w") ``` |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Segfault writing large netcdf files to s3fs 1402002645 | |
| 1272541759 | https://github.com/pydata/xarray/issues/7146#issuecomment-1272541759 | https://api.github.com/repos/pydata/xarray/issues/7146 | IC_kwDOAMm_X85L2XI_ | d1mach 11075246 | 2022-10-09T13:21:22Z | 2022-10-09T13:21:40Z | NONE | The first one results in a segfault: ```python import numpy as np import xarray as xr import pandas as pd N_TIMES = 48 time_vals = pd.date_range("2022-10-06", freq="20 min", periods=N_TIMES) ds = xr.Dataset({"time": ("T", time_vals), 'd1': (["T", "x", "y"], np.zeros((len(time_vals), 201,201)))}) ds.to_netcdf(path="/my_s3_fs/test_netcdf.nc", format="NETCDF4", mode="w") ``` Not sure how to add the 3D var to the second dataset. |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Segfault writing large netcdf files to s3fs 1402002645 | |
| 1272539102 | https://github.com/pydata/xarray/issues/7146#issuecomment-1272539102 | https://api.github.com/repos/pydata/xarray/issues/7146 | IC_kwDOAMm_X85L2Wfe | d1mach 11075246 | 2022-10-09T13:08:26Z | 2022-10-09T13:08:26Z | NONE | Will try to reproduce this with h5py. For the bug to show up the file has to be large enough. That is why my example has a 2D array variable alongside the time dimension. With just the time dimension the script completes without an error. All three cases work without an error: |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Segfault writing large netcdf files to s3fs 1402002645 | |
| 1272514059 | https://github.com/pydata/xarray/issues/7146#issuecomment-1272514059 | https://api.github.com/repos/pydata/xarray/issues/7146 | IC_kwDOAMm_X85L2QYL | d1mach 11075246 | 2022-10-09T10:48:35Z | 2022-10-09T10:48:35Z | NONE | Adding a gdb stackrace.txt from corefile obtained with
|
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Segfault writing large netcdf files to s3fs 1402002645 | |
| 1272365368 | https://github.com/pydata/xarray/issues/7146#issuecomment-1272365368 | https://api.github.com/repos/pydata/xarray/issues/7146 | IC_kwDOAMm_X85L1sE4 | d1mach 11075246 | 2022-10-08T17:37:24Z | 2022-10-08T17:37:24Z | NONE | libnetcdf, netcdf4 and hdf5 are at their latest versions available on conda-forge |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Segfault writing large netcdf files to s3fs 1402002645 | |
| 1272364031 | https://github.com/pydata/xarray/issues/7146#issuecomment-1272364031 | https://api.github.com/repos/pydata/xarray/issues/7146 | IC_kwDOAMm_X85L1rv_ | d1mach 11075246 | 2022-10-08T17:30:41Z | 2022-10-08T17:30:41Z | NONE | Can confirm the issue with xarray 2022.6.0 and dask 2022.9.2. The latest versions available on conda-forge. The issue might be related to netcdf4 and hdf5 libraries. Will try to update this as well. |
{
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
Segfault writing large netcdf files to s3fs 1402002645 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] (
[html_url] TEXT,
[issue_url] TEXT,
[id] INTEGER PRIMARY KEY,
[node_id] TEXT,
[user] INTEGER REFERENCES [users]([id]),
[created_at] TEXT,
[updated_at] TEXT,
[author_association] TEXT,
[body] TEXT,
[reactions] TEXT,
[performed_via_github_app] TEXT,
[issue] INTEGER REFERENCES [issues]([id])
);
CREATE INDEX [idx_issue_comments_issue]
ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
ON [issue_comments] ([user]);
user 1