issue_comments
11 rows where issue = 480753417 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date)
issue 1
- Feature request: time-based rolling window functionality · 11 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
1499591643 | https://github.com/pydata/xarray/issues/3216#issuecomment-1499591643 | https://api.github.com/repos/pydata/xarray/issues/3216 | IC_kwDOAMm_X85ZYfPb | chiaral 8453445 | 2023-04-06T20:34:19Z | 2023-04-06T20:34:47Z | CONTRIBUTOR | Hello!
Just adding a 👍 to this thread - and, since it is an old issue, wondering if this is on |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Feature request: time-based rolling window functionality 480753417 | |
871393973 | https://github.com/pydata/xarray/issues/3216#issuecomment-871393973 | https://api.github.com/repos/pydata/xarray/issues/3216 | MDEyOklzc3VlQ29tbWVudDg3MTM5Mzk3Mw== | hCraker 42781301 | 2021-06-30T13:14:13Z | 2022-02-08T19:16:54Z | NONE | I understand now. I was referring to the part of the opening comment talking about averages in an hour or minute windows. That can be done with a couple lines of code, but you're right that the data has to be evenly spaced. I'm not sure how to make the varied windows work at all (let alone making it efficient) so I will leave this to you all. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Feature request: time-based rolling window functionality 480753417 | |
871665765 | https://github.com/pydata/xarray/issues/3216#issuecomment-871665765 | https://api.github.com/repos/pydata/xarray/issues/3216 | MDEyOklzc3VlQ29tbWVudDg3MTY2NTc2NQ== | max-sixty 5635139 | 2021-06-30T19:21:38Z | 2021-06-30T19:21:38Z | MEMBER | Pandas has this, so it's not intractable. If you'd like the feature, add a 👍 to the issue or help it along by looking at what would be required / starting an implementation. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Feature request: time-based rolling window functionality 480753417 | |
870915351 | https://github.com/pydata/xarray/issues/3216#issuecomment-870915351 | https://api.github.com/repos/pydata/xarray/issues/3216 | MDEyOklzc3VlQ29tbWVudDg3MDkxNTM1MQ== | dcherian 2448579 | 2021-06-29T21:03:27Z | 2021-06-29T21:03:27Z | MEMBER | @hCraker that's not right. rolling works by using fixed length windows so what you are doing only works with evenly spaced data. What's being discussed here is rolling with windows of varying lengths. I don't know how to do that efficiently. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Feature request: time-based rolling window functionality 480753417 | |
870865120 | https://github.com/pydata/xarray/issues/3216#issuecomment-870865120 | https://api.github.com/repos/pydata/xarray/issues/3216 | MDEyOklzc3VlQ29tbWVudDg3MDg2NTEyMA== | hCraker 42781301 | 2021-06-29T19:42:17Z | 2021-06-29T19:42:17Z | NONE | Hi all. This functionality can be done in xarray, but it's not a simple one line call. Currently this sort of functionality is being added to the geocat-comp repository in PR 158 https://github.com/NCAR/geocat-comp/pull/158 which should be merged and added to the July release here in the next few weeks. @dcherian perhaps we could chat about whether or not this should remain in geocat-comp as is or if it could be done more efficiently in xarray's backend |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Feature request: time-based rolling window functionality 480753417 | |
834466193 | https://github.com/pydata/xarray/issues/3216#issuecomment-834466193 | https://api.github.com/repos/pydata/xarray/issues/3216 | MDEyOklzc3VlQ29tbWVudDgzNDQ2NjE5Mw== | bhemmer 22427415 | 2021-05-07T14:38:59Z | 2021-05-07T14:38:59Z | NONE | Is there a chance this might be added? I would also highly appreciate this feature. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Feature request: time-based rolling window functionality 480753417 | |
618332209 | https://github.com/pydata/xarray/issues/3216#issuecomment-618332209 | https://api.github.com/repos/pydata/xarray/issues/3216 | MDEyOklzc3VlQ29tbWVudDYxODMzMjIwOQ== | snbentley 7360639 | 2020-04-23T10:52:40Z | 2020-04-23T10:52:40Z | NONE | This would still be very useful to me in future - for the piece of work I was referring to here I came up with a workaround. I filled in the gaps roughly with NaNs, so that I could identify and remove outliers and other bad data. Only then could I use the resample functionality without smearing these artefacts across good data. However, my solution was quite clunky and slow and was based on the still-mostly-regular resolution of my dataset, rather than any neater general solution in pandas. As I was (and am) also relatively new to Python I did not think this was appropriate to add to xarray myself, but I would like to say that I would definitely use this functionality in future - as would the other colleagues in space physics/meteorology I mentioned this to. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Feature request: time-based rolling window functionality 480753417 | |
618050489 | https://github.com/pydata/xarray/issues/3216#issuecomment-618050489 | https://api.github.com/repos/pydata/xarray/issues/3216 | MDEyOklzc3VlQ29tbWVudDYxODA1MDQ4OQ== | max-sixty 5635139 | 2020-04-22T21:28:59Z | 2020-04-22T21:28:59Z | MEMBER | We would definitely take a PR for this; and it might not be that difficult given it's already implemented in pandas. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Feature request: time-based rolling window functionality 480753417 | |
618006150 | https://github.com/pydata/xarray/issues/3216#issuecomment-618006150 | https://api.github.com/repos/pydata/xarray/issues/3216 | MDEyOklzc3VlQ29tbWVudDYxODAwNjE1MA== | mattrossman 22670878 | 2020-04-22T19:58:30Z | 2020-04-22T19:58:30Z | NONE | I'm surprised this feature still hasn't made its way from pandas to xarray, it's incredibly helpful for datasets that are not evenly sampled. Resampling and calculating the integer window size feels unnecessary for the end goal. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Feature request: time-based rolling window functionality 480753417 | |
521322678 | https://github.com/pydata/xarray/issues/3216#issuecomment-521322678 | https://api.github.com/repos/pydata/xarray/issues/3216 | MDEyOklzc3VlQ29tbWVudDUyMTMyMjY3OA== | snbentley 7360639 | 2019-08-14T16:38:07Z | 2019-08-14T16:38:07Z | NONE | Hi, I did actually just see this - it would solve the unevenly sampled data part but really I need to identify the unphysical values that are not tagged by the quality flags first. Once that has been done then resampling and interpolation would be great - but otherwise I will be spreading the effect of bad data. For this particular set of data I am looking at, I often get individual points which are close to but clearly outliers from the time series so examining a rolling mean would help find these. That is the example I was hoping to solve with this query, but I have already realised that this extends to other problems I will encounter. For example, sudden jumps in the time series (for which I have been recommended to calculate rolling correlation coefficients between two time series) and multiple points jumping all over the place (for which I will probably compare the variance of groups of points and a rolling gradient) (I really don't know why these aren't cleaned better first, but unfortunately that is the way things are) Because I need to clean the data before any analysis, the resampling method would probably allow me to get rid of most but not all the bad data. Then I would have to be extra-cautious and throw out lots of possibly good observations just in case. I will definitely use resampling for the analysis but there are so many ways that this would be helpful at the processing stage. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Feature request: time-based rolling window functionality 480753417 | |
521303238 | https://github.com/pydata/xarray/issues/3216#issuecomment-521303238 | https://api.github.com/repos/pydata/xarray/issues/3216 | MDEyOklzc3VlQ29tbWVudDUyMTMwMzIzOA== | dcherian 2448579 | 2019-08-14T15:47:04Z | 2019-08-14T15:47:04Z | MEMBER | Does |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Feature request: time-based rolling window functionality 480753417 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 7