issue_comments
15 rows where author_association = "MEMBER", issue = 94328498 and user = 1217238 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: reactions, created_at (date), updated_at (date)
issue 1
- open_mfdataset too many files · 15 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
347157526 | https://github.com/pydata/xarray/issues/463#issuecomment-347157526 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDM0NzE1NzUyNg== | shoyer 1217238 | 2017-11-27T11:40:35Z | 2017-11-27T11:40:35Z | MEMBER | Using autoclose=True should also fix this. On Mon, Nov 27, 2017 at 10:26 AM Sebastian Hahn notifications@github.com wrote:
|
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
288832922 | https://github.com/pydata/xarray/issues/463#issuecomment-288832922 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDI4ODgzMjkyMg== | shoyer 1217238 | 2017-03-23T19:22:43Z | 2017-03-23T19:22:43Z | MEMBER | OK, I'm closing this issue as "Fixed" by #1198. Feel free to open new issue for any follow-up concerns. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
263734251 | https://github.com/pydata/xarray/issues/463#issuecomment-263734251 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDI2MzczNDI1MQ== | shoyer 1217238 | 2016-11-29T23:30:02Z | 2016-11-29T23:30:02Z | MEMBER |
Yes, exactly. I plan to merge that PR very shortly, after a few fixes for the failing tests on Windows (less than an hour of work). |
{ "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
263706346 | https://github.com/pydata/xarray/issues/463#issuecomment-263706346 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDI2MzcwNjM0Ng== | shoyer 1217238 | 2016-11-29T21:35:06Z | 2016-11-29T21:35:06Z | MEMBER | @pwolfram NcML is just an XML specification for how variables in a set of NetCDF files can be combined into a single virtual NetCDF file. This would be useful because it would allow building a version of I suspect that even the LRU cache approach would build on |
{ "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
263652409 | https://github.com/pydata/xarray/issues/463#issuecomment-263652409 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDI2MzY1MjQwOQ== | shoyer 1217238 | 2016-11-29T18:17:17Z | 2016-11-29T18:17:17Z | MEMBER |
Sure. This should probably be a different wrapper function than @kmpaul thanks for sharing! This is useful background. There is at least one other option worth considering. Instead of using the open file LRU cache, a simpler option could be to add an optional argument to xarray backends (building on |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
263437709 | https://github.com/pydata/xarray/issues/463#issuecomment-263437709 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDI2MzQzNzcwOQ== | shoyer 1217238 | 2016-11-29T00:19:53Z | 2016-11-29T00:19:53Z | MEMBER |
The LRU cache solution proposed in https://github.com/pydata/xarray/issues/798 would work in either case. It just would have poor performance when accessing a small piece of each of 10^6 files, both to build the graph (because xarray needs to open each file to read the metadata) and to do the actual computation (again, because of the need to open so many files). If you only need a small amount of data from many files, you probably want to reshape your data to minimize the amount of necessary file access no matter what, whether you do that reshaping with PyReshaper or xarray/dask.array/dask-distributed. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
223838593 | https://github.com/pydata/xarray/issues/463#issuecomment-223838593 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDIyMzgzODU5Mw== | shoyer 1217238 | 2016-06-05T21:23:41Z | 2016-06-05T21:23:41Z | MEMBER | @mangecoeur I can take a look. Can you share an example of how you use the |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
223663026 | https://github.com/pydata/xarray/issues/463#issuecomment-223663026 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDIyMzY2MzAyNg== | shoyer 1217238 | 2016-06-03T18:53:22Z | 2016-06-03T18:53:22Z | MEMBER | I suspect you hit this in IPython after rerunning cells, because file handles are only automatically closed when programs exit. You might find it a good idea to explicitly close files by calling .close() (or using a "with" statement) on Datasets opened with open_mfdataset. On Fri, Jun 3, 2016 at 11:08 AM, mangecoeur notifications@github.com wrote:
|
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
143382040 | https://github.com/pydata/xarray/issues/463#issuecomment-143382040 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDE0MzM4MjA0MA== | shoyer 1217238 | 2015-09-26T00:22:51Z | 2015-09-26T00:22:51Z | MEMBER | OK, I think you could also just add an ensured_open() to the repr() method. Right now that class is inheriting it from NDArrayMixin. On Fri, Sep 25, 2015 at 5:11 PM, Christoph Paulik notifications@github.com wrote:
|
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
143347373 | https://github.com/pydata/xarray/issues/463#issuecomment-143347373 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDE0MzM0NzM3Mw== | shoyer 1217238 | 2015-09-25T20:35:38Z | 2015-09-25T20:35:38Z | MEMBER | OK, so the problem is that |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
143325053 | https://github.com/pydata/xarray/issues/463#issuecomment-143325053 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDE0MzMyNTA1Mw== | shoyer 1217238 | 2015-09-25T19:06:51Z | 2015-09-25T19:06:51Z | MEMBER | @cpaulik I wonder if the issue is this section in your
I would put Actually, you probably want to put this in a context manager that automatically closes the file, something like:
|
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
142675701 | https://github.com/pydata/xarray/issues/463#issuecomment-142675701 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDE0MjY3NTcwMQ== | shoyer 1217238 | 2015-09-23T17:41:49Z | 2015-09-23T17:41:49Z | MEMBER | I think we can actually read all the variable metadata (shape and dtype) in when we open the file -- we already do that for reading in attributes. Something like this prototype, which would also be useful for reading compressed netCDF4 files with multiprocessing: https://github.com/blaze/dask/pull/457#issuecomment-123512166 |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
120666380 | https://github.com/pydata/xarray/issues/463#issuecomment-120666380 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDEyMDY2NjM4MA== | shoyer 1217238 | 2015-07-11T22:36:30Z | 2015-07-11T22:36:30Z | MEMBER | Hmm. How big are each of your netCDF files? |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
120448308 | https://github.com/pydata/xarray/issues/463#issuecomment-120448308 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDEyMDQ0ODMwOA== | shoyer 1217238 | 2015-07-10T16:12:52Z | 2015-07-10T16:12:52Z | MEMBER | Sure, you could do this on the scipy backend -- the logic will be essentially the same on both backends. I believe your issue with netCDF4 backend is the same as this one: https://github.com/xray/xray/issues/444. This will be fixed in the next release. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 | |
120443929 | https://github.com/pydata/xarray/issues/463#issuecomment-120443929 | https://api.github.com/repos/pydata/xarray/issues/463 | MDEyOklzc3VlQ29tbWVudDEyMDQ0MzkyOQ== | shoyer 1217238 | 2015-07-10T15:58:41Z | 2015-07-10T15:58:41Z | MEMBER | Yes, this is a known issue, and I agree that it is annoying. We could work around this by opening up (and closing) netCDF files inside the |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
open_mfdataset too many files 94328498 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 1