issue_comments
8 rows where user = 6405510 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: issue_url, created_at (date), updated_at (date)
user 1
- richardotis · 8 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | performed_via_github_app | issue |
---|---|---|---|---|---|---|---|---|---|---|---|
176425322 | https://github.com/pydata/xarray/issues/732#issuecomment-176425322 | https://api.github.com/repos/pydata/xarray/issues/732 | MDEyOklzc3VlQ29tbWVudDE3NjQyNTMyMg== | richardotis 6405510 | 2016-01-28T21:20:45Z | 2016-01-28T21:20:45Z | NONE | For anyone else who runs into this, I built a minimal conda package for xarray 0.7 for Python 3.3. It's in the ``` yaml package: name: xarray version: 0.7.0 source: git_url: https://github.com/pydata/xarray.git git_tag: v0.7.0 requirements: build: - python - setuptools - pandas run: - python - pandas build: script: python setup.py install test: imports: - xarray about: home: https://github.com/pydata/xarray license: Apache summary: 'N-Dimensional Labeled Arrays' ``` |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
0.7 missing Python 3.3 conda package 129525746 | |
176381986 | https://github.com/pydata/xarray/issues/732#issuecomment-176381986 | https://api.github.com/repos/pydata/xarray/issues/732 | MDEyOklzc3VlQ29tbWVudDE3NjM4MTk4Ng== | richardotis 6405510 | 2016-01-28T20:11:06Z | 2016-01-28T20:11:06Z | NONE | That's what I'm going to do. Thanks for the help. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
0.7 missing Python 3.3 conda package 129525746 | |
176378656 | https://github.com/pydata/xarray/issues/732#issuecomment-176378656 | https://api.github.com/repos/pydata/xarray/issues/732 | MDEyOklzc3VlQ29tbWVudDE3NjM3ODY1Ng== | richardotis 6405510 | 2016-01-28T20:04:15Z | 2016-01-28T20:04:15Z | NONE | It occurs to me that, since xarray is in the "defaults" channel, this probably has to do with a policy at Continuum and so I might be better off just hosting a py33 version of xarray in my own conda channel. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
0.7 missing Python 3.3 conda package 129525746 | |
176374548 | https://github.com/pydata/xarray/issues/732#issuecomment-176374548 | https://api.github.com/repos/pydata/xarray/issues/732 | MDEyOklzc3VlQ29tbWVudDE3NjM3NDU0OA== | richardotis 6405510 | 2016-01-28T19:53:18Z | 2016-01-28T19:53:18Z | NONE | That's true. I just looked and saw numpy's last Python 3.3 build was 1.9.2. On the other hand, I feel that the xarray package rename has created a special case. Without the rename, pycalphad could continue its py33 support indefinitely as long as nothing happened to the relevant xarray APIs (not guaranteed, I understand that). I'm happy to drop support for deprecated versions when I have to, but with xarray being a pure Python package and seemingly happily running on the latest py33 versions of its dependencies, I don't think it's too much to ask for one more py33 build after the rename. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
0.7 missing Python 3.3 conda package 129525746 | |
176357482 | https://github.com/pydata/xarray/issues/732#issuecomment-176357482 | https://api.github.com/repos/pydata/xarray/issues/732 | MDEyOklzc3VlQ29tbWVudDE3NjM1NzQ4Mg== | richardotis 6405510 | 2016-01-28T19:28:23Z | 2016-01-28T19:28:23Z | NONE | So it appears that xray hasn't pushed a py33 conda release since 0.5.2, and I didn't notice because conda happily installed that version in my py33 tests and everything worked. With the name change, it can't find a compatible version anymore. For dependency resolution purposes I would appreciate an official py33 release of xarray 0.7, or even "xarray 0.5.2" if there's been some kind of breaking change just so that conda can find the package. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
0.7 missing Python 3.3 conda package 129525746 | |
176316930 | https://github.com/pydata/xarray/issues/732#issuecomment-176316930 | https://api.github.com/repos/pydata/xarray/issues/732 | MDEyOklzc3VlQ29tbWVudDE3NjMxNjkzMA== | richardotis 6405510 | 2016-01-28T18:16:50Z | 2016-01-28T18:16:50Z | NONE | I run py35 personally but my library pycalphad supports py27 and py33+. I noticed this issue when I updated my CI configuration to use "xarray" and only the py33 builds were failing. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
0.7 missing Python 3.3 conda package 129525746 | |
112984626 | https://github.com/pydata/xarray/issues/435#issuecomment-112984626 | https://api.github.com/repos/pydata/xarray/issues/435 | MDEyOklzc3VlQ29tbWVudDExMjk4NDYyNg== | richardotis 6405510 | 2015-06-18T00:16:19Z | 2015-06-18T00:16:19Z | NONE | xray definitely seems to be the correct tool, as you suggested. For the record, this is my first pass at coming up with the Dataset:
Full notebook: https://github.com/richardotis/pycalphad/blob/178f150b492099c32e197b417c11729f12d6dfe8/research/xrayTest.ipynb I decided I'm better off giving each phase its own Some initial queries of the data seem to function well and at a fraction of the memory cost of the pandas-based approach, so I'm feeling optimistic here. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Working with labeled N-dimensional data with combinatoric independent variables 88868867 | |
112811117 | https://github.com/pydata/xarray/issues/435#issuecomment-112811117 | https://api.github.com/repos/pydata/xarray/issues/435 | MDEyOklzc3VlQ29tbWVudDExMjgxMTExNw== | richardotis 6405510 | 2015-06-17T13:58:59Z | 2015-06-17T13:58:59Z | NONE | Thank you for your thoughts. While composing my response I realized I'm actually concerned about two distinct data representation problems. 1. Energies computed for any number of phases at discrete temperatures, pressures and compositions in a system ("energy surface data"). This is intermediate data in the computation. 2. Result of constrained equilibrium computations using the data in (1) The shape of the data in (2) would be something like (condition axis 1, condition axis 2, ..., condition axis n). Conditions can be independent or dependent variables (the solver can work backwards), and not all combinations of conditions result in an answer. For example, I want to map the phase relations of a 4-component system in T-P-x space. I choose 50 temperatures and pressures, plus 100 points per independent composition axis (here I fix the total system size so that 1 composition variable is dependent). So then the shape of my equilibrium data would be (50, 50, 100, 100, 100). But what is the value of each element, the equilibrium result? The equilibrium result is also multi-dimensional. I need to store the computed chemical potentials for each component (1-dimensional), the fraction of each stable phase (1-dimensional), and references to the corresponding physical states in the energy surface data (1-dimensional). Going back to (1), phases can also have "internal" composition variables that map to the overall composition in a non-invertible way, i.e., two physical states can have the same overall composition but different internal compositions. The way I've been handling this is by adding more columns to my DataFrames, but it's not a sustainable approach for reasons we've both mentioned. The data in (1) makes the most sense to me as a "ragged ndarray", where the internal degrees of freedom of each phase are free to be be different but still mapping to global composition coordinates. For (2), I imagine a "result object" bundled up inside all the conditions dimensions, but I need to be able to slice and search the derived/computed quantities just as easily as the independent variables. Does xray make sense for either or both of these cases? |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Working with labeled N-dimensional data with combinatoric independent variables 88868867 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [performed_via_github_app] TEXT, [issue] INTEGER REFERENCES [issues]([id]) ); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
issue 2