issue_comments: 507536039
This data as json
| html_url | issue_url | id | node_id | user | created_at | updated_at | author_association | body | reactions | performed_via_github_app | issue |
|---|---|---|---|---|---|---|---|---|---|---|---|
| https://github.com/pydata/xarray/pull/3070#issuecomment-507536039 | https://api.github.com/repos/pydata/xarray/issues/3070 | 507536039 | MDEyOklzc3VlQ29tbWVudDUwNzUzNjAzOQ== | 5635139 | 2019-07-02T06:19:55Z | 2019-07-02T06:19:55Z | MEMBER | Hmmm, this is my code. Checking what I missed. We use the similar code internally and it works (I think...) Do we have any flaky tests? I only see one that's marked, and that's skipped anyway (maybe because of this problem): ``` rg flaky azure-pipelines.yml 24: pytest_extra_flags: --run-flaky --run-network-tests setup.cfg 13: flaky: flaky tests conftest.py 8: parser.addoption("--run-flaky", action="store_true", 9: help="runs flaky tests") 16: if not config.getoption("--run-flaky"): 17: skip_flaky = pytest.mark.skip( 18: reason="set --run-flaky option to run flaky tests") 20: if "flaky" in item.keywords: 21: item.add_marker(skip_flaky) doc/whats-new.rst
1847:- Enhanced tests suite by use of xarray/tests/init.py 112:flaky = pytest.mark.flaky xarray/tests/test_plot.py 30:@pytest.mark.flaky 31:@pytest.mark.skip(reason='maybe flaky') ``` |
{
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} |
463021297 |