home / github

Menu
  • GraphQL API
  • Search all tables

commits

Table actions
  • GraphQL API for commits

1 row where author = 9658781 sorted by author_date descending

✖
✖

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: author_date (date), committer_date (date)

repo 1

  • xarray 1

author 1

  • jendrikjoe · 1 ✖
sha message author_date ▲ committer_date raw_author raw_committer repo author committer
18f35da6ac14b078772357b6bc11d3af5296afba Appending to zarr store (#2706) * Initial version of appending to zarr store * Added docs * Resolve PEP8 incompliances * Added write and append test for mode 'a' * Merged repaired master * Resolved pep8 issue * Put target store encoding in appended variable * Rewrite test with appending along time dimension * Add chunk_size parameter for rechunking appended coordinate * Add chunk_dim test * Add type check and tests for it. In append mode storing any datatype apart from number subtypes and fixed size strings raises and error. * Add documentation * Add test for compute=False and commented it out * Remove python 3.7 string formatting * Fix PEP8 incompliance * Add missing whitespaces * allowed for compute=False when appending to a zarr store * Fixed empty array data error When using create_append_test_data we used np.arange(...), which was incidently also the default value of the zarr array when fill_value is set to None. So appending to the data with compute=False and then expecting an error when asserting the source and target to be the same failed the tests. Using random data passes the tests * flake8 fixes * removed chunk_dim argument to to_zarr function * implemented requested changes * Update xarray/backends/api.py Co-Authored-By: Stephan Hoyer <shoyer@gmail.com> * added contributors and example of using append to zarr * fixed docs fail * fixed docs * removed unnecessary condition * attempt at clean string encoding and variable length strings * implemented suggestions * * append_dim does not need to be specified if creating a new array with Dataset.to_zarr(store, mode='a') * cleand up to_zarr append mode tests * raise ValueError when append_dim is not a valid dimension * flake8 fix * removed unused comment * * raise error when appending with encoding provided for existing variable * add test for encoding consistency when appending * implemented: https://github.com/pydata/xarray/pull/2706#discussion_r… 2019-06-29T23:43:14Z 2019-06-29T23:43:14Z Jendrik Jördening 1fb9e7ae004febb59559d36136b2695db0a5fcd2 Stephan Hoyer f10b21bed2846b879806f87039b77245b18e7671 xarray 13221727 jendrikjoe 9658781 shoyer 1217238

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [commits] (
   [sha] TEXT PRIMARY KEY,
   [message] TEXT,
   [author_date] TEXT,
   [committer_date] TEXT,
   [raw_author] TEXT REFERENCES [raw_authors]([id]),
   [raw_committer] TEXT REFERENCES [raw_authors]([id]),
   [repo] INTEGER REFERENCES [repos]([id]),
   [author] INTEGER REFERENCES [users]([id]),
   [committer] INTEGER REFERENCES [users]([id])
);
CREATE INDEX [idx_commits_committer]
    ON [commits] ([committer]);
CREATE INDEX [idx_commits_author]
    ON [commits] ([author]);
CREATE INDEX [idx_commits_repo]
    ON [commits] ([repo]);
CREATE INDEX [idx_commits_raw_committer]
    ON [commits] ([raw_committer]);
CREATE INDEX [idx_commits_raw_author]
    ON [commits] ([raw_author]);
Powered by Datasette · Queries took 2391.081ms · About: xarray-datasette
  • Sort ascending
  • Sort descending
  • Facet by this
  • Hide this column
  • Show all columns
  • Show not-blank rows