On Sat, Mar 24, 2018 at 07:33:40AM +0100, Nguyễn Thái Ngọc Duy wrote:

> +unsigned long oe_get_size_slow(struct packing_data *pack,
> +                            const struct object_entry *e)
> +{
> +     struct packed_git *p;
> +     struct pack_window *w_curs;
> +     unsigned char *buf;
> +     enum object_type type;
> +     unsigned long used, avail, size;
> +
> +     if (e->type_ != OBJ_OFS_DELTA && e->type_ != OBJ_REF_DELTA) {
> +             read_lock();
> +             if (sha1_object_info(e->idx.oid.hash, &size) < 0)
> +                     die(_("unable to get size of %s"),
> +                         oid_to_hex(&e->idx.oid));
> +             read_unlock();
> +             return size;
> +     }
> +
> +     p = oe_in_pack(pack, e);
> +     if (!p)
> +             die("BUG: when e->type is a delta, it must belong to a pack");
> +
> +     read_lock();
> +     w_curs = NULL;
> +     buf = use_pack(p, &w_curs, e->in_pack_offset, &avail);
> +     used = unpack_object_header_buffer(buf, avail, &type, &size);
> +     if (used == 0)
> +             die(_("unable to parse object header of %s"),
> +                 oid_to_hex(&e->idx.oid));
> +
> +     unuse_pack(&w_curs);
> +     read_unlock();
> +     return size;
> +}

It took me a while to figure out why this treated deltas and non-deltas
differently. At first I thought it was an optimization (since we can
find non-delta sizes quickly by looking at the headers).  But I think
it's just that you want to know the size of the actual _delta_, not the
reconstructed object. And there's no way to ask sha1_object_info() for
that.

Perhaps the _extended version of that function should learn an
OBJECT_INFO_NO_DEREF flag or something to tell it return the true delta
type and size. Then this whole function could just become a single call.

But short of that, it's probably worth a comment explaining what's going
on.

> +static void prepare_in_pack_by_idx(struct packing_data *pdata)
> +{
> +     struct packed_git **mapping, *p;
> +     int cnt = 0, nr = 1 << OE_IN_PACK_BITS;
> +
> +     if (getenv("GIT_TEST_FULL_IN_PACK_ARRAY")) {
> +             /*
> +              * leave in_pack_by_idx NULL to force in_pack[] to be
> +              * used instead
> +              */
> +             return;
> +     }

Minor nit, but can we use git_env_bool() here? It's just as easy, and
it's less surprising in some corner cases.

>  struct object_entry *packlist_alloc(struct packing_data *pdata,
>                                   const unsigned char *sha1,
>                                   uint32_t index_pos)
>  {
>       struct object_entry *new_entry;
>  
> +     if (!pdata->nr_objects) {
> +             prepare_in_pack_by_idx(pdata);
> +             if (getenv("GIT_TEST_OE_SIZE_BITS")) {
> +                     int bits = atoi(getenv("GIT_TEST_OE_SIZE_BITS"));;
> +                     pdata->oe_size_limit = 1 << bits;
> +             }
> +             if (!pdata->oe_size_limit)
> +                     pdata->oe_size_limit = 1 << OE_SIZE_BITS;
> +     }

Ditto here; I think this could just be:

  pdata->oe_size_limit = git_env_ulong("GIT_TEST_OE_SIZE_BITS",
                                       (1 << OE_SIZE_BITS));

>       if (pdata->nr_objects >= pdata->nr_alloc) {
>               pdata->nr_alloc = (pdata->nr_alloc  + 1024) * 3 / 2;
>               REALLOC_ARRAY(pdata->objects, pdata->nr_alloc);
> +
> +             if (!pdata->in_pack_by_idx)
> +                     REALLOC_ARRAY(pdata->in_pack, pdata->nr_alloc);
>       }

I was going to complain that we don't use ALLOC_GROW() here, but
actually that part is in the context. ;)

> @@ -35,7 +36,9 @@ enum dfs_state {
>   *
>   * "size" is the uncompressed object size. Compressed size of the raw
>   * data for an object in a pack is not stored anywhere but is computed
> - * and made available when reverse .idx is made.
> + * and made available when reverse .idx is made. Note that when an
> + * delta is reused, "size" is the uncompressed _delta_ size, not the
> + * canonical one after the delta has been applied.

s/an delta/a delta/

> +Running tests with special setups
> +---------------------------------
> +
> +The whole test suite could be run to test some special features
> +that cannot be easily covered by a few specific test cases. These
> +could be enabled by running the test suite with correct GIT_TEST_
> +environment set.
> +
> +GIT_TEST_SPLIT_INDEX forces split-index mode on the whole test suite.
> +
> +GIT_TEST_FULL_IN_PACK_ARRAY exercises the uncommon pack-objects code
> +path where there are more than 1024 packs even if the actual number of
> +packs in repository is below this limit.
> +
> +GIT_TEST_OE_SIZE_BITS=<bits> exercises the uncommon pack-objects
> +code path where we do not cache objecct size in memory and read it
> +from existing packs on demand. This normally only happens when the
> +object size is over 2GB. This variable forces the code path on any
> +object larger than 2^<bits> bytes.

It's nice to have these available to test the uncommon cases. But I have
a feeling nobody will ever run them, since it requires extra effort (and
takes a full test run).

I see there's a one-off test for GIT_TEST_FULL_IN_PACK_ARRAY, which I
think is a good idea, since it makes sure the code is exercised in a
normal test suite run. Should we do the same for GIT_TEST_OE_SIZE_BITS?

Also, s/objecct/object/. :)

> [...]

I haven't done an in-depth read of each patch yet; this was just what
jumped out at me from reading the interdiff.

-Peff

Reply via email to