6a1e32d532
When we serve a fetch, we pass the "wants" and "haves" from
the fetch negotiation to pack-objects. That tells us not
only which objects we need to send, but we also use the
boundary commits as "preferred bases": their trees and blobs
are candidates for delta bases, both for reusing on-disk
deltas and for finding new ones.
However, this misses some opportunities. Modulo some special
cases like shallow or partial clones, we know that every
object reachable from the "haves" could be a preferred base.
We don't use all of them for two reasons:
1. It's expensive to traverse the whole history and
enumerate all of the objects the other side has.
2. The delta search is expensive, so we want to keep the
number of candidate bases sane. The boundary commits
are the most likely to work.
When we have reachability bitmaps, though, reason 1 no
longer applies. We can efficiently compute the set of
reachable objects on the other side (and in fact already did
so as part of the bitmap set-difference to get the list of
interesting objects). And using this set conveniently
covers the shallow and partial cases, since we have to
disable the use of bitmaps for those anyway.
The second reason argues against using these bases in the
search for new deltas. But there's one case where we can use
this information for free: when we have an existing on-disk
delta that we're considering reusing, we can do so if we
know the other side has the base object. This in fact saves
time during the delta search, because it's one less delta we
have to compute.
And that's exactly what this patch does: when we're
considering whether to reuse an on-disk delta, if bitmaps
tell us the other side has the object (and we're making a
thin-pack), then we reuse it.
Here are the results on p5311 using linux.git, which
simulates a client fetching after `N` days since their last
fetch:
Test origin HEAD
--------------------------------------------------------------------------
5311.3: server (1 days) 0.27(0.27+0.04) 0.12(0.09+0.03) -55.6%
5311.4: size (1 days) 0.9M 237.0K -73.7%
5311.5: client (1 days) 0.04(0.05+0.00) 0.10(0.10+0.00) +150.0%
5311.7: server (2 days) 0.34(0.42+0.04) 0.13(0.10+0.03) -61.8%
5311.8: size (2 days) 1.5M 347.7K -76.5%
5311.9: client (2 days) 0.07(0.08+0.00) 0.16(0.15+0.01) +128.6%
5311.11: server (4 days) 0.56(0.77+0.08) 0.13(0.10+0.02) -76.8%
5311.12: size (4 days) 2.8M 566.6K -79.8%
5311.13: client (4 days) 0.13(0.15+0.00) 0.34(0.31+0.02) +161.5%
5311.15: server (8 days) 0.97(1.39+0.11) 0.30(0.25+0.05) -69.1%
5311.16: size (8 days) 4.3M 1.0M -76.0%
5311.17: client (8 days) 0.20(0.22+0.01) 0.53(0.52+0.01) +165.0%
5311.19: server (16 days) 1.52(2.51+0.12) 0.30(0.26+0.03) -80.3%
5311.20: size (16 days) 8.0M 2.0M -74.5%
5311.21: client (16 days) 0.40(0.47+0.03) 1.01(0.98+0.04) +152.5%
5311.23: server (32 days) 2.40(4.44+0.20) 0.31(0.26+0.04) -87.1%
5311.24: size (32 days) 14.1M 4.1M -70.9%
5311.25: client (32 days) 0.70(0.90+0.03) 1.81(1.75+0.06) +158.6%
5311.27: server (64 days) 11.76(26.57+0.29) 0.55(0.50+0.08) -95.3%
5311.28: size (64 days) 89.4M 47.4M -47.0%
5311.29: client (64 days) 5.71(9.31+0.27) 15.20(15.20+0.32) +166.2%
5311.31: server (128 days) 16.15(36.87+0.40) 0.91(0.82+0.14) -94.4%
5311.32: size (128 days) 134.8M 100.4M -25.5%
5311.33: client (128 days) 9.42(16.86+0.49) 25.34(25.80+0.46) +169.0%
In all cases we save CPU time on the server (sometimes
significant) and the resulting pack is smaller. We do spend
more CPU time on the client side, because it has to
reconstruct more deltas. But that's the right tradeoff to
make, since clients tend to outnumber servers. It just means
the thin pack mechanism is doing its job.
From the user's perspective, the end-to-end time of the
operation will generally be faster. E.g., in the 128-day
case, we saved 15s on the server at a cost of 16s on the
client. Since the resulting pack is 34MB smaller, this is a
net win if the network speed is less than 270Mbit/s. And
that's actually the worst case. The 64-day case saves just
over 11s at a cost of just under 11s. So it's a slight win
at any network speed, and the 40MB saved is pure bonus. That
trend continues for the smaller fetches.
The implementation itself is mostly straightforward, with
the new logic going into check_object(). But there are two
tricky bits.
The first is that check_object() needs access to the
relevant information (the thin flag and bitmap result). We
can do this by pushing these into program-lifetime globals.
The second is that the rest of the code assumes that any
reused delta will point to another "struct object_entry" as
its base. But of course the case we are interested in here
is the one where don't have such an entry!
I looked at a number of options that didn't quite work:
- we could use a flag to signal a reused delta, but it's
not a single bit. We have to actually store the oid of
the base, which is normally done by pointing to the
existing object_entry. And we'd have to modify all the
code which looks at deltas.
- we could add the reused bases to the end of the existing
object_entry array. While this does create some extra
work as later stages consider the extra entries, it's
actually not too bad (we're not sending them, so they
don't cost much in the delta search, and at most we'd
have 2*N of them).
But there's a more subtle problem. Adding to the existing
array means we might need to grow it with realloc, which
could move the earlier entries around. While many of the
references to other entries are done by integer index,
some (including ones on the stack) use pointers, which
would become invalidated.
This isn't insurmountable, but it would require quite a
bit of refactoring (and it's hard to know that you've got
it all, since it may work _most_ of the time and then
fail subtly based on memory allocation patterns).
- we could allocate a new one-off entry for the base. In
fact, this is what an earlier version of this patch did.
However, since the refactoring brought in by ad635e82d6
(Merge branch 'nd/pack-objects-pack-struct', 2018-05-23),
the delta_idx code requires that both entries be in the
main packing list.
So taking all of those options into account, what I ended up
with is a separate list of "external bases" that are not
part of the main packing list. Each delta entry that points
to an external base has a single-bit flag to do so; we have a
little breathing room in the bitfield section of
object_entry.
This lets us limit the change primarily to the oe_delta()
and oe_set_delta_ext() functions. And as a bonus, most of
the rest of the code does not consider these dummy entries
at all, saving both runtime CPU and code complexity.
Signed-off-by: Jeff King <peff@peff.net>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
367 lines
10 KiB
C
367 lines
10 KiB
C
#ifndef PACK_OBJECTS_H
|
|
#define PACK_OBJECTS_H
|
|
|
|
#include "object-store.h"
|
|
#include "pack.h"
|
|
|
|
#define DEFAULT_DELTA_CACHE_SIZE (256 * 1024 * 1024)
|
|
|
|
#define OE_DFS_STATE_BITS 2
|
|
#define OE_DEPTH_BITS 12
|
|
#define OE_IN_PACK_BITS 10
|
|
#define OE_Z_DELTA_BITS 20
|
|
/*
|
|
* Note that oe_set_size() becomes expensive when the given size is
|
|
* above this limit. Don't lower it too much.
|
|
*/
|
|
#define OE_SIZE_BITS 31
|
|
#define OE_DELTA_SIZE_BITS 20
|
|
|
|
/*
|
|
* State flags for depth-first search used for analyzing delta cycles.
|
|
*
|
|
* The depth is measured in delta-links to the base (so if A is a delta
|
|
* against B, then A has a depth of 1, and B a depth of 0).
|
|
*/
|
|
enum dfs_state {
|
|
DFS_NONE = 0,
|
|
DFS_ACTIVE,
|
|
DFS_DONE,
|
|
DFS_NUM_STATES
|
|
};
|
|
|
|
/*
|
|
* The size of struct nearly determines pack-objects's memory
|
|
* consumption. This struct is packed tight for that reason. When you
|
|
* add or reorder something in this struct, think a bit about this.
|
|
*
|
|
* basic object info
|
|
* -----------------
|
|
* idx.oid is filled up before delta searching starts. idx.crc32 is
|
|
* only valid after the object is written out and will be used for
|
|
* generating the index. idx.offset will be both gradually set and
|
|
* used in writing phase (base objects get offset first, then deltas
|
|
* refer to them)
|
|
*
|
|
* "size" is the uncompressed object size. Compressed size of the raw
|
|
* data for an object in a pack is not stored anywhere but is computed
|
|
* and made available when reverse .idx is made. Note that when a
|
|
* delta is reused, "size" is the uncompressed _delta_ size, not the
|
|
* canonical one after the delta has been applied.
|
|
*
|
|
* "hash" contains a path name hash which is used for sorting the
|
|
* delta list and also during delta searching. Once prepare_pack()
|
|
* returns it's no longer needed.
|
|
*
|
|
* source pack info
|
|
* ----------------
|
|
* The (in_pack, in_pack_offset) tuple contains the location of the
|
|
* object in the source pack. in_pack_header_size allows quickly
|
|
* skipping the header and going straight to the zlib stream.
|
|
*
|
|
* "type" and "in_pack_type" both describe object type. in_pack_type
|
|
* may contain a delta type, while type is always the canonical type.
|
|
*
|
|
* deltas
|
|
* ------
|
|
* Delta links (delta, delta_child and delta_sibling) are created to
|
|
* reflect that delta graph from the source pack then updated or added
|
|
* during delta searching phase when we find better deltas.
|
|
*
|
|
* delta_child and delta_sibling are last needed in
|
|
* compute_write_order(). "delta" and "delta_size" must remain valid
|
|
* at object writing phase in case the delta is not cached.
|
|
*
|
|
* If a delta is cached in memory and is compressed, delta_data points
|
|
* to the data and z_delta_size contains the compressed size. If it's
|
|
* uncompressed [1], z_delta_size must be zero. delta_size is always
|
|
* the uncompressed size and must be valid even if the delta is not
|
|
* cached.
|
|
*
|
|
* [1] during try_delta phase we don't bother with compressing because
|
|
* the delta could be quickly replaced with a better one.
|
|
*/
|
|
struct object_entry {
|
|
struct pack_idx_entry idx;
|
|
void *delta_data; /* cached delta (uncompressed) */
|
|
off_t in_pack_offset;
|
|
uint32_t hash; /* name hint hash */
|
|
unsigned size_:OE_SIZE_BITS;
|
|
unsigned size_valid:1;
|
|
uint32_t delta_idx; /* delta base object */
|
|
uint32_t delta_child_idx; /* deltified objects who bases me */
|
|
uint32_t delta_sibling_idx; /* other deltified objects who
|
|
* uses the same base as me
|
|
*/
|
|
unsigned delta_size_:OE_DELTA_SIZE_BITS; /* delta data size (uncompressed) */
|
|
unsigned delta_size_valid:1;
|
|
unsigned in_pack_idx:OE_IN_PACK_BITS; /* already in pack */
|
|
unsigned z_delta_size:OE_Z_DELTA_BITS;
|
|
unsigned type_valid:1;
|
|
unsigned type_:TYPE_BITS;
|
|
unsigned no_try_delta:1;
|
|
unsigned in_pack_type:TYPE_BITS; /* could be delta */
|
|
unsigned preferred_base:1; /*
|
|
* we do not pack this, but is available
|
|
* to be used as the base object to delta
|
|
* objects against.
|
|
*/
|
|
unsigned tagged:1; /* near the very tip of refs */
|
|
unsigned filled:1; /* assigned write-order */
|
|
unsigned dfs_state:OE_DFS_STATE_BITS;
|
|
unsigned char in_pack_header_size;
|
|
unsigned depth:OE_DEPTH_BITS;
|
|
unsigned ext_base:1; /* delta_idx points outside packlist */
|
|
|
|
/*
|
|
* pahole results on 64-bit linux (gcc and clang)
|
|
*
|
|
* size: 80, bit_padding: 20 bits, holes: 8 bits
|
|
*
|
|
* and on 32-bit (gcc)
|
|
*
|
|
* size: 76, bit_padding: 20 bits, holes: 8 bits
|
|
*/
|
|
};
|
|
|
|
struct packing_data {
|
|
struct object_entry *objects;
|
|
uint32_t nr_objects, nr_alloc;
|
|
|
|
int32_t *index;
|
|
uint32_t index_size;
|
|
|
|
unsigned int *in_pack_pos;
|
|
|
|
/*
|
|
* Only one of these can be non-NULL and they have different
|
|
* sizes. if in_pack_by_idx is allocated, oe_in_pack() returns
|
|
* the pack of an object using in_pack_idx field. If not,
|
|
* in_pack[] array is used the same way as in_pack_pos[]
|
|
*/
|
|
struct packed_git **in_pack_by_idx;
|
|
struct packed_git **in_pack;
|
|
|
|
/*
|
|
* This list contains entries for bases which we know the other side
|
|
* has (e.g., via reachability bitmaps), but which aren't in our
|
|
* "objects" list.
|
|
*/
|
|
struct object_entry *ext_bases;
|
|
uint32_t nr_ext, alloc_ext;
|
|
|
|
uintmax_t oe_size_limit;
|
|
};
|
|
|
|
void prepare_packing_data(struct packing_data *pdata);
|
|
struct object_entry *packlist_alloc(struct packing_data *pdata,
|
|
const unsigned char *sha1,
|
|
uint32_t index_pos);
|
|
|
|
struct object_entry *packlist_find(struct packing_data *pdata,
|
|
const unsigned char *sha1,
|
|
uint32_t *index_pos);
|
|
|
|
static inline uint32_t pack_name_hash(const char *name)
|
|
{
|
|
uint32_t c, hash = 0;
|
|
|
|
if (!name)
|
|
return 0;
|
|
|
|
/*
|
|
* This effectively just creates a sortable number from the
|
|
* last sixteen non-whitespace characters. Last characters
|
|
* count "most", so things that end in ".c" sort together.
|
|
*/
|
|
while ((c = *name++) != 0) {
|
|
if (isspace(c))
|
|
continue;
|
|
hash = (hash >> 2) + (c << 24);
|
|
}
|
|
return hash;
|
|
}
|
|
|
|
static inline enum object_type oe_type(const struct object_entry *e)
|
|
{
|
|
return e->type_valid ? e->type_ : OBJ_BAD;
|
|
}
|
|
|
|
static inline void oe_set_type(struct object_entry *e,
|
|
enum object_type type)
|
|
{
|
|
if (type >= OBJ_ANY)
|
|
BUG("OBJ_ANY cannot be set in pack-objects code");
|
|
|
|
e->type_valid = type >= OBJ_NONE;
|
|
e->type_ = (unsigned)type;
|
|
}
|
|
|
|
static inline unsigned int oe_in_pack_pos(const struct packing_data *pack,
|
|
const struct object_entry *e)
|
|
{
|
|
return pack->in_pack_pos[e - pack->objects];
|
|
}
|
|
|
|
static inline void oe_set_in_pack_pos(const struct packing_data *pack,
|
|
const struct object_entry *e,
|
|
unsigned int pos)
|
|
{
|
|
pack->in_pack_pos[e - pack->objects] = pos;
|
|
}
|
|
|
|
static inline struct packed_git *oe_in_pack(const struct packing_data *pack,
|
|
const struct object_entry *e)
|
|
{
|
|
if (pack->in_pack_by_idx)
|
|
return pack->in_pack_by_idx[e->in_pack_idx];
|
|
else
|
|
return pack->in_pack[e - pack->objects];
|
|
}
|
|
|
|
void oe_map_new_pack(struct packing_data *pack,
|
|
struct packed_git *p);
|
|
static inline void oe_set_in_pack(struct packing_data *pack,
|
|
struct object_entry *e,
|
|
struct packed_git *p)
|
|
{
|
|
if (!p->index)
|
|
oe_map_new_pack(pack, p);
|
|
if (pack->in_pack_by_idx)
|
|
e->in_pack_idx = p->index;
|
|
else
|
|
pack->in_pack[e - pack->objects] = p;
|
|
}
|
|
|
|
static inline struct object_entry *oe_delta(
|
|
const struct packing_data *pack,
|
|
const struct object_entry *e)
|
|
{
|
|
if (!e->delta_idx)
|
|
return NULL;
|
|
if (e->ext_base)
|
|
return &pack->ext_bases[e->delta_idx - 1];
|
|
else
|
|
return &pack->objects[e->delta_idx - 1];
|
|
}
|
|
|
|
static inline void oe_set_delta(struct packing_data *pack,
|
|
struct object_entry *e,
|
|
struct object_entry *delta)
|
|
{
|
|
if (delta)
|
|
e->delta_idx = (delta - pack->objects) + 1;
|
|
else
|
|
e->delta_idx = 0;
|
|
}
|
|
|
|
void oe_set_delta_ext(struct packing_data *pack,
|
|
struct object_entry *e,
|
|
const unsigned char *sha1);
|
|
|
|
static inline struct object_entry *oe_delta_child(
|
|
const struct packing_data *pack,
|
|
const struct object_entry *e)
|
|
{
|
|
if (e->delta_child_idx)
|
|
return &pack->objects[e->delta_child_idx - 1];
|
|
return NULL;
|
|
}
|
|
|
|
static inline void oe_set_delta_child(struct packing_data *pack,
|
|
struct object_entry *e,
|
|
struct object_entry *delta)
|
|
{
|
|
if (delta)
|
|
e->delta_child_idx = (delta - pack->objects) + 1;
|
|
else
|
|
e->delta_child_idx = 0;
|
|
}
|
|
|
|
static inline struct object_entry *oe_delta_sibling(
|
|
const struct packing_data *pack,
|
|
const struct object_entry *e)
|
|
{
|
|
if (e->delta_sibling_idx)
|
|
return &pack->objects[e->delta_sibling_idx - 1];
|
|
return NULL;
|
|
}
|
|
|
|
static inline void oe_set_delta_sibling(struct packing_data *pack,
|
|
struct object_entry *e,
|
|
struct object_entry *delta)
|
|
{
|
|
if (delta)
|
|
e->delta_sibling_idx = (delta - pack->objects) + 1;
|
|
else
|
|
e->delta_sibling_idx = 0;
|
|
}
|
|
|
|
unsigned long oe_get_size_slow(struct packing_data *pack,
|
|
const struct object_entry *e);
|
|
static inline unsigned long oe_size(struct packing_data *pack,
|
|
const struct object_entry *e)
|
|
{
|
|
if (e->size_valid)
|
|
return e->size_;
|
|
|
|
return oe_get_size_slow(pack, e);
|
|
}
|
|
|
|
static inline int oe_size_less_than(struct packing_data *pack,
|
|
const struct object_entry *lhs,
|
|
unsigned long rhs)
|
|
{
|
|
if (lhs->size_valid)
|
|
return lhs->size_ < rhs;
|
|
if (rhs < pack->oe_size_limit) /* rhs < 2^x <= lhs ? */
|
|
return 0;
|
|
return oe_get_size_slow(pack, lhs) < rhs;
|
|
}
|
|
|
|
static inline int oe_size_greater_than(struct packing_data *pack,
|
|
const struct object_entry *lhs,
|
|
unsigned long rhs)
|
|
{
|
|
if (lhs->size_valid)
|
|
return lhs->size_ > rhs;
|
|
if (rhs < pack->oe_size_limit) /* rhs < 2^x <= lhs ? */
|
|
return 1;
|
|
return oe_get_size_slow(pack, lhs) > rhs;
|
|
}
|
|
|
|
static inline void oe_set_size(struct packing_data *pack,
|
|
struct object_entry *e,
|
|
unsigned long size)
|
|
{
|
|
if (size < pack->oe_size_limit) {
|
|
e->size_ = size;
|
|
e->size_valid = 1;
|
|
} else {
|
|
e->size_valid = 0;
|
|
if (oe_get_size_slow(pack, e) != size)
|
|
BUG("'size' is supposed to be the object size!");
|
|
}
|
|
}
|
|
|
|
static inline unsigned long oe_delta_size(struct packing_data *pack,
|
|
const struct object_entry *e)
|
|
{
|
|
if (e->delta_size_valid)
|
|
return e->delta_size_;
|
|
return oe_size(pack, e);
|
|
}
|
|
|
|
static inline void oe_set_delta_size(struct packing_data *pack,
|
|
struct object_entry *e,
|
|
unsigned long size)
|
|
{
|
|
e->delta_size_ = size;
|
|
e->delta_size_valid = e->delta_size_ == size;
|
|
if (!e->delta_size_valid && size != oe_size(pack, e))
|
|
BUG("this can only happen in check_object() "
|
|
"where delta size is the same as entry size");
|
|
}
|
|
|
|
#endif
|