248 lines
		
	
	
		
			7.0 KiB
		
	
	
	
		
			C
		
	
	
			
		
		
	
	
			248 lines
		
	
	
		
			7.0 KiB
		
	
	
	
		
			C
		
	
	
| #include "cache.h"
 | |
| #include "pack-revindex.h"
 | |
| 
 | |
| /*
 | |
|  * Pack index for existing packs give us easy access to the offsets into
 | |
|  * corresponding pack file where each object's data starts, but the entries
 | |
|  * do not store the size of the compressed representation (uncompressed
 | |
|  * size is easily available by examining the pack entry header).  It is
 | |
|  * also rather expensive to find the sha1 for an object given its offset.
 | |
|  *
 | |
|  * We build a hashtable of existing packs (pack_revindex), and keep reverse
 | |
|  * index here -- pack index file is sorted by object name mapping to offset;
 | |
|  * this pack_revindex[].revindex array is a list of offset/index_nr pairs
 | |
|  * ordered by offset, so if you know the offset of an object, next offset
 | |
|  * is where its packed representation ends and the index_nr can be used to
 | |
|  * get the object sha1 from the main index.
 | |
|  */
 | |
| 
 | |
| static struct pack_revindex *pack_revindex;
 | |
| static int pack_revindex_hashsz;
 | |
| 
 | |
| static int pack_revindex_ix(struct packed_git *p)
 | |
| {
 | |
| 	unsigned long ui = (unsigned long)p;
 | |
| 	int i;
 | |
| 
 | |
| 	ui = ui ^ (ui >> 16); /* defeat structure alignment */
 | |
| 	i = (int)(ui % pack_revindex_hashsz);
 | |
| 	while (pack_revindex[i].p) {
 | |
| 		if (pack_revindex[i].p == p)
 | |
| 			return i;
 | |
| 		if (++i == pack_revindex_hashsz)
 | |
| 			i = 0;
 | |
| 	}
 | |
| 	return -1 - i;
 | |
| }
 | |
| 
 | |
| static void init_pack_revindex(void)
 | |
| {
 | |
| 	int num;
 | |
| 	struct packed_git *p;
 | |
| 
 | |
| 	for (num = 0, p = packed_git; p; p = p->next)
 | |
| 		num++;
 | |
| 	if (!num)
 | |
| 		return;
 | |
| 	pack_revindex_hashsz = num * 11;
 | |
| 	pack_revindex = xcalloc(pack_revindex_hashsz, sizeof(*pack_revindex));
 | |
| 	for (p = packed_git; p; p = p->next) {
 | |
| 		num = pack_revindex_ix(p);
 | |
| 		num = - 1 - num;
 | |
| 		pack_revindex[num].p = p;
 | |
| 	}
 | |
| 	/* revindex elements are lazily initialized */
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * This is a least-significant-digit radix sort.
 | |
|  *
 | |
|  * It sorts each of the "n" items in "entries" by its offset field. The "max"
 | |
|  * parameter must be at least as large as the largest offset in the array,
 | |
|  * and lets us quit the sort early.
 | |
|  */
 | |
| static void sort_revindex(struct revindex_entry *entries, unsigned n, off_t max)
 | |
| {
 | |
| 	/*
 | |
| 	 * We use a "digit" size of 16 bits. That keeps our memory
 | |
| 	 * usage reasonable, and we can generally (for a 4G or smaller
 | |
| 	 * packfile) quit after two rounds of radix-sorting.
 | |
| 	 */
 | |
| #define DIGIT_SIZE (16)
 | |
| #define BUCKETS (1 << DIGIT_SIZE)
 | |
| 	/*
 | |
| 	 * We want to know the bucket that a[i] will go into when we are using
 | |
| 	 * the digit that is N bits from the (least significant) end.
 | |
| 	 */
 | |
| #define BUCKET_FOR(a, i, bits) (((a)[(i)].offset >> (bits)) & (BUCKETS-1))
 | |
| 
 | |
| 	/*
 | |
| 	 * We need O(n) temporary storage. Rather than do an extra copy of the
 | |
| 	 * partial results into "entries", we sort back and forth between the
 | |
| 	 * real array and temporary storage. In each iteration of the loop, we
 | |
| 	 * keep track of them with alias pointers, always sorting from "from"
 | |
| 	 * to "to".
 | |
| 	 */
 | |
| 	struct revindex_entry *tmp = xmalloc(n * sizeof(*tmp));
 | |
| 	struct revindex_entry *from = entries, *to = tmp;
 | |
| 	int bits;
 | |
| 	unsigned *pos = xmalloc(BUCKETS * sizeof(*pos));
 | |
| 
 | |
| 	/*
 | |
| 	 * If (max >> bits) is zero, then we know that the radix digit we are
 | |
| 	 * on (and any higher) will be zero for all entries, and our loop will
 | |
| 	 * be a no-op, as everybody lands in the same zero-th bucket.
 | |
| 	 */
 | |
| 	for (bits = 0; max >> bits; bits += DIGIT_SIZE) {
 | |
| 		struct revindex_entry *swap;
 | |
| 		unsigned i;
 | |
| 
 | |
| 		memset(pos, 0, BUCKETS * sizeof(*pos));
 | |
| 
 | |
| 		/*
 | |
| 		 * We want pos[i] to store the index of the last element that
 | |
| 		 * will go in bucket "i" (actually one past the last element).
 | |
| 		 * To do this, we first count the items that will go in each
 | |
| 		 * bucket, which gives us a relative offset from the last
 | |
| 		 * bucket. We can then cumulatively add the index from the
 | |
| 		 * previous bucket to get the true index.
 | |
| 		 */
 | |
| 		for (i = 0; i < n; i++)
 | |
| 			pos[BUCKET_FOR(from, i, bits)]++;
 | |
| 		for (i = 1; i < BUCKETS; i++)
 | |
| 			pos[i] += pos[i-1];
 | |
| 
 | |
| 		/*
 | |
| 		 * Now we can drop the elements into their correct buckets (in
 | |
| 		 * our temporary array).  We iterate the pos counter backwards
 | |
| 		 * to avoid using an extra index to count up. And since we are
 | |
| 		 * going backwards there, we must also go backwards through the
 | |
| 		 * array itself, to keep the sort stable.
 | |
| 		 *
 | |
| 		 * Note that we use an unsigned iterator to make sure we can
 | |
| 		 * handle 2^32-1 objects, even on a 32-bit system. But this
 | |
| 		 * means we cannot use the more obvious "i >= 0" loop condition
 | |
| 		 * for counting backwards, and must instead check for
 | |
| 		 * wrap-around with UINT_MAX.
 | |
| 		 */
 | |
| 		for (i = n - 1; i != UINT_MAX; i--)
 | |
| 			to[--pos[BUCKET_FOR(from, i, bits)]] = from[i];
 | |
| 
 | |
| 		/*
 | |
| 		 * Now "to" contains the most sorted list, so we swap "from" and
 | |
| 		 * "to" for the next iteration.
 | |
| 		 */
 | |
| 		swap = from;
 | |
| 		from = to;
 | |
| 		to = swap;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * If we ended with our data in the original array, great. If not,
 | |
| 	 * we have to move it back from the temporary storage.
 | |
| 	 */
 | |
| 	if (from != entries)
 | |
| 		memcpy(entries, tmp, n * sizeof(*entries));
 | |
| 	free(tmp);
 | |
| 	free(pos);
 | |
| 
 | |
| #undef BUCKET_FOR
 | |
| #undef BUCKETS
 | |
| #undef DIGIT_SIZE
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Ordered list of offsets of objects in the pack.
 | |
|  */
 | |
| static void create_pack_revindex(struct pack_revindex *rix)
 | |
| {
 | |
| 	struct packed_git *p = rix->p;
 | |
| 	unsigned num_ent = p->num_objects;
 | |
| 	unsigned i;
 | |
| 	const char *index = p->index_data;
 | |
| 
 | |
| 	rix->revindex = xmalloc(sizeof(*rix->revindex) * (num_ent + 1));
 | |
| 	index += 4 * 256;
 | |
| 
 | |
| 	if (p->index_version > 1) {
 | |
| 		const uint32_t *off_32 =
 | |
| 			(uint32_t *)(index + 8 + p->num_objects * (20 + 4));
 | |
| 		const uint32_t *off_64 = off_32 + p->num_objects;
 | |
| 		for (i = 0; i < num_ent; i++) {
 | |
| 			uint32_t off = ntohl(*off_32++);
 | |
| 			if (!(off & 0x80000000)) {
 | |
| 				rix->revindex[i].offset = off;
 | |
| 			} else {
 | |
| 				rix->revindex[i].offset =
 | |
| 					((uint64_t)ntohl(*off_64++)) << 32;
 | |
| 				rix->revindex[i].offset |=
 | |
| 					ntohl(*off_64++);
 | |
| 			}
 | |
| 			rix->revindex[i].nr = i;
 | |
| 		}
 | |
| 	} else {
 | |
| 		for (i = 0; i < num_ent; i++) {
 | |
| 			uint32_t hl = *((uint32_t *)(index + 24 * i));
 | |
| 			rix->revindex[i].offset = ntohl(hl);
 | |
| 			rix->revindex[i].nr = i;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	/* This knows the pack format -- the 20-byte trailer
 | |
| 	 * follows immediately after the last object data.
 | |
| 	 */
 | |
| 	rix->revindex[num_ent].offset = p->pack_size - 20;
 | |
| 	rix->revindex[num_ent].nr = -1;
 | |
| 	sort_revindex(rix->revindex, num_ent, p->pack_size);
 | |
| }
 | |
| 
 | |
| struct pack_revindex *revindex_for_pack(struct packed_git *p)
 | |
| {
 | |
| 	int num;
 | |
| 	struct pack_revindex *rix;
 | |
| 
 | |
| 	if (!pack_revindex_hashsz)
 | |
| 		init_pack_revindex();
 | |
| 
 | |
| 	num = pack_revindex_ix(p);
 | |
| 	if (num < 0)
 | |
| 		die("internal error: pack revindex fubar");
 | |
| 
 | |
| 	rix = &pack_revindex[num];
 | |
| 	if (!rix->revindex)
 | |
| 		create_pack_revindex(rix);
 | |
| 
 | |
| 	return rix;
 | |
| }
 | |
| 
 | |
| int find_revindex_position(struct pack_revindex *pridx, off_t ofs)
 | |
| {
 | |
| 	int lo = 0;
 | |
| 	int hi = pridx->p->num_objects + 1;
 | |
| 	struct revindex_entry *revindex = pridx->revindex;
 | |
| 
 | |
| 	do {
 | |
| 		unsigned mi = lo + (hi - lo) / 2;
 | |
| 		if (revindex[mi].offset == ofs) {
 | |
| 			return mi;
 | |
| 		} else if (ofs < revindex[mi].offset)
 | |
| 			hi = mi;
 | |
| 		else
 | |
| 			lo = mi + 1;
 | |
| 	} while (lo < hi);
 | |
| 
 | |
| 	error("bad offset for revindex");
 | |
| 	return -1;
 | |
| }
 | |
| 
 | |
| struct revindex_entry *find_pack_revindex(struct packed_git *p, off_t ofs)
 | |
| {
 | |
| 	struct pack_revindex *pridx = revindex_for_pack(p);
 | |
| 	int pos = find_revindex_position(pridx, ofs);
 | |
| 
 | |
| 	if (pos < 0)
 | |
| 		return NULL;
 | |
| 
 | |
| 	return pridx->revindex + pos;
 | |
| }
 |