Skip to content

Commit

Permalink
midx: add pack_perm to write_midx_context
Browse files Browse the repository at this point in the history
In an effort to align write_midx_internal() with the chunk-format API,
continue to group necessary data into "struct write_midx_context". This
change collects the "uint32_t *pack_perm" and large_offsets_needed bit
into the context.

Update write_midx_object_offsets() to match chunk_write_fn.

Signed-off-by: Derrick Stolee <[email protected]>
Signed-off-by: Junio C Hamano <[email protected]>
  • Loading branch information
derrickstolee authored and gitster committed Feb 18, 2021
1 parent 31bda9a commit 7a3ada1
Showing 1 changed file with 21 additions and 19 deletions.
40 changes: 21 additions & 19 deletions midx.c
Original file line number Diff line number Diff line change
Expand Up @@ -461,6 +461,9 @@ struct write_midx_context {

struct pack_midx_entry *entries;
uint32_t entries_nr;

uint32_t *pack_perm;
unsigned large_offsets_needed:1;
};

static void add_pack_to_midx(const char *full_path, size_t full_path_len,
Expand Down Expand Up @@ -736,27 +739,27 @@ static size_t write_midx_oid_lookup(struct hashfile *f,
return written;
}

static size_t write_midx_object_offsets(struct hashfile *f, int large_offset_needed,
uint32_t *perm,
struct pack_midx_entry *objects, uint32_t nr_objects)
static size_t write_midx_object_offsets(struct hashfile *f,
void *data)
{
struct pack_midx_entry *list = objects;
struct write_midx_context *ctx = data;
struct pack_midx_entry *list = ctx->entries;
uint32_t i, nr_large_offset = 0;
size_t written = 0;

for (i = 0; i < nr_objects; i++) {
for (i = 0; i < ctx->entries_nr; i++) {
struct pack_midx_entry *obj = list++;

if (perm[obj->pack_int_id] == PACK_EXPIRED)
if (ctx->pack_perm[obj->pack_int_id] == PACK_EXPIRED)
BUG("object %s is in an expired pack with int-id %d",
oid_to_hex(&obj->oid),
obj->pack_int_id);

hashwrite_be32(f, perm[obj->pack_int_id]);
hashwrite_be32(f, ctx->pack_perm[obj->pack_int_id]);

if (large_offset_needed && obj->offset >> 31)
if (ctx->large_offsets_needed && obj->offset >> 31)
hashwrite_be32(f, MIDX_LARGE_OFFSET_NEEDED | nr_large_offset++);
else if (!large_offset_needed && obj->offset >> 32)
else if (!ctx->large_offsets_needed && obj->offset >> 32)
BUG("object %s requires a large offset (%"PRIx64") but the MIDX is not writing large offsets!",
oid_to_hex(&obj->oid),
obj->offset);
Expand Down Expand Up @@ -805,13 +808,11 @@ static int write_midx_internal(const char *object_dir, struct multi_pack_index *
struct hashfile *f = NULL;
struct lock_file lk;
struct write_midx_context ctx = { 0 };
uint32_t *pack_perm = NULL;
uint64_t written = 0;
uint32_t chunk_ids[MIDX_MAX_CHUNKS + 1];
uint64_t chunk_offsets[MIDX_MAX_CHUNKS + 1];
uint32_t num_large_offsets = 0;
struct progress *progress = NULL;
int large_offsets_needed = 0;
int pack_name_concat_len = 0;
int dropped_packs = 0;
int result = 0;
Expand Down Expand Up @@ -857,11 +858,12 @@ static int write_midx_internal(const char *object_dir, struct multi_pack_index *

ctx.entries = get_sorted_entries(ctx.m, ctx.info, ctx.nr, &ctx.entries_nr);

ctx.large_offsets_needed = 0;
for (i = 0; i < ctx.entries_nr; i++) {
if (ctx.entries[i].offset > 0x7fffffff)
num_large_offsets++;
if (ctx.entries[i].offset > 0xffffffff)
large_offsets_needed = 1;
ctx.large_offsets_needed = 1;
}

QSORT(ctx.info, ctx.nr, pack_info_compare);
Expand Down Expand Up @@ -900,13 +902,13 @@ static int write_midx_internal(const char *object_dir, struct multi_pack_index *
*
* pack_perm[old_id] = new_id
*/
ALLOC_ARRAY(pack_perm, ctx.nr);
ALLOC_ARRAY(ctx.pack_perm, ctx.nr);
for (i = 0; i < ctx.nr; i++) {
if (ctx.info[i].expired) {
dropped_packs++;
pack_perm[ctx.info[i].orig_pack_int_id] = PACK_EXPIRED;
ctx.pack_perm[ctx.info[i].orig_pack_int_id] = PACK_EXPIRED;
} else {
pack_perm[ctx.info[i].orig_pack_int_id] = i - dropped_packs;
ctx.pack_perm[ctx.info[i].orig_pack_int_id] = i - dropped_packs;
}
}

Expand All @@ -927,7 +929,7 @@ static int write_midx_internal(const char *object_dir, struct multi_pack_index *
close_midx(ctx.m);

cur_chunk = 0;
num_chunks = large_offsets_needed ? 5 : 4;
num_chunks = ctx.large_offsets_needed ? 5 : 4;

if (ctx.nr - dropped_packs == 0) {
error(_("no pack files to index."));
Expand All @@ -954,7 +956,7 @@ static int write_midx_internal(const char *object_dir, struct multi_pack_index *

cur_chunk++;
chunk_offsets[cur_chunk] = chunk_offsets[cur_chunk - 1] + ctx.entries_nr * MIDX_CHUNK_OFFSET_WIDTH;
if (large_offsets_needed) {
if (ctx.large_offsets_needed) {
chunk_ids[cur_chunk] = MIDX_CHUNKID_LARGEOFFSETS;

cur_chunk++;
Expand Down Expand Up @@ -1004,7 +1006,7 @@ static int write_midx_internal(const char *object_dir, struct multi_pack_index *
break;

case MIDX_CHUNKID_OBJECTOFFSETS:
written += write_midx_object_offsets(f, large_offsets_needed, pack_perm, ctx.entries, ctx.entries_nr);
written += write_midx_object_offsets(f, &ctx);
break;

case MIDX_CHUNKID_LARGEOFFSETS:
Expand Down Expand Up @@ -1039,7 +1041,7 @@ static int write_midx_internal(const char *object_dir, struct multi_pack_index *

free(ctx.info);
free(ctx.entries);
free(pack_perm);
free(ctx.pack_perm);
free(midx_name);
return result;
}
Expand Down

0 comments on commit 7a3ada1

Please sign in to comment.