|  | #define USE_THE_REPOSITORY_VARIABLE | 
|  | #define DISABLE_SIGN_COMPARE_WARNINGS | 
|  |  | 
|  | #include "git-compat-util.h" | 
|  | #include "hex.h" | 
|  | #include "repository.h" | 
|  | #include "tempfile.h" | 
|  | #include "lockfile.h" | 
|  | #include "object-store-ll.h" | 
|  | #include "commit.h" | 
|  | #include "tag.h" | 
|  | #include "pkt-line.h" | 
|  | #include "refs.h" | 
|  | #include "oid-array.h" | 
|  | #include "path.h" | 
|  | #include "diff.h" | 
|  | #include "revision.h" | 
|  | #include "commit-slab.h" | 
|  | #include "list-objects.h" | 
|  | #include "commit-reach.h" | 
|  | #include "shallow.h" | 
|  | #include "statinfo.h" | 
|  | #include "trace.h" | 
|  |  | 
|  | void set_alternate_shallow_file(struct repository *r, const char *path, int override) | 
|  | { | 
|  | if (r->parsed_objects->is_shallow != -1) | 
|  | BUG("is_repository_shallow must not be called before set_alternate_shallow_file"); | 
|  | if (r->parsed_objects->alternate_shallow_file && !override) | 
|  | return; | 
|  | free(r->parsed_objects->alternate_shallow_file); | 
|  | r->parsed_objects->alternate_shallow_file = xstrdup_or_null(path); | 
|  | } | 
|  |  | 
|  | int register_shallow(struct repository *r, const struct object_id *oid) | 
|  | { | 
|  | struct commit_graft *graft = | 
|  | xmalloc(sizeof(struct commit_graft)); | 
|  | struct commit *commit = lookup_commit(r, oid); | 
|  |  | 
|  | oidcpy(&graft->oid, oid); | 
|  | graft->nr_parent = -1; | 
|  | if (commit && commit->object.parsed) { | 
|  | free_commit_list(commit->parents); | 
|  | commit->parents = NULL; | 
|  | } | 
|  | return register_commit_graft(r, graft, 0); | 
|  | } | 
|  |  | 
|  | int unregister_shallow(const struct object_id *oid) | 
|  | { | 
|  | int pos = commit_graft_pos(the_repository, oid); | 
|  | if (pos < 0) | 
|  | return -1; | 
|  | free(the_repository->parsed_objects->grafts[pos]); | 
|  | if (pos + 1 < the_repository->parsed_objects->grafts_nr) | 
|  | MOVE_ARRAY(the_repository->parsed_objects->grafts + pos, | 
|  | the_repository->parsed_objects->grafts + pos + 1, | 
|  | the_repository->parsed_objects->grafts_nr - pos - 1); | 
|  | the_repository->parsed_objects->grafts_nr--; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | int is_repository_shallow(struct repository *r) | 
|  | { | 
|  | FILE *fp; | 
|  | char buf[1024]; | 
|  | const char *path = r->parsed_objects->alternate_shallow_file; | 
|  |  | 
|  | if (r->parsed_objects->is_shallow >= 0) | 
|  | return r->parsed_objects->is_shallow; | 
|  |  | 
|  | if (!path) | 
|  | path = git_path_shallow(r); | 
|  | /* | 
|  | * fetch-pack sets '--shallow-file ""' as an indicator that no | 
|  | * shallow file should be used. We could just open it and it | 
|  | * will likely fail. But let's do an explicit check instead. | 
|  | */ | 
|  | if (!*path || (fp = fopen(path, "r")) == NULL) { | 
|  | stat_validity_clear(r->parsed_objects->shallow_stat); | 
|  | r->parsed_objects->is_shallow = 0; | 
|  | return r->parsed_objects->is_shallow; | 
|  | } | 
|  | stat_validity_update(r->parsed_objects->shallow_stat, fileno(fp)); | 
|  | r->parsed_objects->is_shallow = 1; | 
|  |  | 
|  | while (fgets(buf, sizeof(buf), fp)) { | 
|  | struct object_id oid; | 
|  | if (get_oid_hex(buf, &oid)) | 
|  | die("bad shallow line: %s", buf); | 
|  | register_shallow(r, &oid); | 
|  | } | 
|  | fclose(fp); | 
|  | return r->parsed_objects->is_shallow; | 
|  | } | 
|  |  | 
|  | static void reset_repository_shallow(struct repository *r) | 
|  | { | 
|  | r->parsed_objects->is_shallow = -1; | 
|  | stat_validity_clear(r->parsed_objects->shallow_stat); | 
|  | parsed_object_pool_reset_commit_grafts(r->parsed_objects); | 
|  | } | 
|  |  | 
|  | int commit_shallow_file(struct repository *r, struct shallow_lock *lk) | 
|  | { | 
|  | int res = commit_lock_file(&lk->lock); | 
|  | reset_repository_shallow(r); | 
|  |  | 
|  | /* | 
|  | * Update in-memory data structures with the new shallow information, | 
|  | * including unparsing all commits that now have grafts. | 
|  | */ | 
|  | is_repository_shallow(r); | 
|  |  | 
|  | return res; | 
|  | } | 
|  |  | 
|  | void rollback_shallow_file(struct repository *r, struct shallow_lock *lk) | 
|  | { | 
|  | rollback_lock_file(&lk->lock); | 
|  | reset_repository_shallow(r); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * TODO: use "int" elemtype instead of "int *" when/if commit-slab | 
|  | * supports a "valid" flag. | 
|  | */ | 
|  | define_commit_slab(commit_depth, int *); | 
|  | static void free_depth_in_slab(int **ptr) | 
|  | { | 
|  | FREE_AND_NULL(*ptr); | 
|  | } | 
|  | struct commit_list *get_shallow_commits(struct object_array *heads, int depth, | 
|  | int shallow_flag, int not_shallow_flag) | 
|  | { | 
|  | int i = 0, cur_depth = 0; | 
|  | struct commit_list *result = NULL; | 
|  | struct object_array stack = OBJECT_ARRAY_INIT; | 
|  | struct commit *commit = NULL; | 
|  | struct commit_graft *graft; | 
|  | struct commit_depth depths; | 
|  |  | 
|  | init_commit_depth(&depths); | 
|  | while (commit || i < heads->nr || stack.nr) { | 
|  | struct commit_list *p; | 
|  | if (!commit) { | 
|  | if (i < heads->nr) { | 
|  | int **depth_slot; | 
|  | commit = (struct commit *) | 
|  | deref_tag(the_repository, | 
|  | heads->objects[i++].item, | 
|  | NULL, 0); | 
|  | if (!commit || commit->object.type != OBJ_COMMIT) { | 
|  | commit = NULL; | 
|  | continue; | 
|  | } | 
|  | depth_slot = commit_depth_at(&depths, commit); | 
|  | if (!*depth_slot) | 
|  | *depth_slot = xmalloc(sizeof(int)); | 
|  | **depth_slot = 0; | 
|  | cur_depth = 0; | 
|  | } else { | 
|  | commit = (struct commit *) | 
|  | object_array_pop(&stack); | 
|  | cur_depth = **commit_depth_at(&depths, commit); | 
|  | } | 
|  | } | 
|  | parse_commit_or_die(commit); | 
|  | cur_depth++; | 
|  | if ((depth != INFINITE_DEPTH && cur_depth >= depth) || | 
|  | (is_repository_shallow(the_repository) && !commit->parents && | 
|  | (graft = lookup_commit_graft(the_repository, &commit->object.oid)) != NULL && | 
|  | graft->nr_parent < 0)) { | 
|  | commit_list_insert(commit, &result); | 
|  | commit->object.flags |= shallow_flag; | 
|  | commit = NULL; | 
|  | continue; | 
|  | } | 
|  | commit->object.flags |= not_shallow_flag; | 
|  | for (p = commit->parents, commit = NULL; p; p = p->next) { | 
|  | int **depth_slot = commit_depth_at(&depths, p->item); | 
|  | if (!*depth_slot) { | 
|  | *depth_slot = xmalloc(sizeof(int)); | 
|  | **depth_slot = cur_depth; | 
|  | } else { | 
|  | if (cur_depth >= **depth_slot) | 
|  | continue; | 
|  | **depth_slot = cur_depth; | 
|  | } | 
|  | if (p->next) | 
|  | add_object_array(&p->item->object, | 
|  | NULL, &stack); | 
|  | else { | 
|  | commit = p->item; | 
|  | cur_depth = **commit_depth_at(&depths, commit); | 
|  | } | 
|  | } | 
|  | } | 
|  | deep_clear_commit_depth(&depths, free_depth_in_slab); | 
|  |  | 
|  | return result; | 
|  | } | 
|  |  | 
|  | static void show_commit(struct commit *commit, void *data) | 
|  | { | 
|  | commit_list_insert(commit, data); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Given rev-list arguments, run rev-list. All reachable commits | 
|  | * except border ones are marked with not_shallow_flag. Border commits | 
|  | * are marked with shallow_flag. The list of border/shallow commits | 
|  | * are also returned. | 
|  | */ | 
|  | struct commit_list *get_shallow_commits_by_rev_list(int ac, const char **av, | 
|  | int shallow_flag, | 
|  | int not_shallow_flag) | 
|  | { | 
|  | struct commit_list *result = NULL, *p; | 
|  | struct commit_list *not_shallow_list = NULL; | 
|  | struct rev_info revs; | 
|  | int both_flags = shallow_flag | not_shallow_flag; | 
|  |  | 
|  | /* | 
|  | * SHALLOW (excluded) and NOT_SHALLOW (included) should not be | 
|  | * set at this point. But better be safe than sorry. | 
|  | */ | 
|  | clear_object_flags(both_flags); | 
|  |  | 
|  | is_repository_shallow(the_repository); /* make sure shallows are read */ | 
|  |  | 
|  | repo_init_revisions(the_repository, &revs, NULL); | 
|  | save_commit_buffer = 0; | 
|  | setup_revisions(ac, av, &revs, NULL); | 
|  |  | 
|  | if (prepare_revision_walk(&revs)) | 
|  | die("revision walk setup failed"); | 
|  | traverse_commit_list(&revs, show_commit, NULL, ¬_shallow_list); | 
|  |  | 
|  | if (!not_shallow_list) | 
|  | die("no commits selected for shallow requests"); | 
|  |  | 
|  | /* Mark all reachable commits as NOT_SHALLOW */ | 
|  | for (p = not_shallow_list; p; p = p->next) | 
|  | p->item->object.flags |= not_shallow_flag; | 
|  |  | 
|  | /* | 
|  | * mark border commits SHALLOW + NOT_SHALLOW. | 
|  | * We cannot clear NOT_SHALLOW right now. Imagine border | 
|  | * commit A is processed first, then commit B, whose parent is | 
|  | * A, later. If NOT_SHALLOW on A is cleared at step 1, B | 
|  | * itself is considered border at step 2, which is incorrect. | 
|  | */ | 
|  | for (p = not_shallow_list; p; p = p->next) { | 
|  | struct commit *c = p->item; | 
|  | struct commit_list *parent; | 
|  |  | 
|  | if (repo_parse_commit(the_repository, c)) | 
|  | die("unable to parse commit %s", | 
|  | oid_to_hex(&c->object.oid)); | 
|  |  | 
|  | for (parent = c->parents; parent; parent = parent->next) | 
|  | if (!(parent->item->object.flags & not_shallow_flag)) { | 
|  | c->object.flags |= shallow_flag; | 
|  | commit_list_insert(c, &result); | 
|  | break; | 
|  | } | 
|  | } | 
|  | free_commit_list(not_shallow_list); | 
|  |  | 
|  | /* | 
|  | * Now we can clean up NOT_SHALLOW on border commits. Having | 
|  | * both flags set can confuse the caller. | 
|  | */ | 
|  | for (p = result; p; p = p->next) { | 
|  | struct object *o = &p->item->object; | 
|  | if ((o->flags & both_flags) == both_flags) | 
|  | o->flags &= ~not_shallow_flag; | 
|  | } | 
|  | release_revisions(&revs); | 
|  | return result; | 
|  | } | 
|  |  | 
|  | static void check_shallow_file_for_update(struct repository *r) | 
|  | { | 
|  | if (r->parsed_objects->is_shallow == -1) | 
|  | BUG("shallow must be initialized by now"); | 
|  |  | 
|  | if (!stat_validity_check(r->parsed_objects->shallow_stat, | 
|  | git_path_shallow(r))) | 
|  | die("shallow file has changed since we read it"); | 
|  | } | 
|  |  | 
|  | #define SEEN_ONLY 1 | 
|  | #define VERBOSE   2 | 
|  | #define QUICK 4 | 
|  |  | 
|  | struct write_shallow_data { | 
|  | struct strbuf *out; | 
|  | int use_pack_protocol; | 
|  | int count; | 
|  | unsigned flags; | 
|  | }; | 
|  |  | 
|  | static int write_one_shallow(const struct commit_graft *graft, void *cb_data) | 
|  | { | 
|  | struct write_shallow_data *data = cb_data; | 
|  | const char *hex = oid_to_hex(&graft->oid); | 
|  | if (graft->nr_parent != -1) | 
|  | return 0; | 
|  | if (data->flags & QUICK) { | 
|  | if (!repo_has_object_file(the_repository, &graft->oid)) | 
|  | return 0; | 
|  | } else if (data->flags & SEEN_ONLY) { | 
|  | struct commit *c = lookup_commit(the_repository, &graft->oid); | 
|  | if (!c || !(c->object.flags & SEEN)) { | 
|  | if (data->flags & VERBOSE) | 
|  | printf("Removing %s from .git/shallow\n", | 
|  | oid_to_hex(&c->object.oid)); | 
|  | return 0; | 
|  | } | 
|  | } | 
|  | data->count++; | 
|  | if (data->use_pack_protocol) | 
|  | packet_buf_write(data->out, "shallow %s", hex); | 
|  | else { | 
|  | strbuf_addstr(data->out, hex); | 
|  | strbuf_addch(data->out, '\n'); | 
|  | } | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int write_shallow_commits_1(struct strbuf *out, int use_pack_protocol, | 
|  | const struct oid_array *extra, | 
|  | unsigned flags) | 
|  | { | 
|  | struct write_shallow_data data; | 
|  | int i; | 
|  | data.out = out; | 
|  | data.use_pack_protocol = use_pack_protocol; | 
|  | data.count = 0; | 
|  | data.flags = flags; | 
|  | for_each_commit_graft(write_one_shallow, &data); | 
|  | if (!extra) | 
|  | return data.count; | 
|  | for (i = 0; i < extra->nr; i++) { | 
|  | strbuf_addstr(out, oid_to_hex(extra->oid + i)); | 
|  | strbuf_addch(out, '\n'); | 
|  | data.count++; | 
|  | } | 
|  | return data.count; | 
|  | } | 
|  |  | 
|  | int write_shallow_commits(struct strbuf *out, int use_pack_protocol, | 
|  | const struct oid_array *extra) | 
|  | { | 
|  | return write_shallow_commits_1(out, use_pack_protocol, extra, 0); | 
|  | } | 
|  |  | 
|  | const char *setup_temporary_shallow(const struct oid_array *extra) | 
|  | { | 
|  | struct tempfile *temp; | 
|  | struct strbuf sb = STRBUF_INIT; | 
|  |  | 
|  | if (write_shallow_commits(&sb, 0, extra)) { | 
|  | temp = xmks_tempfile(git_path("shallow_XXXXXX")); | 
|  |  | 
|  | if (write_in_full(temp->fd, sb.buf, sb.len) < 0 || | 
|  | close_tempfile_gently(temp) < 0) | 
|  | die_errno("failed to write to %s", | 
|  | get_tempfile_path(temp)); | 
|  | strbuf_release(&sb); | 
|  | return get_tempfile_path(temp); | 
|  | } | 
|  | /* | 
|  | * is_repository_shallow() sees empty string as "no shallow | 
|  | * file". | 
|  | */ | 
|  | return ""; | 
|  | } | 
|  |  | 
|  | void setup_alternate_shallow(struct shallow_lock *shallow_lock, | 
|  | const char **alternate_shallow_file, | 
|  | const struct oid_array *extra) | 
|  | { | 
|  | struct strbuf sb = STRBUF_INIT; | 
|  | int fd; | 
|  |  | 
|  | fd = hold_lock_file_for_update(&shallow_lock->lock, | 
|  | git_path_shallow(the_repository), | 
|  | LOCK_DIE_ON_ERROR); | 
|  | check_shallow_file_for_update(the_repository); | 
|  | if (write_shallow_commits(&sb, 0, extra)) { | 
|  | if (write_in_full(fd, sb.buf, sb.len) < 0) | 
|  | die_errno("failed to write to %s", | 
|  | get_lock_file_path(&shallow_lock->lock)); | 
|  | *alternate_shallow_file = get_lock_file_path(&shallow_lock->lock); | 
|  | } else | 
|  | /* | 
|  | * is_repository_shallow() sees empty string as "no | 
|  | * shallow file". | 
|  | */ | 
|  | *alternate_shallow_file = ""; | 
|  | strbuf_release(&sb); | 
|  | } | 
|  |  | 
|  | static int advertise_shallow_grafts_cb(const struct commit_graft *graft, void *cb) | 
|  | { | 
|  | int fd = *(int *)cb; | 
|  | if (graft->nr_parent == -1) | 
|  | packet_write_fmt(fd, "shallow %s\n", oid_to_hex(&graft->oid)); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | void advertise_shallow_grafts(int fd) | 
|  | { | 
|  | if (!is_repository_shallow(the_repository)) | 
|  | return; | 
|  | for_each_commit_graft(advertise_shallow_grafts_cb, &fd); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * mark_reachable_objects() should have been run prior to this and all | 
|  | * reachable commits marked as "SEEN", except when quick_prune is non-zero, | 
|  | * in which case lines are excised from the shallow file if they refer to | 
|  | * commits that do not exist (any longer). | 
|  | */ | 
|  | void prune_shallow(unsigned options) | 
|  | { | 
|  | struct shallow_lock shallow_lock = SHALLOW_LOCK_INIT; | 
|  | struct strbuf sb = STRBUF_INIT; | 
|  | unsigned flags = SEEN_ONLY; | 
|  | int fd; | 
|  |  | 
|  | if (options & PRUNE_QUICK) | 
|  | flags |= QUICK; | 
|  |  | 
|  | if (options & PRUNE_SHOW_ONLY) { | 
|  | flags |= VERBOSE; | 
|  | write_shallow_commits_1(&sb, 0, NULL, flags); | 
|  | strbuf_release(&sb); | 
|  | return; | 
|  | } | 
|  | fd = hold_lock_file_for_update(&shallow_lock.lock, | 
|  | git_path_shallow(the_repository), | 
|  | LOCK_DIE_ON_ERROR); | 
|  | check_shallow_file_for_update(the_repository); | 
|  | if (write_shallow_commits_1(&sb, 0, NULL, flags)) { | 
|  | if (write_in_full(fd, sb.buf, sb.len) < 0) | 
|  | die_errno("failed to write to %s", | 
|  | get_lock_file_path(&shallow_lock.lock)); | 
|  | commit_shallow_file(the_repository, &shallow_lock); | 
|  | } else { | 
|  | unlink(git_path_shallow(the_repository)); | 
|  | rollback_shallow_file(the_repository, &shallow_lock); | 
|  | } | 
|  | strbuf_release(&sb); | 
|  | } | 
|  |  | 
|  | struct trace_key trace_shallow = TRACE_KEY_INIT(SHALLOW); | 
|  |  | 
|  | /* | 
|  | * Step 1, split sender shallow commits into "ours" and "theirs" | 
|  | * Step 2, clean "ours" based on .git/shallow | 
|  | */ | 
|  | void prepare_shallow_info(struct shallow_info *info, struct oid_array *sa) | 
|  | { | 
|  | int i; | 
|  | trace_printf_key(&trace_shallow, "shallow: prepare_shallow_info\n"); | 
|  | memset(info, 0, sizeof(*info)); | 
|  | info->shallow = sa; | 
|  | if (!sa) | 
|  | return; | 
|  | ALLOC_ARRAY(info->ours, sa->nr); | 
|  | ALLOC_ARRAY(info->theirs, sa->nr); | 
|  | for (i = 0; i < sa->nr; i++) { | 
|  | if (repo_has_object_file(the_repository, sa->oid + i)) { | 
|  | struct commit_graft *graft; | 
|  | graft = lookup_commit_graft(the_repository, | 
|  | &sa->oid[i]); | 
|  | if (graft && graft->nr_parent < 0) | 
|  | continue; | 
|  | info->ours[info->nr_ours++] = i; | 
|  | } else | 
|  | info->theirs[info->nr_theirs++] = i; | 
|  | } | 
|  | } | 
|  |  | 
|  | void clear_shallow_info(struct shallow_info *info) | 
|  | { | 
|  | if (info->used_shallow) { | 
|  | for (size_t i = 0; i < info->shallow->nr; i++) | 
|  | free(info->used_shallow[i]); | 
|  | free(info->used_shallow); | 
|  | } | 
|  |  | 
|  | free(info->need_reachability_test); | 
|  | free(info->reachable); | 
|  | free(info->shallow_ref); | 
|  | free(info->ours); | 
|  | free(info->theirs); | 
|  | } | 
|  |  | 
|  | /* Step 4, remove non-existent ones in "theirs" after getting the pack */ | 
|  |  | 
|  | void remove_nonexistent_theirs_shallow(struct shallow_info *info) | 
|  | { | 
|  | struct object_id *oid = info->shallow->oid; | 
|  | int i, dst; | 
|  | trace_printf_key(&trace_shallow, "shallow: remove_nonexistent_theirs_shallow\n"); | 
|  | for (i = dst = 0; i < info->nr_theirs; i++) { | 
|  | if (i != dst) | 
|  | info->theirs[dst] = info->theirs[i]; | 
|  | if (repo_has_object_file(the_repository, oid + info->theirs[i])) | 
|  | dst++; | 
|  | } | 
|  | info->nr_theirs = dst; | 
|  | } | 
|  |  | 
|  | define_commit_slab(ref_bitmap, uint32_t *); | 
|  |  | 
|  | #define POOL_SIZE (512 * 1024) | 
|  |  | 
|  | struct paint_info { | 
|  | struct ref_bitmap ref_bitmap; | 
|  | unsigned nr_bits; | 
|  | char **pools; | 
|  | char *free, *end; | 
|  | unsigned pool_count; | 
|  | }; | 
|  |  | 
|  | static uint32_t *paint_alloc(struct paint_info *info) | 
|  | { | 
|  | unsigned nr = DIV_ROUND_UP(info->nr_bits, 32); | 
|  | unsigned size = nr * sizeof(uint32_t); | 
|  | void *p; | 
|  | if (!info->pool_count || size > info->end - info->free) { | 
|  | if (size > POOL_SIZE) | 
|  | BUG("pool size too small for %d in paint_alloc()", | 
|  | size); | 
|  | info->pool_count++; | 
|  | REALLOC_ARRAY(info->pools, info->pool_count); | 
|  | info->free = xmalloc(POOL_SIZE); | 
|  | info->pools[info->pool_count - 1] = info->free; | 
|  | info->end = info->free + POOL_SIZE; | 
|  | } | 
|  | p = info->free; | 
|  | info->free += size; | 
|  | return p; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Given a commit SHA-1, walk down to parents until either SEEN, | 
|  | * UNINTERESTING or BOTTOM is hit. Set the id-th bit in ref_bitmap for | 
|  | * all walked commits. | 
|  | */ | 
|  | static void paint_down(struct paint_info *info, const struct object_id *oid, | 
|  | unsigned int id) | 
|  | { | 
|  | unsigned int i, nr; | 
|  | struct commit_list *head = NULL; | 
|  | int bitmap_nr = DIV_ROUND_UP(info->nr_bits, 32); | 
|  | size_t bitmap_size = st_mult(sizeof(uint32_t), bitmap_nr); | 
|  | struct commit *c = lookup_commit_reference_gently(the_repository, oid, | 
|  | 1); | 
|  | uint32_t *tmp; /* to be freed before return */ | 
|  | uint32_t *bitmap; | 
|  |  | 
|  | if (!c) | 
|  | return; | 
|  |  | 
|  | tmp = xmalloc(bitmap_size); | 
|  | bitmap = paint_alloc(info); | 
|  | memset(bitmap, 0, bitmap_size); | 
|  | bitmap[id / 32] |= (1U << (id % 32)); | 
|  | commit_list_insert(c, &head); | 
|  | while (head) { | 
|  | struct commit_list *p; | 
|  | struct commit *c = pop_commit(&head); | 
|  | uint32_t **refs = ref_bitmap_at(&info->ref_bitmap, c); | 
|  |  | 
|  | /* XXX check "UNINTERESTING" from pack bitmaps if available */ | 
|  | if (c->object.flags & (SEEN | UNINTERESTING)) | 
|  | continue; | 
|  | else | 
|  | c->object.flags |= SEEN; | 
|  |  | 
|  | if (!*refs) | 
|  | *refs = bitmap; | 
|  | else { | 
|  | memcpy(tmp, *refs, bitmap_size); | 
|  | for (i = 0; i < bitmap_nr; i++) | 
|  | tmp[i] |= bitmap[i]; | 
|  | if (memcmp(tmp, *refs, bitmap_size)) { | 
|  | *refs = paint_alloc(info); | 
|  | memcpy(*refs, tmp, bitmap_size); | 
|  | } | 
|  | } | 
|  |  | 
|  | if (c->object.flags & BOTTOM) | 
|  | continue; | 
|  |  | 
|  | if (repo_parse_commit(the_repository, c)) | 
|  | die("unable to parse commit %s", | 
|  | oid_to_hex(&c->object.oid)); | 
|  |  | 
|  | for (p = c->parents; p; p = p->next) { | 
|  | if (p->item->object.flags & SEEN) | 
|  | continue; | 
|  | commit_list_insert(p->item, &head); | 
|  | } | 
|  | } | 
|  |  | 
|  | nr = get_max_object_index(); | 
|  | for (i = 0; i < nr; i++) { | 
|  | struct object *o = get_indexed_object(i); | 
|  | if (o && o->type == OBJ_COMMIT) | 
|  | o->flags &= ~SEEN; | 
|  | } | 
|  |  | 
|  | free(tmp); | 
|  | } | 
|  |  | 
|  | static int mark_uninteresting(const char *refname UNUSED, | 
|  | const char *referent UNUSED, | 
|  | const struct object_id *oid, | 
|  | int flags UNUSED, | 
|  | void *cb_data UNUSED) | 
|  | { | 
|  | struct commit *commit = lookup_commit_reference_gently(the_repository, | 
|  | oid, 1); | 
|  | if (!commit) | 
|  | return 0; | 
|  | commit->object.flags |= UNINTERESTING; | 
|  | mark_parents_uninteresting(NULL, commit); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void post_assign_shallow(struct shallow_info *info, | 
|  | struct ref_bitmap *ref_bitmap, | 
|  | int *ref_status); | 
|  | /* | 
|  | * Step 6(+7), associate shallow commits with new refs | 
|  | * | 
|  | * info->ref must be initialized before calling this function. | 
|  | * | 
|  | * If used is not NULL, it's an array of info->shallow->nr | 
|  | * bitmaps. The n-th bit set in the m-th bitmap if ref[n] needs the | 
|  | * m-th shallow commit from info->shallow. | 
|  | * | 
|  | * If used is NULL, "ours" and "theirs" are updated. And if ref_status | 
|  | * is not NULL it's an array of ref->nr ints. ref_status[i] is true if | 
|  | * the ref needs some shallow commits from either info->ours or | 
|  | * info->theirs. | 
|  | */ | 
|  | void assign_shallow_commits_to_refs(struct shallow_info *info, | 
|  | uint32_t **used, int *ref_status) | 
|  | { | 
|  | struct object_id *oid = info->shallow->oid; | 
|  | struct oid_array *ref = info->ref; | 
|  | unsigned int i, nr; | 
|  | int *shallow, nr_shallow = 0; | 
|  | struct paint_info pi; | 
|  |  | 
|  | trace_printf_key(&trace_shallow, "shallow: assign_shallow_commits_to_refs\n"); | 
|  | ALLOC_ARRAY(shallow, info->nr_ours + info->nr_theirs); | 
|  | for (i = 0; i < info->nr_ours; i++) | 
|  | shallow[nr_shallow++] = info->ours[i]; | 
|  | for (i = 0; i < info->nr_theirs; i++) | 
|  | shallow[nr_shallow++] = info->theirs[i]; | 
|  |  | 
|  | /* | 
|  | * Prepare the commit graph to track what refs can reach what | 
|  | * (new) shallow commits. | 
|  | */ | 
|  | nr = get_max_object_index(); | 
|  | for (i = 0; i < nr; i++) { | 
|  | struct object *o = get_indexed_object(i); | 
|  | if (!o || o->type != OBJ_COMMIT) | 
|  | continue; | 
|  |  | 
|  | o->flags &= ~(UNINTERESTING | BOTTOM | SEEN); | 
|  | } | 
|  |  | 
|  | memset(&pi, 0, sizeof(pi)); | 
|  | init_ref_bitmap(&pi.ref_bitmap); | 
|  | pi.nr_bits = ref->nr; | 
|  |  | 
|  | /* | 
|  | * "--not --all" to cut short the traversal if new refs | 
|  | * connect to old refs. If not (e.g. force ref updates) it'll | 
|  | * have to go down to the current shallow commits. | 
|  | */ | 
|  | refs_head_ref(get_main_ref_store(the_repository), mark_uninteresting, | 
|  | NULL); | 
|  | refs_for_each_ref(get_main_ref_store(the_repository), | 
|  | mark_uninteresting, NULL); | 
|  |  | 
|  | /* Mark potential bottoms so we won't go out of bound */ | 
|  | for (i = 0; i < nr_shallow; i++) { | 
|  | struct commit *c = lookup_commit(the_repository, | 
|  | &oid[shallow[i]]); | 
|  | c->object.flags |= BOTTOM; | 
|  | } | 
|  |  | 
|  | for (i = 0; i < ref->nr; i++) | 
|  | paint_down(&pi, ref->oid + i, i); | 
|  |  | 
|  | if (used) { | 
|  | int bitmap_size = DIV_ROUND_UP(pi.nr_bits, 32) * sizeof(uint32_t); | 
|  | memset(used, 0, sizeof(*used) * info->shallow->nr); | 
|  | for (i = 0; i < nr_shallow; i++) { | 
|  | const struct commit *c = lookup_commit(the_repository, | 
|  | &oid[shallow[i]]); | 
|  | uint32_t **map = ref_bitmap_at(&pi.ref_bitmap, c); | 
|  | if (*map) | 
|  | used[shallow[i]] = xmemdupz(*map, bitmap_size); | 
|  | } | 
|  | /* | 
|  | * unreachable shallow commits are not removed from | 
|  | * "ours" and "theirs". The user is supposed to run | 
|  | * step 7 on every ref separately and not trust "ours" | 
|  | * and "theirs" any more. | 
|  | */ | 
|  | } else | 
|  | post_assign_shallow(info, &pi.ref_bitmap, ref_status); | 
|  |  | 
|  | clear_ref_bitmap(&pi.ref_bitmap); | 
|  | for (i = 0; i < pi.pool_count; i++) | 
|  | free(pi.pools[i]); | 
|  | free(pi.pools); | 
|  | free(shallow); | 
|  | } | 
|  |  | 
|  | struct commit_array { | 
|  | struct commit **commits; | 
|  | int nr, alloc; | 
|  | }; | 
|  |  | 
|  | static int add_ref(const char *refname UNUSED, | 
|  | const char *referent UNUSED, | 
|  | const struct object_id *oid, | 
|  | int flags UNUSED, | 
|  | void *cb_data) | 
|  | { | 
|  | struct commit_array *ca = cb_data; | 
|  | ALLOC_GROW(ca->commits, ca->nr + 1, ca->alloc); | 
|  | ca->commits[ca->nr] = lookup_commit_reference_gently(the_repository, | 
|  | oid, 1); | 
|  | if (ca->commits[ca->nr]) | 
|  | ca->nr++; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void update_refstatus(int *ref_status, int nr, uint32_t *bitmap) | 
|  | { | 
|  | unsigned int i; | 
|  | if (!ref_status) | 
|  | return; | 
|  | for (i = 0; i < nr; i++) | 
|  | if (bitmap[i / 32] & (1U << (i % 32))) | 
|  | ref_status[i]++; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Step 7, reachability test on "ours" at commit level | 
|  | */ | 
|  | static void post_assign_shallow(struct shallow_info *info, | 
|  | struct ref_bitmap *ref_bitmap, | 
|  | int *ref_status) | 
|  | { | 
|  | struct object_id *oid = info->shallow->oid; | 
|  | struct commit *c; | 
|  | uint32_t **bitmap; | 
|  | int dst, i, j; | 
|  | int bitmap_nr = DIV_ROUND_UP(info->ref->nr, 32); | 
|  | struct commit_array ca; | 
|  |  | 
|  | trace_printf_key(&trace_shallow, "shallow: post_assign_shallow\n"); | 
|  | if (ref_status) | 
|  | memset(ref_status, 0, sizeof(*ref_status) * info->ref->nr); | 
|  |  | 
|  | /* Remove unreachable shallow commits from "theirs" */ | 
|  | for (i = dst = 0; i < info->nr_theirs; i++) { | 
|  | if (i != dst) | 
|  | info->theirs[dst] = info->theirs[i]; | 
|  | c = lookup_commit(the_repository, &oid[info->theirs[i]]); | 
|  | bitmap = ref_bitmap_at(ref_bitmap, c); | 
|  | if (!*bitmap) | 
|  | continue; | 
|  | for (j = 0; j < bitmap_nr; j++) | 
|  | if (bitmap[0][j]) { | 
|  | update_refstatus(ref_status, info->ref->nr, *bitmap); | 
|  | dst++; | 
|  | break; | 
|  | } | 
|  | } | 
|  | info->nr_theirs = dst; | 
|  |  | 
|  | memset(&ca, 0, sizeof(ca)); | 
|  | refs_head_ref(get_main_ref_store(the_repository), add_ref, &ca); | 
|  | refs_for_each_ref(get_main_ref_store(the_repository), add_ref, &ca); | 
|  |  | 
|  | /* Remove unreachable shallow commits from "ours" */ | 
|  | for (i = dst = 0; i < info->nr_ours; i++) { | 
|  | if (i != dst) | 
|  | info->ours[dst] = info->ours[i]; | 
|  | c = lookup_commit(the_repository, &oid[info->ours[i]]); | 
|  | bitmap = ref_bitmap_at(ref_bitmap, c); | 
|  | if (!*bitmap) | 
|  | continue; | 
|  | for (j = 0; j < bitmap_nr; j++) | 
|  | if (bitmap[0][j]) { | 
|  | /* Step 7, reachability test at commit level */ | 
|  | int ret = repo_in_merge_bases_many(the_repository, c, ca.nr, ca.commits, 1); | 
|  | if (ret < 0) | 
|  | exit(128); | 
|  | if (!ret) { | 
|  | update_refstatus(ref_status, info->ref->nr, *bitmap); | 
|  | dst++; | 
|  | break; | 
|  | } | 
|  | } | 
|  | } | 
|  | info->nr_ours = dst; | 
|  |  | 
|  | free(ca.commits); | 
|  | } | 
|  |  | 
|  | /* (Delayed) step 7, reachability test at commit level */ | 
|  | int delayed_reachability_test(struct shallow_info *si, int c) | 
|  | { | 
|  | if (si->need_reachability_test[c]) { | 
|  | struct commit *commit = lookup_commit(the_repository, | 
|  | &si->shallow->oid[c]); | 
|  |  | 
|  | if (!si->commits) { | 
|  | struct commit_array ca; | 
|  |  | 
|  | memset(&ca, 0, sizeof(ca)); | 
|  | refs_head_ref(get_main_ref_store(the_repository), | 
|  | add_ref, &ca); | 
|  | refs_for_each_ref(get_main_ref_store(the_repository), | 
|  | add_ref, &ca); | 
|  | si->commits = ca.commits; | 
|  | si->nr_commits = ca.nr; | 
|  | } | 
|  |  | 
|  | si->reachable[c] = repo_in_merge_bases_many(the_repository, | 
|  | commit, | 
|  | si->nr_commits, | 
|  | si->commits, | 
|  | 1); | 
|  | if (si->reachable[c] < 0) | 
|  | exit(128); | 
|  | si->need_reachability_test[c] = 0; | 
|  | } | 
|  | return si->reachable[c]; | 
|  | } |