commit: add repository argument to lookup_commit_reference_gently
[git/git.git] / shallow.c
1 #include "cache.h"
2 #include "repository.h"
3 #include "tempfile.h"
4 #include "lockfile.h"
5 #include "object-store.h"
6 #include "commit.h"
7 #include "tag.h"
8 #include "pkt-line.h"
9 #include "remote.h"
10 #include "refs.h"
11 #include "sha1-array.h"
12 #include "diff.h"
13 #include "revision.h"
14 #include "commit-slab.h"
15 #include "revision.h"
16 #include "list-objects.h"
17 #include "commit-slab.h"
18 #include "repository.h"
19
20 void set_alternate_shallow_file(struct repository *r, const char *path, int override)
21 {
22 if (r->parsed_objects->is_shallow != -1)
23 BUG("is_repository_shallow must not be called before set_alternate_shallow_file");
24 if (r->parsed_objects->alternate_shallow_file && !override)
25 return;
26 free(r->parsed_objects->alternate_shallow_file);
27 r->parsed_objects->alternate_shallow_file = xstrdup_or_null(path);
28 }
29
30 int register_shallow(struct repository *r, const struct object_id *oid)
31 {
32 struct commit_graft *graft =
33 xmalloc(sizeof(struct commit_graft));
34 struct commit *commit = lookup_commit(oid);
35
36 oidcpy(&graft->oid, oid);
37 graft->nr_parent = -1;
38 if (commit && commit->object.parsed)
39 commit->parents = NULL;
40 return register_commit_graft(r, graft, 0);
41 }
42
43 int is_repository_shallow(struct repository *r)
44 {
45 FILE *fp;
46 char buf[1024];
47 const char *path = r->parsed_objects->alternate_shallow_file;
48
49 if (r->parsed_objects->is_shallow >= 0)
50 return r->parsed_objects->is_shallow;
51
52 if (!path)
53 path = git_path_shallow(r);
54 /*
55 * fetch-pack sets '--shallow-file ""' as an indicator that no
56 * shallow file should be used. We could just open it and it
57 * will likely fail. But let's do an explicit check instead.
58 */
59 if (!*path || (fp = fopen(path, "r")) == NULL) {
60 stat_validity_clear(r->parsed_objects->shallow_stat);
61 r->parsed_objects->is_shallow = 0;
62 return r->parsed_objects->is_shallow;
63 }
64 stat_validity_update(r->parsed_objects->shallow_stat, fileno(fp));
65 r->parsed_objects->is_shallow = 1;
66
67 while (fgets(buf, sizeof(buf), fp)) {
68 struct object_id oid;
69 if (get_oid_hex(buf, &oid))
70 die("bad shallow line: %s", buf);
71 register_shallow(r, &oid);
72 }
73 fclose(fp);
74 return r->parsed_objects->is_shallow;
75 }
76
77 /*
78 * TODO: use "int" elemtype instead of "int *" when/if commit-slab
79 * supports a "valid" flag.
80 */
81 define_commit_slab(commit_depth, int *);
82 struct commit_list *get_shallow_commits(struct object_array *heads, int depth,
83 int shallow_flag, int not_shallow_flag)
84 {
85 int i = 0, cur_depth = 0;
86 struct commit_list *result = NULL;
87 struct object_array stack = OBJECT_ARRAY_INIT;
88 struct commit *commit = NULL;
89 struct commit_graft *graft;
90 struct commit_depth depths;
91
92 init_commit_depth(&depths);
93 while (commit || i < heads->nr || stack.nr) {
94 struct commit_list *p;
95 if (!commit) {
96 if (i < heads->nr) {
97 int **depth_slot;
98 commit = (struct commit *)
99 deref_tag(heads->objects[i++].item, NULL, 0);
100 if (!commit || commit->object.type != OBJ_COMMIT) {
101 commit = NULL;
102 continue;
103 }
104 depth_slot = commit_depth_at(&depths, commit);
105 if (!*depth_slot)
106 *depth_slot = xmalloc(sizeof(int));
107 **depth_slot = 0;
108 cur_depth = 0;
109 } else {
110 commit = (struct commit *)
111 object_array_pop(&stack);
112 cur_depth = **commit_depth_at(&depths, commit);
113 }
114 }
115 parse_commit_or_die(commit);
116 cur_depth++;
117 if ((depth != INFINITE_DEPTH && cur_depth >= depth) ||
118 (is_repository_shallow(the_repository) && !commit->parents &&
119 (graft = lookup_commit_graft(the_repository, &commit->object.oid)) != NULL &&
120 graft->nr_parent < 0)) {
121 commit_list_insert(commit, &result);
122 commit->object.flags |= shallow_flag;
123 commit = NULL;
124 continue;
125 }
126 commit->object.flags |= not_shallow_flag;
127 for (p = commit->parents, commit = NULL; p; p = p->next) {
128 int **depth_slot = commit_depth_at(&depths, p->item);
129 if (!*depth_slot) {
130 *depth_slot = xmalloc(sizeof(int));
131 **depth_slot = cur_depth;
132 } else {
133 if (cur_depth >= **depth_slot)
134 continue;
135 **depth_slot = cur_depth;
136 }
137 if (p->next)
138 add_object_array(&p->item->object,
139 NULL, &stack);
140 else {
141 commit = p->item;
142 cur_depth = **commit_depth_at(&depths, commit);
143 }
144 }
145 }
146 for (i = 0; i < depths.slab_count; i++) {
147 int j;
148
149 for (j = 0; j < depths.slab_size; j++)
150 free(depths.slab[i][j]);
151 }
152 clear_commit_depth(&depths);
153
154 return result;
155 }
156
157 static void show_commit(struct commit *commit, void *data)
158 {
159 commit_list_insert(commit, data);
160 }
161
162 /*
163 * Given rev-list arguments, run rev-list. All reachable commits
164 * except border ones are marked with not_shallow_flag. Border commits
165 * are marked with shallow_flag. The list of border/shallow commits
166 * are also returned.
167 */
168 struct commit_list *get_shallow_commits_by_rev_list(int ac, const char **av,
169 int shallow_flag,
170 int not_shallow_flag)
171 {
172 struct commit_list *result = NULL, *p;
173 struct commit_list *not_shallow_list = NULL;
174 struct rev_info revs;
175 int both_flags = shallow_flag | not_shallow_flag;
176
177 /*
178 * SHALLOW (excluded) and NOT_SHALLOW (included) should not be
179 * set at this point. But better be safe than sorry.
180 */
181 clear_object_flags(both_flags);
182
183 is_repository_shallow(the_repository); /* make sure shallows are read */
184
185 init_revisions(&revs, NULL);
186 save_commit_buffer = 0;
187 setup_revisions(ac, av, &revs, NULL);
188
189 if (prepare_revision_walk(&revs))
190 die("revision walk setup failed");
191 traverse_commit_list(&revs, show_commit, NULL, &not_shallow_list);
192
193 if (!not_shallow_list)
194 die("no commits selected for shallow requests");
195
196 /* Mark all reachable commits as NOT_SHALLOW */
197 for (p = not_shallow_list; p; p = p->next)
198 p->item->object.flags |= not_shallow_flag;
199
200 /*
201 * mark border commits SHALLOW + NOT_SHALLOW.
202 * We cannot clear NOT_SHALLOW right now. Imagine border
203 * commit A is processed first, then commit B, whose parent is
204 * A, later. If NOT_SHALLOW on A is cleared at step 1, B
205 * itself is considered border at step 2, which is incorrect.
206 */
207 for (p = not_shallow_list; p; p = p->next) {
208 struct commit *c = p->item;
209 struct commit_list *parent;
210
211 if (parse_commit(c))
212 die("unable to parse commit %s",
213 oid_to_hex(&c->object.oid));
214
215 for (parent = c->parents; parent; parent = parent->next)
216 if (!(parent->item->object.flags & not_shallow_flag)) {
217 c->object.flags |= shallow_flag;
218 commit_list_insert(c, &result);
219 break;
220 }
221 }
222 free_commit_list(not_shallow_list);
223
224 /*
225 * Now we can clean up NOT_SHALLOW on border commits. Having
226 * both flags set can confuse the caller.
227 */
228 for (p = result; p; p = p->next) {
229 struct object *o = &p->item->object;
230 if ((o->flags & both_flags) == both_flags)
231 o->flags &= ~not_shallow_flag;
232 }
233 return result;
234 }
235
236 static void check_shallow_file_for_update(struct repository *r)
237 {
238 if (r->parsed_objects->is_shallow == -1)
239 BUG("shallow must be initialized by now");
240
241 if (!stat_validity_check(r->parsed_objects->shallow_stat, git_path_shallow(the_repository)))
242 die("shallow file has changed since we read it");
243 }
244
245 #define SEEN_ONLY 1
246 #define VERBOSE 2
247
248 struct write_shallow_data {
249 struct strbuf *out;
250 int use_pack_protocol;
251 int count;
252 unsigned flags;
253 };
254
255 static int write_one_shallow(const struct commit_graft *graft, void *cb_data)
256 {
257 struct write_shallow_data *data = cb_data;
258 const char *hex = oid_to_hex(&graft->oid);
259 if (graft->nr_parent != -1)
260 return 0;
261 if (data->flags & SEEN_ONLY) {
262 struct commit *c = lookup_commit(&graft->oid);
263 if (!c || !(c->object.flags & SEEN)) {
264 if (data->flags & VERBOSE)
265 printf("Removing %s from .git/shallow\n",
266 oid_to_hex(&c->object.oid));
267 return 0;
268 }
269 }
270 data->count++;
271 if (data->use_pack_protocol)
272 packet_buf_write(data->out, "shallow %s", hex);
273 else {
274 strbuf_addstr(data->out, hex);
275 strbuf_addch(data->out, '\n');
276 }
277 return 0;
278 }
279
280 static int write_shallow_commits_1(struct strbuf *out, int use_pack_protocol,
281 const struct oid_array *extra,
282 unsigned flags)
283 {
284 struct write_shallow_data data;
285 int i;
286 data.out = out;
287 data.use_pack_protocol = use_pack_protocol;
288 data.count = 0;
289 data.flags = flags;
290 for_each_commit_graft(write_one_shallow, &data);
291 if (!extra)
292 return data.count;
293 for (i = 0; i < extra->nr; i++) {
294 strbuf_addstr(out, oid_to_hex(extra->oid + i));
295 strbuf_addch(out, '\n');
296 data.count++;
297 }
298 return data.count;
299 }
300
301 int write_shallow_commits(struct strbuf *out, int use_pack_protocol,
302 const struct oid_array *extra)
303 {
304 return write_shallow_commits_1(out, use_pack_protocol, extra, 0);
305 }
306
307 const char *setup_temporary_shallow(const struct oid_array *extra)
308 {
309 struct tempfile *temp;
310 struct strbuf sb = STRBUF_INIT;
311
312 if (write_shallow_commits(&sb, 0, extra)) {
313 temp = xmks_tempfile(git_path("shallow_XXXXXX"));
314
315 if (write_in_full(temp->fd, sb.buf, sb.len) < 0 ||
316 close_tempfile_gently(temp) < 0)
317 die_errno("failed to write to %s",
318 get_tempfile_path(temp));
319 strbuf_release(&sb);
320 return get_tempfile_path(temp);
321 }
322 /*
323 * is_repository_shallow() sees empty string as "no shallow
324 * file".
325 */
326 return "";
327 }
328
329 void setup_alternate_shallow(struct lock_file *shallow_lock,
330 const char **alternate_shallow_file,
331 const struct oid_array *extra)
332 {
333 struct strbuf sb = STRBUF_INIT;
334 int fd;
335
336 fd = hold_lock_file_for_update(shallow_lock,
337 git_path_shallow(the_repository),
338 LOCK_DIE_ON_ERROR);
339 check_shallow_file_for_update(the_repository);
340 if (write_shallow_commits(&sb, 0, extra)) {
341 if (write_in_full(fd, sb.buf, sb.len) < 0)
342 die_errno("failed to write to %s",
343 get_lock_file_path(shallow_lock));
344 *alternate_shallow_file = get_lock_file_path(shallow_lock);
345 } else
346 /*
347 * is_repository_shallow() sees empty string as "no
348 * shallow file".
349 */
350 *alternate_shallow_file = "";
351 strbuf_release(&sb);
352 }
353
354 static int advertise_shallow_grafts_cb(const struct commit_graft *graft, void *cb)
355 {
356 int fd = *(int *)cb;
357 if (graft->nr_parent == -1)
358 packet_write_fmt(fd, "shallow %s\n", oid_to_hex(&graft->oid));
359 return 0;
360 }
361
362 void advertise_shallow_grafts(int fd)
363 {
364 if (!is_repository_shallow(the_repository))
365 return;
366 for_each_commit_graft(advertise_shallow_grafts_cb, &fd);
367 }
368
369 /*
370 * mark_reachable_objects() should have been run prior to this and all
371 * reachable commits marked as "SEEN".
372 */
373 void prune_shallow(int show_only)
374 {
375 struct lock_file shallow_lock = LOCK_INIT;
376 struct strbuf sb = STRBUF_INIT;
377 int fd;
378
379 if (show_only) {
380 write_shallow_commits_1(&sb, 0, NULL, SEEN_ONLY | VERBOSE);
381 strbuf_release(&sb);
382 return;
383 }
384 fd = hold_lock_file_for_update(&shallow_lock,
385 git_path_shallow(the_repository),
386 LOCK_DIE_ON_ERROR);
387 check_shallow_file_for_update(the_repository);
388 if (write_shallow_commits_1(&sb, 0, NULL, SEEN_ONLY)) {
389 if (write_in_full(fd, sb.buf, sb.len) < 0)
390 die_errno("failed to write to %s",
391 get_lock_file_path(&shallow_lock));
392 commit_lock_file(&shallow_lock);
393 } else {
394 unlink(git_path_shallow(the_repository));
395 rollback_lock_file(&shallow_lock);
396 }
397 strbuf_release(&sb);
398 }
399
400 struct trace_key trace_shallow = TRACE_KEY_INIT(SHALLOW);
401
402 /*
403 * Step 1, split sender shallow commits into "ours" and "theirs"
404 * Step 2, clean "ours" based on .git/shallow
405 */
406 void prepare_shallow_info(struct shallow_info *info, struct oid_array *sa)
407 {
408 int i;
409 trace_printf_key(&trace_shallow, "shallow: prepare_shallow_info\n");
410 memset(info, 0, sizeof(*info));
411 info->shallow = sa;
412 if (!sa)
413 return;
414 ALLOC_ARRAY(info->ours, sa->nr);
415 ALLOC_ARRAY(info->theirs, sa->nr);
416 for (i = 0; i < sa->nr; i++) {
417 if (has_object_file(sa->oid + i)) {
418 struct commit_graft *graft;
419 graft = lookup_commit_graft(the_repository,
420 &sa->oid[i]);
421 if (graft && graft->nr_parent < 0)
422 continue;
423 info->ours[info->nr_ours++] = i;
424 } else
425 info->theirs[info->nr_theirs++] = i;
426 }
427 }
428
429 void clear_shallow_info(struct shallow_info *info)
430 {
431 free(info->ours);
432 free(info->theirs);
433 }
434
435 /* Step 4, remove non-existent ones in "theirs" after getting the pack */
436
437 void remove_nonexistent_theirs_shallow(struct shallow_info *info)
438 {
439 struct object_id *oid = info->shallow->oid;
440 int i, dst;
441 trace_printf_key(&trace_shallow, "shallow: remove_nonexistent_theirs_shallow\n");
442 for (i = dst = 0; i < info->nr_theirs; i++) {
443 if (i != dst)
444 info->theirs[dst] = info->theirs[i];
445 if (has_object_file(oid + info->theirs[i]))
446 dst++;
447 }
448 info->nr_theirs = dst;
449 }
450
451 define_commit_slab(ref_bitmap, uint32_t *);
452
453 #define POOL_SIZE (512 * 1024)
454
455 struct paint_info {
456 struct ref_bitmap ref_bitmap;
457 unsigned nr_bits;
458 char **pools;
459 char *free, *end;
460 unsigned pool_count;
461 };
462
463 static uint32_t *paint_alloc(struct paint_info *info)
464 {
465 unsigned nr = DIV_ROUND_UP(info->nr_bits, 32);
466 unsigned size = nr * sizeof(uint32_t);
467 void *p;
468 if (!info->pool_count || size > info->end - info->free) {
469 if (size > POOL_SIZE)
470 BUG("pool size too small for %d in paint_alloc()",
471 size);
472 info->pool_count++;
473 REALLOC_ARRAY(info->pools, info->pool_count);
474 info->free = xmalloc(POOL_SIZE);
475 info->pools[info->pool_count - 1] = info->free;
476 info->end = info->free + POOL_SIZE;
477 }
478 p = info->free;
479 info->free += size;
480 return p;
481 }
482
483 /*
484 * Given a commit SHA-1, walk down to parents until either SEEN,
485 * UNINTERESTING or BOTTOM is hit. Set the id-th bit in ref_bitmap for
486 * all walked commits.
487 */
488 static void paint_down(struct paint_info *info, const struct object_id *oid,
489 unsigned int id)
490 {
491 unsigned int i, nr;
492 struct commit_list *head = NULL;
493 int bitmap_nr = DIV_ROUND_UP(info->nr_bits, 32);
494 size_t bitmap_size = st_mult(sizeof(uint32_t), bitmap_nr);
495 struct commit *c = lookup_commit_reference_gently(the_repository, oid,
496 1);
497 uint32_t *tmp; /* to be freed before return */
498 uint32_t *bitmap;
499
500 if (!c)
501 return;
502
503 tmp = xmalloc(bitmap_size);
504 bitmap = paint_alloc(info);
505 memset(bitmap, 0, bitmap_size);
506 bitmap[id / 32] |= (1U << (id % 32));
507 commit_list_insert(c, &head);
508 while (head) {
509 struct commit_list *p;
510 struct commit *c = pop_commit(&head);
511 uint32_t **refs = ref_bitmap_at(&info->ref_bitmap, c);
512
513 /* XXX check "UNINTERESTING" from pack bitmaps if available */
514 if (c->object.flags & (SEEN | UNINTERESTING))
515 continue;
516 else
517 c->object.flags |= SEEN;
518
519 if (*refs == NULL)
520 *refs = bitmap;
521 else {
522 memcpy(tmp, *refs, bitmap_size);
523 for (i = 0; i < bitmap_nr; i++)
524 tmp[i] |= bitmap[i];
525 if (memcmp(tmp, *refs, bitmap_size)) {
526 *refs = paint_alloc(info);
527 memcpy(*refs, tmp, bitmap_size);
528 }
529 }
530
531 if (c->object.flags & BOTTOM)
532 continue;
533
534 if (parse_commit(c))
535 die("unable to parse commit %s",
536 oid_to_hex(&c->object.oid));
537
538 for (p = c->parents; p; p = p->next) {
539 if (p->item->object.flags & SEEN)
540 continue;
541 commit_list_insert(p->item, &head);
542 }
543 }
544
545 nr = get_max_object_index();
546 for (i = 0; i < nr; i++) {
547 struct object *o = get_indexed_object(i);
548 if (o && o->type == OBJ_COMMIT)
549 o->flags &= ~SEEN;
550 }
551
552 free(tmp);
553 }
554
555 static int mark_uninteresting(const char *refname, const struct object_id *oid,
556 int flags, void *cb_data)
557 {
558 struct commit *commit = lookup_commit_reference_gently(the_repository,
559 oid, 1);
560 if (!commit)
561 return 0;
562 commit->object.flags |= UNINTERESTING;
563 mark_parents_uninteresting(commit);
564 return 0;
565 }
566
567 static void post_assign_shallow(struct shallow_info *info,
568 struct ref_bitmap *ref_bitmap,
569 int *ref_status);
570 /*
571 * Step 6(+7), associate shallow commits with new refs
572 *
573 * info->ref must be initialized before calling this function.
574 *
575 * If used is not NULL, it's an array of info->shallow->nr
576 * bitmaps. The n-th bit set in the m-th bitmap if ref[n] needs the
577 * m-th shallow commit from info->shallow.
578 *
579 * If used is NULL, "ours" and "theirs" are updated. And if ref_status
580 * is not NULL it's an array of ref->nr ints. ref_status[i] is true if
581 * the ref needs some shallow commits from either info->ours or
582 * info->theirs.
583 */
584 void assign_shallow_commits_to_refs(struct shallow_info *info,
585 uint32_t **used, int *ref_status)
586 {
587 struct object_id *oid = info->shallow->oid;
588 struct oid_array *ref = info->ref;
589 unsigned int i, nr;
590 int *shallow, nr_shallow = 0;
591 struct paint_info pi;
592
593 trace_printf_key(&trace_shallow, "shallow: assign_shallow_commits_to_refs\n");
594 ALLOC_ARRAY(shallow, info->nr_ours + info->nr_theirs);
595 for (i = 0; i < info->nr_ours; i++)
596 shallow[nr_shallow++] = info->ours[i];
597 for (i = 0; i < info->nr_theirs; i++)
598 shallow[nr_shallow++] = info->theirs[i];
599
600 /*
601 * Prepare the commit graph to track what refs can reach what
602 * (new) shallow commits.
603 */
604 nr = get_max_object_index();
605 for (i = 0; i < nr; i++) {
606 struct object *o = get_indexed_object(i);
607 if (!o || o->type != OBJ_COMMIT)
608 continue;
609
610 o->flags &= ~(UNINTERESTING | BOTTOM | SEEN);
611 }
612
613 memset(&pi, 0, sizeof(pi));
614 init_ref_bitmap(&pi.ref_bitmap);
615 pi.nr_bits = ref->nr;
616
617 /*
618 * "--not --all" to cut short the traversal if new refs
619 * connect to old refs. If not (e.g. force ref updates) it'll
620 * have to go down to the current shallow commits.
621 */
622 head_ref(mark_uninteresting, NULL);
623 for_each_ref(mark_uninteresting, NULL);
624
625 /* Mark potential bottoms so we won't go out of bound */
626 for (i = 0; i < nr_shallow; i++) {
627 struct commit *c = lookup_commit(&oid[shallow[i]]);
628 c->object.flags |= BOTTOM;
629 }
630
631 for (i = 0; i < ref->nr; i++)
632 paint_down(&pi, ref->oid + i, i);
633
634 if (used) {
635 int bitmap_size = DIV_ROUND_UP(pi.nr_bits, 32) * sizeof(uint32_t);
636 memset(used, 0, sizeof(*used) * info->shallow->nr);
637 for (i = 0; i < nr_shallow; i++) {
638 const struct commit *c = lookup_commit(&oid[shallow[i]]);
639 uint32_t **map = ref_bitmap_at(&pi.ref_bitmap, c);
640 if (*map)
641 used[shallow[i]] = xmemdupz(*map, bitmap_size);
642 }
643 /*
644 * unreachable shallow commits are not removed from
645 * "ours" and "theirs". The user is supposed to run
646 * step 7 on every ref separately and not trust "ours"
647 * and "theirs" any more.
648 */
649 } else
650 post_assign_shallow(info, &pi.ref_bitmap, ref_status);
651
652 clear_ref_bitmap(&pi.ref_bitmap);
653 for (i = 0; i < pi.pool_count; i++)
654 free(pi.pools[i]);
655 free(pi.pools);
656 free(shallow);
657 }
658
659 struct commit_array {
660 struct commit **commits;
661 int nr, alloc;
662 };
663
664 static int add_ref(const char *refname, const struct object_id *oid,
665 int flags, void *cb_data)
666 {
667 struct commit_array *ca = cb_data;
668 ALLOC_GROW(ca->commits, ca->nr + 1, ca->alloc);
669 ca->commits[ca->nr] = lookup_commit_reference_gently(the_repository,
670 oid, 1);
671 if (ca->commits[ca->nr])
672 ca->nr++;
673 return 0;
674 }
675
676 static void update_refstatus(int *ref_status, int nr, uint32_t *bitmap)
677 {
678 unsigned int i;
679 if (!ref_status)
680 return;
681 for (i = 0; i < nr; i++)
682 if (bitmap[i / 32] & (1U << (i % 32)))
683 ref_status[i]++;
684 }
685
686 /*
687 * Step 7, reachability test on "ours" at commit level
688 */
689 static void post_assign_shallow(struct shallow_info *info,
690 struct ref_bitmap *ref_bitmap,
691 int *ref_status)
692 {
693 struct object_id *oid = info->shallow->oid;
694 struct commit *c;
695 uint32_t **bitmap;
696 int dst, i, j;
697 int bitmap_nr = DIV_ROUND_UP(info->ref->nr, 32);
698 struct commit_array ca;
699
700 trace_printf_key(&trace_shallow, "shallow: post_assign_shallow\n");
701 if (ref_status)
702 memset(ref_status, 0, sizeof(*ref_status) * info->ref->nr);
703
704 /* Remove unreachable shallow commits from "theirs" */
705 for (i = dst = 0; i < info->nr_theirs; i++) {
706 if (i != dst)
707 info->theirs[dst] = info->theirs[i];
708 c = lookup_commit(&oid[info->theirs[i]]);
709 bitmap = ref_bitmap_at(ref_bitmap, c);
710 if (!*bitmap)
711 continue;
712 for (j = 0; j < bitmap_nr; j++)
713 if (bitmap[0][j]) {
714 update_refstatus(ref_status, info->ref->nr, *bitmap);
715 dst++;
716 break;
717 }
718 }
719 info->nr_theirs = dst;
720
721 memset(&ca, 0, sizeof(ca));
722 head_ref(add_ref, &ca);
723 for_each_ref(add_ref, &ca);
724
725 /* Remove unreachable shallow commits from "ours" */
726 for (i = dst = 0; i < info->nr_ours; i++) {
727 if (i != dst)
728 info->ours[dst] = info->ours[i];
729 c = lookup_commit(&oid[info->ours[i]]);
730 bitmap = ref_bitmap_at(ref_bitmap, c);
731 if (!*bitmap)
732 continue;
733 for (j = 0; j < bitmap_nr; j++)
734 if (bitmap[0][j] &&
735 /* Step 7, reachability test at commit level */
736 !in_merge_bases_many(c, ca.nr, ca.commits)) {
737 update_refstatus(ref_status, info->ref->nr, *bitmap);
738 dst++;
739 break;
740 }
741 }
742 info->nr_ours = dst;
743
744 free(ca.commits);
745 }
746
747 /* (Delayed) step 7, reachability test at commit level */
748 int delayed_reachability_test(struct shallow_info *si, int c)
749 {
750 if (si->need_reachability_test[c]) {
751 struct commit *commit = lookup_commit(&si->shallow->oid[c]);
752
753 if (!si->commits) {
754 struct commit_array ca;
755
756 memset(&ca, 0, sizeof(ca));
757 head_ref(add_ref, &ca);
758 for_each_ref(add_ref, &ca);
759 si->commits = ca.commits;
760 si->nr_commits = ca.nr;
761 }
762
763 si->reachable[c] = in_merge_bases_many(commit,
764 si->nr_commits,
765 si->commits);
766 si->need_reachability_test[c] = 0;
767 }
768 return si->reachable[c];
769 }