The `object_directory` structure is used as an access point for a single
object directory like ".git/objects". While the structure isn't yet
fully self-contained, the intent is for it to eventually contain all
information required to access objects in one specific location.
While the name "object directory" is a good fit for now, this will
change over time as we continue with the agenda to make pluggable object
databases a thing. Eventually, objects may not be accessed via any kind
of directory at all anymore, but they could instead be backed by any
kind of durable storage mechanism. While it seems quite far-fetched for
now, it is thinkable that eventually this might even be some form of a
database, for example.
As such, the current name of this structure will become worse over time
as we evolve into the direction of pluggable ODBs. Immediate next steps
will start to carve out proper self-contained object directories, which
requires us to pass in these object directories as parameters. Based on
our modern naming schema this means that those functions should then be
named after their subsystem, which means that we would start to bake the
current name into the codebase more and more.
Let's preempt this by renaming the structure. There have been a couple
alternatives that were discussed:
- `odb_backend` was discarded because it led to the association that
one object database has a single backend, but the model is that one
alternate has one backend. Furthermore, "backend" is more about the
actual backing implementation and less about the high-level concept.
- `odb_alternate` was discarded because it is a bit of a stretch to
also call the main object directory an "alternate".
Instead, pick `odb_source` as the new name. It makes it sufficiently
clear that there can be multiple sources and does not cause confusion
when mixed with the already-existing "alternate" terminology.
In the future, this change allows us to easily introduce for example a
`odb_files_source` and other format-specific implementations.
Signed-off-by: Patrick Steinhardt <ps@pks.im>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
641 lines
16 KiB
C
641 lines
16 KiB
C
#define USE_THE_REPOSITORY_VARIABLE
|
|
#define DISABLE_SIGN_COMPARE_WARNINGS
|
|
|
|
#include "git-compat-util.h"
|
|
#include "lockfile.h"
|
|
#include "bundle.h"
|
|
#include "environment.h"
|
|
#include "gettext.h"
|
|
#include "hex.h"
|
|
#include "object-store.h"
|
|
#include "repository.h"
|
|
#include "object.h"
|
|
#include "commit.h"
|
|
#include "diff.h"
|
|
#include "revision.h"
|
|
#include "list-objects.h"
|
|
#include "run-command.h"
|
|
#include "refs.h"
|
|
#include "strvec.h"
|
|
#include "list-objects-filter-options.h"
|
|
#include "connected.h"
|
|
#include "write-or-die.h"
|
|
|
|
static const char v2_bundle_signature[] = "# v2 git bundle\n";
|
|
static const char v3_bundle_signature[] = "# v3 git bundle\n";
|
|
static struct {
|
|
int version;
|
|
const char *signature;
|
|
} bundle_sigs[] = {
|
|
{ 2, v2_bundle_signature },
|
|
{ 3, v3_bundle_signature },
|
|
};
|
|
|
|
void bundle_header_init(struct bundle_header *header)
|
|
{
|
|
struct bundle_header blank = BUNDLE_HEADER_INIT;
|
|
memcpy(header, &blank, sizeof(*header));
|
|
}
|
|
|
|
void bundle_header_release(struct bundle_header *header)
|
|
{
|
|
string_list_clear(&header->prerequisites, 1);
|
|
string_list_clear(&header->references, 1);
|
|
list_objects_filter_release(&header->filter);
|
|
}
|
|
|
|
static int parse_capability(struct bundle_header *header, const char *capability)
|
|
{
|
|
const char *arg;
|
|
if (skip_prefix(capability, "object-format=", &arg)) {
|
|
int algo = hash_algo_by_name(arg);
|
|
if (algo == GIT_HASH_UNKNOWN)
|
|
return error(_("unrecognized bundle hash algorithm: %s"), arg);
|
|
header->hash_algo = &hash_algos[algo];
|
|
return 0;
|
|
}
|
|
if (skip_prefix(capability, "filter=", &arg)) {
|
|
parse_list_objects_filter(&header->filter, arg);
|
|
return 0;
|
|
}
|
|
return error(_("unknown capability '%s'"), capability);
|
|
}
|
|
|
|
static int parse_bundle_signature(struct bundle_header *header, const char *line)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(bundle_sigs); i++) {
|
|
if (!strcmp(line, bundle_sigs[i].signature)) {
|
|
header->version = bundle_sigs[i].version;
|
|
return 0;
|
|
}
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
int read_bundle_header_fd(int fd, struct bundle_header *header,
|
|
const char *report_path)
|
|
{
|
|
struct strbuf buf = STRBUF_INIT;
|
|
int status = 0;
|
|
|
|
/* The bundle header begins with the signature */
|
|
if (strbuf_getwholeline_fd(&buf, fd, '\n') ||
|
|
parse_bundle_signature(header, buf.buf)) {
|
|
if (report_path)
|
|
error(_("'%s' does not look like a v2 or v3 bundle file"),
|
|
report_path);
|
|
status = -1;
|
|
goto abort;
|
|
}
|
|
|
|
/*
|
|
* The default hash format for bundles is SHA1, unless told otherwise
|
|
* by an "object-format=" capability, which is being handled in
|
|
* `parse_capability()`.
|
|
*/
|
|
header->hash_algo = &hash_algos[GIT_HASH_SHA1];
|
|
|
|
/* The bundle header ends with an empty line */
|
|
while (!strbuf_getwholeline_fd(&buf, fd, '\n') &&
|
|
buf.len && buf.buf[0] != '\n') {
|
|
struct object_id oid;
|
|
int is_prereq = 0;
|
|
const char *p;
|
|
|
|
strbuf_rtrim(&buf);
|
|
|
|
if (header->version == 3 && *buf.buf == '@') {
|
|
if (parse_capability(header, buf.buf + 1)) {
|
|
status = -1;
|
|
break;
|
|
}
|
|
continue;
|
|
}
|
|
|
|
if (*buf.buf == '-') {
|
|
is_prereq = 1;
|
|
strbuf_remove(&buf, 0, 1);
|
|
}
|
|
|
|
/*
|
|
* Tip lines have object name, SP, and refname.
|
|
* Prerequisites have object name that is optionally
|
|
* followed by SP and subject line.
|
|
*/
|
|
if (parse_oid_hex_algop(buf.buf, &oid, &p, header->hash_algo) ||
|
|
(*p && !isspace(*p)) ||
|
|
(!is_prereq && !*p)) {
|
|
if (report_path)
|
|
error(_("unrecognized header: %s%s (%d)"),
|
|
(is_prereq ? "-" : ""), buf.buf, (int)buf.len);
|
|
status = -1;
|
|
break;
|
|
} else {
|
|
struct object_id *dup = oiddup(&oid);
|
|
if (is_prereq)
|
|
string_list_append(&header->prerequisites, "")->util = dup;
|
|
else
|
|
string_list_append(&header->references, p + 1)->util = dup;
|
|
}
|
|
}
|
|
|
|
abort:
|
|
if (status) {
|
|
close(fd);
|
|
fd = -1;
|
|
}
|
|
strbuf_release(&buf);
|
|
return fd;
|
|
}
|
|
|
|
int read_bundle_header(const char *path, struct bundle_header *header)
|
|
{
|
|
int fd = open(path, O_RDONLY);
|
|
|
|
if (fd < 0)
|
|
return error(_("could not open '%s'"), path);
|
|
return read_bundle_header_fd(fd, header, path);
|
|
}
|
|
|
|
int is_bundle(const char *path, int quiet)
|
|
{
|
|
struct bundle_header header = BUNDLE_HEADER_INIT;
|
|
int fd = open(path, O_RDONLY);
|
|
|
|
if (fd < 0)
|
|
return 0;
|
|
fd = read_bundle_header_fd(fd, &header, quiet ? NULL : path);
|
|
if (fd >= 0)
|
|
close(fd);
|
|
bundle_header_release(&header);
|
|
return (fd >= 0);
|
|
}
|
|
|
|
static int list_refs(struct string_list *r, int argc, const char **argv)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < r->nr; i++) {
|
|
struct object_id *oid;
|
|
const char *name;
|
|
|
|
if (argc > 1) {
|
|
int j;
|
|
for (j = 1; j < argc; j++)
|
|
if (!strcmp(r->items[i].string, argv[j]))
|
|
break;
|
|
if (j == argc)
|
|
continue;
|
|
}
|
|
|
|
oid = r->items[i].util;
|
|
name = r->items[i].string;
|
|
printf("%s %s\n", oid_to_hex(oid), name);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/* Remember to update object flag allocation in object.h */
|
|
#define PREREQ_MARK (1u<<16)
|
|
|
|
struct string_list_iterator {
|
|
struct string_list *list;
|
|
size_t cur;
|
|
};
|
|
|
|
static const struct object_id *iterate_ref_map(void *cb_data)
|
|
{
|
|
struct string_list_iterator *iter = cb_data;
|
|
|
|
if (iter->cur >= iter->list->nr)
|
|
return NULL;
|
|
|
|
return iter->list->items[iter->cur++].util;
|
|
}
|
|
|
|
int verify_bundle(struct repository *r,
|
|
struct bundle_header *header,
|
|
enum verify_bundle_flags flags)
|
|
{
|
|
/*
|
|
* Do fast check, then if any prereqs are missing then go line by line
|
|
* to be verbose about the errors
|
|
*/
|
|
struct string_list *p = &header->prerequisites;
|
|
int i, ret = 0;
|
|
const char *message = _("Repository lacks these prerequisite commits:");
|
|
struct string_list_iterator iter = {
|
|
.list = p,
|
|
};
|
|
struct check_connected_options opts = {
|
|
.quiet = 1,
|
|
};
|
|
|
|
if (!r || !r->objects || !r->objects->sources)
|
|
return error(_("need a repository to verify a bundle"));
|
|
|
|
for (i = 0; i < p->nr; i++) {
|
|
struct string_list_item *e = p->items + i;
|
|
const char *name = e->string;
|
|
struct object_id *oid = e->util;
|
|
struct object *o = parse_object(r, oid);
|
|
if (o)
|
|
continue;
|
|
ret++;
|
|
if (flags & VERIFY_BUNDLE_QUIET)
|
|
continue;
|
|
if (ret == 1)
|
|
error("%s", message);
|
|
error("%s %s", oid_to_hex(oid), name);
|
|
}
|
|
if (ret)
|
|
goto cleanup;
|
|
|
|
if ((ret = check_connected(iterate_ref_map, &iter, &opts)))
|
|
error(_("some prerequisite commits exist in the object store, "
|
|
"but are not connected to the repository's history"));
|
|
|
|
/* TODO: preserve this verbose language. */
|
|
if (flags & VERIFY_BUNDLE_VERBOSE) {
|
|
struct string_list *r;
|
|
|
|
r = &header->references;
|
|
printf_ln(Q_("The bundle contains this ref:",
|
|
"The bundle contains these %"PRIuMAX" refs:",
|
|
r->nr),
|
|
(uintmax_t)r->nr);
|
|
list_refs(r, 0, NULL);
|
|
|
|
r = &header->prerequisites;
|
|
if (!r->nr) {
|
|
printf_ln(_("The bundle records a complete history."));
|
|
} else {
|
|
printf_ln(Q_("The bundle requires this ref:",
|
|
"The bundle requires these %"PRIuMAX" refs:",
|
|
r->nr),
|
|
(uintmax_t)r->nr);
|
|
list_refs(r, 0, NULL);
|
|
}
|
|
|
|
printf_ln(_("The bundle uses this hash algorithm: %s"),
|
|
header->hash_algo->name);
|
|
if (header->filter.choice)
|
|
printf_ln(_("The bundle uses this filter: %s"),
|
|
list_objects_filter_spec(&header->filter));
|
|
}
|
|
cleanup:
|
|
return ret;
|
|
}
|
|
|
|
int list_bundle_refs(struct bundle_header *header, int argc, const char **argv)
|
|
{
|
|
return list_refs(&header->references, argc, argv);
|
|
}
|
|
|
|
static int is_tag_in_date_range(struct object *tag, struct rev_info *revs)
|
|
{
|
|
unsigned long size;
|
|
enum object_type type;
|
|
char *buf = NULL, *line, *lineend;
|
|
timestamp_t date;
|
|
int result = 1;
|
|
|
|
if (revs->max_age == -1 && revs->min_age == -1)
|
|
goto out;
|
|
|
|
buf = repo_read_object_file(the_repository, &tag->oid, &type, &size);
|
|
if (!buf)
|
|
goto out;
|
|
line = memmem(buf, size, "\ntagger ", 8);
|
|
if (!line++)
|
|
goto out;
|
|
lineend = memchr(line, '\n', buf + size - line);
|
|
line = memchr(line, '>', lineend ? lineend - line : buf + size - line);
|
|
if (!line++)
|
|
goto out;
|
|
date = parse_timestamp(line, NULL, 10);
|
|
result = (revs->max_age == -1 || revs->max_age < date) &&
|
|
(revs->min_age == -1 || revs->min_age > date);
|
|
out:
|
|
free(buf);
|
|
return result;
|
|
}
|
|
|
|
|
|
/* Write the pack data to bundle_fd */
|
|
static int write_pack_data(int bundle_fd, struct rev_info *revs, struct strvec *pack_options)
|
|
{
|
|
struct child_process pack_objects = CHILD_PROCESS_INIT;
|
|
int i;
|
|
|
|
strvec_pushl(&pack_objects.args,
|
|
"pack-objects",
|
|
"--stdout", "--thin", "--delta-base-offset",
|
|
NULL);
|
|
strvec_pushv(&pack_objects.args, pack_options->v);
|
|
if (revs->filter.choice)
|
|
strvec_pushf(&pack_objects.args, "--filter=%s",
|
|
list_objects_filter_spec(&revs->filter));
|
|
pack_objects.in = -1;
|
|
pack_objects.out = bundle_fd;
|
|
pack_objects.git_cmd = 1;
|
|
|
|
/*
|
|
* start_command() will close our descriptor if it's >1. Duplicate it
|
|
* to avoid surprising the caller.
|
|
*/
|
|
if (pack_objects.out > 1) {
|
|
pack_objects.out = dup(pack_objects.out);
|
|
if (pack_objects.out < 0) {
|
|
error_errno(_("unable to dup bundle descriptor"));
|
|
child_process_clear(&pack_objects);
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
if (start_command(&pack_objects))
|
|
return error(_("Could not spawn pack-objects"));
|
|
|
|
for (i = 0; i < revs->pending.nr; i++) {
|
|
struct object *object = revs->pending.objects[i].item;
|
|
if (object->flags & UNINTERESTING)
|
|
write_or_die(pack_objects.in, "^", 1);
|
|
write_or_die(pack_objects.in, oid_to_hex(&object->oid), the_hash_algo->hexsz);
|
|
write_or_die(pack_objects.in, "\n", 1);
|
|
}
|
|
close(pack_objects.in);
|
|
if (finish_command(&pack_objects))
|
|
return error(_("pack-objects died"));
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Write out bundle refs based on the tips already
|
|
* parsed into revs.pending. As a side effect, may
|
|
* manipulate revs.pending to include additional
|
|
* necessary objects (like tags).
|
|
*
|
|
* Returns the number of refs written, or negative
|
|
* on error.
|
|
*/
|
|
static int write_bundle_refs(int bundle_fd, struct rev_info *revs)
|
|
{
|
|
int i;
|
|
int ref_count = 0;
|
|
struct strset objects = STRSET_INIT;
|
|
|
|
for (i = 0; i < revs->pending.nr; i++) {
|
|
struct object_array_entry *e = revs->pending.objects + i;
|
|
struct object_id oid;
|
|
char *ref;
|
|
const char *display_ref;
|
|
int flag;
|
|
|
|
if (e->item->flags & UNINTERESTING)
|
|
continue;
|
|
if (repo_dwim_ref(the_repository, e->name, strlen(e->name),
|
|
&oid, &ref, 0) != 1)
|
|
goto skip_write_ref;
|
|
if (refs_read_ref_full(get_main_ref_store(the_repository), e->name, RESOLVE_REF_READING, &oid, &flag))
|
|
flag = 0;
|
|
display_ref = (flag & REF_ISSYMREF) ? e->name : ref;
|
|
|
|
if (strset_contains(&objects, display_ref))
|
|
goto skip_write_ref;
|
|
|
|
if (e->item->type == OBJ_TAG &&
|
|
!is_tag_in_date_range(e->item, revs)) {
|
|
e->item->flags |= UNINTERESTING;
|
|
goto skip_write_ref;
|
|
}
|
|
|
|
/*
|
|
* Make sure the refs we wrote out is correct; --max-count and
|
|
* other limiting options could have prevented all the tips
|
|
* from getting output.
|
|
*
|
|
* Non commit objects such as tags and blobs do not have
|
|
* this issue as they are not affected by those extra
|
|
* constraints.
|
|
*/
|
|
if (!(e->item->flags & SHOWN) && e->item->type == OBJ_COMMIT) {
|
|
warning(_("ref '%s' is excluded by the rev-list options"),
|
|
e->name);
|
|
goto skip_write_ref;
|
|
}
|
|
|
|
ref_count++;
|
|
strset_add(&objects, display_ref);
|
|
write_or_die(bundle_fd, oid_to_hex(&e->item->oid), the_hash_algo->hexsz);
|
|
write_or_die(bundle_fd, " ", 1);
|
|
write_or_die(bundle_fd, display_ref, strlen(display_ref));
|
|
write_or_die(bundle_fd, "\n", 1);
|
|
skip_write_ref:
|
|
free(ref);
|
|
}
|
|
|
|
strset_clear(&objects);
|
|
|
|
/* end header */
|
|
write_or_die(bundle_fd, "\n", 1);
|
|
return ref_count;
|
|
}
|
|
|
|
struct bundle_prerequisites_info {
|
|
struct object_array *pending;
|
|
int fd;
|
|
};
|
|
|
|
static void write_bundle_prerequisites(struct commit *commit, void *data)
|
|
{
|
|
struct bundle_prerequisites_info *bpi = data;
|
|
struct object *object;
|
|
struct pretty_print_context ctx = { 0 };
|
|
struct strbuf buf = STRBUF_INIT;
|
|
|
|
if (!(commit->object.flags & BOUNDARY))
|
|
return;
|
|
strbuf_addf(&buf, "-%s ", oid_to_hex(&commit->object.oid));
|
|
write_or_die(bpi->fd, buf.buf, buf.len);
|
|
|
|
ctx.fmt = CMIT_FMT_ONELINE;
|
|
ctx.output_encoding = get_log_output_encoding();
|
|
strbuf_reset(&buf);
|
|
pretty_print_commit(&ctx, commit, &buf);
|
|
strbuf_trim(&buf);
|
|
|
|
object = (struct object *)commit;
|
|
object->flags |= UNINTERESTING;
|
|
add_object_array_with_path(object, buf.buf, bpi->pending, S_IFINVALID,
|
|
NULL);
|
|
strbuf_addch(&buf, '\n');
|
|
write_or_die(bpi->fd, buf.buf, buf.len);
|
|
strbuf_release(&buf);
|
|
}
|
|
|
|
int create_bundle(struct repository *r, const char *path,
|
|
int argc, const char **argv, struct strvec *pack_options, int version)
|
|
{
|
|
struct lock_file lock = LOCK_INIT;
|
|
int bundle_fd = -1;
|
|
int bundle_to_stdout;
|
|
int ref_count = 0;
|
|
struct rev_info revs, revs_copy;
|
|
int min_version = 2;
|
|
struct bundle_prerequisites_info bpi;
|
|
int ret;
|
|
int i;
|
|
|
|
/* init revs to list objects for pack-objects later */
|
|
save_commit_buffer = 0;
|
|
repo_init_revisions(r, &revs, NULL);
|
|
|
|
/*
|
|
* Pre-initialize the '--objects' flag so we can parse a
|
|
* --filter option successfully.
|
|
*/
|
|
revs.tree_objects = revs.blob_objects = 1;
|
|
|
|
argc = setup_revisions(argc, argv, &revs, NULL);
|
|
|
|
/*
|
|
* Reasons to require version 3:
|
|
*
|
|
* 1. @object-format is required because our hash algorithm is not
|
|
* SHA1.
|
|
* 2. @filter is required because we parsed an object filter.
|
|
*/
|
|
if (the_hash_algo != &hash_algos[GIT_HASH_SHA1] || revs.filter.choice)
|
|
min_version = 3;
|
|
|
|
if (argc > 1) {
|
|
ret = error(_("unrecognized argument: %s"), argv[1]);
|
|
goto out;
|
|
}
|
|
|
|
bundle_to_stdout = !strcmp(path, "-");
|
|
if (bundle_to_stdout)
|
|
bundle_fd = 1;
|
|
else
|
|
bundle_fd = hold_lock_file_for_update(&lock, path,
|
|
LOCK_DIE_ON_ERROR);
|
|
|
|
if (version == -1)
|
|
version = min_version;
|
|
|
|
if (version < 2 || version > 3) {
|
|
die(_("unsupported bundle version %d"), version);
|
|
} else if (version < min_version) {
|
|
die(_("cannot write bundle version %d with algorithm %s"), version, the_hash_algo->name);
|
|
} else if (version == 2) {
|
|
write_or_die(bundle_fd, v2_bundle_signature, strlen(v2_bundle_signature));
|
|
} else {
|
|
const char *capability = "@object-format=";
|
|
write_or_die(bundle_fd, v3_bundle_signature, strlen(v3_bundle_signature));
|
|
write_or_die(bundle_fd, capability, strlen(capability));
|
|
write_or_die(bundle_fd, the_hash_algo->name, strlen(the_hash_algo->name));
|
|
write_or_die(bundle_fd, "\n", 1);
|
|
|
|
if (revs.filter.choice) {
|
|
const char *value = expand_list_objects_filter_spec(&revs.filter);
|
|
capability = "@filter=";
|
|
write_or_die(bundle_fd, capability, strlen(capability));
|
|
write_or_die(bundle_fd, value, strlen(value));
|
|
write_or_die(bundle_fd, "\n", 1);
|
|
}
|
|
}
|
|
|
|
/* save revs.pending in revs_copy for later use */
|
|
memcpy(&revs_copy, &revs, sizeof(revs));
|
|
revs_copy.pending.nr = 0;
|
|
revs_copy.pending.alloc = 0;
|
|
revs_copy.pending.objects = NULL;
|
|
for (i = 0; i < revs.pending.nr; i++) {
|
|
struct object_array_entry *e = revs.pending.objects + i;
|
|
if (e)
|
|
add_object_array_with_path(e->item, e->name,
|
|
&revs_copy.pending,
|
|
e->mode, e->path);
|
|
}
|
|
|
|
/* write prerequisites */
|
|
revs.boundary = 1;
|
|
if (prepare_revision_walk(&revs))
|
|
die("revision walk setup failed");
|
|
bpi.fd = bundle_fd;
|
|
bpi.pending = &revs_copy.pending;
|
|
|
|
/*
|
|
* Remove any object walking here. We only care about commits and
|
|
* tags here. The revs_copy has the right instances of these values.
|
|
*/
|
|
revs.blob_objects = revs.tree_objects = 0;
|
|
traverse_commit_list(&revs, write_bundle_prerequisites, NULL, &bpi);
|
|
|
|
/* write bundle refs */
|
|
ref_count = write_bundle_refs(bundle_fd, &revs_copy);
|
|
if (!ref_count) {
|
|
die(_("Refusing to create empty bundle."));
|
|
} else if (ref_count < 0) {
|
|
ret = -1;
|
|
goto out;
|
|
}
|
|
|
|
/* write pack */
|
|
if (write_pack_data(bundle_fd, &revs_copy, pack_options)) {
|
|
ret = -1;
|
|
goto out;
|
|
}
|
|
|
|
if (!bundle_to_stdout) {
|
|
if (commit_lock_file(&lock))
|
|
die_errno(_("cannot create '%s'"), path);
|
|
}
|
|
|
|
ret = 0;
|
|
|
|
out:
|
|
object_array_clear(&revs_copy.pending);
|
|
release_revisions(&revs);
|
|
rollback_lock_file(&lock);
|
|
return ret;
|
|
}
|
|
|
|
int unbundle(struct repository *r, struct bundle_header *header,
|
|
int bundle_fd, struct strvec *extra_index_pack_args,
|
|
struct unbundle_opts *opts)
|
|
{
|
|
struct child_process ip = CHILD_PROCESS_INIT;
|
|
struct unbundle_opts opts_fallback = { 0 };
|
|
|
|
if (!opts)
|
|
opts = &opts_fallback;
|
|
|
|
if (verify_bundle(r, header, opts->flags)) {
|
|
close(bundle_fd);
|
|
return -1;
|
|
}
|
|
|
|
strvec_pushl(&ip.args, "index-pack", "--fix-thin", "--stdin", NULL);
|
|
|
|
/* If there is a filter, then we need to create the promisor pack. */
|
|
if (header->filter.choice)
|
|
strvec_push(&ip.args, "--promisor=from-bundle");
|
|
|
|
if (opts->flags & VERIFY_BUNDLE_FSCK)
|
|
strvec_pushf(&ip.args, "--fsck-objects%s",
|
|
opts->fsck_msg_types ? opts->fsck_msg_types : "");
|
|
|
|
if (extra_index_pack_args)
|
|
strvec_pushv(&ip.args, extra_index_pack_args->v);
|
|
|
|
ip.in = bundle_fd;
|
|
ip.no_stdout = 1;
|
|
ip.git_cmd = 1;
|
|
if (run_command(&ip))
|
|
return error(_("index-pack died"));
|
|
return 0;
|
|
}
|