|  | #include "cache.h" | 
|  | #include "repository.h" | 
|  | #include "commit.h" | 
|  | #include "tag.h" | 
|  | #include "blob.h" | 
|  | #include "http.h" | 
|  | #include "refs.h" | 
|  | #include "diff.h" | 
|  | #include "revision.h" | 
|  | #include "exec-cmd.h" | 
|  | #include "remote.h" | 
|  | #include "list-objects.h" | 
|  | #include "sigchain.h" | 
|  | #include "argv-array.h" | 
|  | #include "packfile.h" | 
|  | #include "object-store.h" | 
|  | #include "commit-reach.h" | 
|  |  | 
|  | #ifdef EXPAT_NEEDS_XMLPARSE_H | 
|  | #include <xmlparse.h> | 
|  | #else | 
|  | #include <expat.h> | 
|  | #endif | 
|  |  | 
|  | static const char http_push_usage[] = | 
|  | "git http-push [--all] [--dry-run] [--force] [--verbose] <remote> [<head>...]\n"; | 
|  |  | 
|  | #ifndef XML_STATUS_OK | 
|  | enum XML_Status { | 
|  | XML_STATUS_OK = 1, | 
|  | XML_STATUS_ERROR = 0 | 
|  | }; | 
|  | #define XML_STATUS_OK    1 | 
|  | #define XML_STATUS_ERROR 0 | 
|  | #endif | 
|  |  | 
|  | #define PREV_BUF_SIZE 4096 | 
|  |  | 
|  | /* DAV methods */ | 
|  | #define DAV_LOCK "LOCK" | 
|  | #define DAV_MKCOL "MKCOL" | 
|  | #define DAV_MOVE "MOVE" | 
|  | #define DAV_PROPFIND "PROPFIND" | 
|  | #define DAV_PUT "PUT" | 
|  | #define DAV_UNLOCK "UNLOCK" | 
|  | #define DAV_DELETE "DELETE" | 
|  |  | 
|  | /* DAV lock flags */ | 
|  | #define DAV_PROP_LOCKWR (1u << 0) | 
|  | #define DAV_PROP_LOCKEX (1u << 1) | 
|  | #define DAV_LOCK_OK (1u << 2) | 
|  |  | 
|  | /* DAV XML properties */ | 
|  | #define DAV_CTX_LOCKENTRY ".multistatus.response.propstat.prop.supportedlock.lockentry" | 
|  | #define DAV_CTX_LOCKTYPE_WRITE ".multistatus.response.propstat.prop.supportedlock.lockentry.locktype.write" | 
|  | #define DAV_CTX_LOCKTYPE_EXCLUSIVE ".multistatus.response.propstat.prop.supportedlock.lockentry.lockscope.exclusive" | 
|  | #define DAV_ACTIVELOCK_OWNER ".prop.lockdiscovery.activelock.owner.href" | 
|  | #define DAV_ACTIVELOCK_TIMEOUT ".prop.lockdiscovery.activelock.timeout" | 
|  | #define DAV_ACTIVELOCK_TOKEN ".prop.lockdiscovery.activelock.locktoken.href" | 
|  | #define DAV_PROPFIND_RESP ".multistatus.response" | 
|  | #define DAV_PROPFIND_NAME ".multistatus.response.href" | 
|  | #define DAV_PROPFIND_COLLECTION ".multistatus.response.propstat.prop.resourcetype.collection" | 
|  |  | 
|  | /* DAV request body templates */ | 
|  | #define PROPFIND_SUPPORTEDLOCK_REQUEST "<?xml version=\"1.0\" encoding=\"utf-8\" ?>\n<D:propfind xmlns:D=\"DAV:\">\n<D:prop xmlns:R=\"%s\">\n<D:supportedlock/>\n</D:prop>\n</D:propfind>" | 
|  | #define PROPFIND_ALL_REQUEST "<?xml version=\"1.0\" encoding=\"utf-8\" ?>\n<D:propfind xmlns:D=\"DAV:\">\n<D:allprop/>\n</D:propfind>" | 
|  | #define LOCK_REQUEST "<?xml version=\"1.0\" encoding=\"utf-8\" ?>\n<D:lockinfo xmlns:D=\"DAV:\">\n<D:lockscope><D:exclusive/></D:lockscope>\n<D:locktype><D:write/></D:locktype>\n<D:owner>\n<D:href>mailto:%s</D:href>\n</D:owner>\n</D:lockinfo>" | 
|  |  | 
|  | #define LOCK_TIME 600 | 
|  | #define LOCK_REFRESH 30 | 
|  |  | 
|  | /* Remember to update object flag allocation in object.h */ | 
|  | #define LOCAL    (1u<<16) | 
|  | #define REMOTE   (1u<<17) | 
|  | #define FETCHING (1u<<18) | 
|  | #define PUSHING  (1u<<19) | 
|  |  | 
|  | /* We allow "recursive" symbolic refs. Only within reason, though */ | 
|  | #define MAXDEPTH 5 | 
|  |  | 
|  | static int pushing; | 
|  | static int aborted; | 
|  | static signed char remote_dir_exists[256]; | 
|  |  | 
|  | static int push_verbosely; | 
|  | static int push_all = MATCH_REFS_NONE; | 
|  | static int force_all; | 
|  | static int dry_run; | 
|  | static int helper_status; | 
|  |  | 
|  | static struct object_list *objects; | 
|  |  | 
|  | struct repo { | 
|  | char *url; | 
|  | char *path; | 
|  | int path_len; | 
|  | int has_info_refs; | 
|  | int can_update_info_refs; | 
|  | int has_info_packs; | 
|  | struct packed_git *packs; | 
|  | struct remote_lock *locks; | 
|  | }; | 
|  |  | 
|  | static struct repo *repo; | 
|  |  | 
|  | enum transfer_state { | 
|  | NEED_FETCH, | 
|  | RUN_FETCH_LOOSE, | 
|  | RUN_FETCH_PACKED, | 
|  | NEED_PUSH, | 
|  | RUN_MKCOL, | 
|  | RUN_PUT, | 
|  | RUN_MOVE, | 
|  | ABORTED, | 
|  | COMPLETE | 
|  | }; | 
|  |  | 
|  | struct transfer_request { | 
|  | struct object *obj; | 
|  | char *url; | 
|  | char *dest; | 
|  | struct remote_lock *lock; | 
|  | struct curl_slist *headers; | 
|  | struct buffer buffer; | 
|  | enum transfer_state state; | 
|  | CURLcode curl_result; | 
|  | char errorstr[CURL_ERROR_SIZE]; | 
|  | long http_code; | 
|  | void *userData; | 
|  | struct active_request_slot *slot; | 
|  | struct transfer_request *next; | 
|  | }; | 
|  |  | 
|  | static struct transfer_request *request_queue_head; | 
|  |  | 
|  | struct xml_ctx { | 
|  | char *name; | 
|  | int len; | 
|  | char *cdata; | 
|  | void (*userFunc)(struct xml_ctx *ctx, int tag_closed); | 
|  | void *userData; | 
|  | }; | 
|  |  | 
|  | struct remote_lock { | 
|  | char *url; | 
|  | char *owner; | 
|  | char *token; | 
|  | char tmpfile_suffix[41]; | 
|  | time_t start_time; | 
|  | long timeout; | 
|  | int refreshing; | 
|  | struct remote_lock *next; | 
|  | }; | 
|  |  | 
|  | /* Flags that control remote_ls processing */ | 
|  | #define PROCESS_FILES (1u << 0) | 
|  | #define PROCESS_DIRS  (1u << 1) | 
|  | #define RECURSIVE     (1u << 2) | 
|  |  | 
|  | /* Flags that remote_ls passes to callback functions */ | 
|  | #define IS_DIR (1u << 0) | 
|  |  | 
|  | struct remote_ls_ctx { | 
|  | char *path; | 
|  | void (*userFunc)(struct remote_ls_ctx *ls); | 
|  | void *userData; | 
|  | int flags; | 
|  | char *dentry_name; | 
|  | int dentry_flags; | 
|  | struct remote_ls_ctx *parent; | 
|  | }; | 
|  |  | 
|  | /* get_dav_token_headers options */ | 
|  | enum dav_header_flag { | 
|  | DAV_HEADER_IF = (1u << 0), | 
|  | DAV_HEADER_LOCK = (1u << 1), | 
|  | DAV_HEADER_TIMEOUT = (1u << 2) | 
|  | }; | 
|  |  | 
|  | static char *xml_entities(const char *s) | 
|  | { | 
|  | struct strbuf buf = STRBUF_INIT; | 
|  | strbuf_addstr_xml_quoted(&buf, s); | 
|  | return strbuf_detach(&buf, NULL); | 
|  | } | 
|  |  | 
|  | static void curl_setup_http_get(CURL *curl, const char *url, | 
|  | const char *custom_req) | 
|  | { | 
|  | curl_easy_setopt(curl, CURLOPT_HTTPGET, 1); | 
|  | curl_easy_setopt(curl, CURLOPT_URL, url); | 
|  | curl_easy_setopt(curl, CURLOPT_CUSTOMREQUEST, custom_req); | 
|  | curl_easy_setopt(curl, CURLOPT_WRITEFUNCTION, fwrite_null); | 
|  | } | 
|  |  | 
|  | static void curl_setup_http(CURL *curl, const char *url, | 
|  | const char *custom_req, struct buffer *buffer, | 
|  | curl_write_callback write_fn) | 
|  | { | 
|  | curl_easy_setopt(curl, CURLOPT_PUT, 1); | 
|  | curl_easy_setopt(curl, CURLOPT_URL, url); | 
|  | curl_easy_setopt(curl, CURLOPT_INFILE, buffer); | 
|  | curl_easy_setopt(curl, CURLOPT_INFILESIZE, buffer->buf.len); | 
|  | curl_easy_setopt(curl, CURLOPT_READFUNCTION, fread_buffer); | 
|  | #ifndef NO_CURL_IOCTL | 
|  | curl_easy_setopt(curl, CURLOPT_IOCTLFUNCTION, ioctl_buffer); | 
|  | curl_easy_setopt(curl, CURLOPT_IOCTLDATA, buffer); | 
|  | #endif | 
|  | curl_easy_setopt(curl, CURLOPT_WRITEFUNCTION, write_fn); | 
|  | curl_easy_setopt(curl, CURLOPT_NOBODY, 0); | 
|  | curl_easy_setopt(curl, CURLOPT_CUSTOMREQUEST, custom_req); | 
|  | curl_easy_setopt(curl, CURLOPT_UPLOAD, 1); | 
|  | } | 
|  |  | 
|  | static struct curl_slist *get_dav_token_headers(struct remote_lock *lock, enum dav_header_flag options) | 
|  | { | 
|  | struct strbuf buf = STRBUF_INIT; | 
|  | struct curl_slist *dav_headers = http_copy_default_headers(); | 
|  |  | 
|  | if (options & DAV_HEADER_IF) { | 
|  | strbuf_addf(&buf, "If: (<%s>)", lock->token); | 
|  | dav_headers = curl_slist_append(dav_headers, buf.buf); | 
|  | strbuf_reset(&buf); | 
|  | } | 
|  | if (options & DAV_HEADER_LOCK) { | 
|  | strbuf_addf(&buf, "Lock-Token: <%s>", lock->token); | 
|  | dav_headers = curl_slist_append(dav_headers, buf.buf); | 
|  | strbuf_reset(&buf); | 
|  | } | 
|  | if (options & DAV_HEADER_TIMEOUT) { | 
|  | strbuf_addf(&buf, "Timeout: Second-%ld", lock->timeout); | 
|  | dav_headers = curl_slist_append(dav_headers, buf.buf); | 
|  | strbuf_reset(&buf); | 
|  | } | 
|  | strbuf_release(&buf); | 
|  |  | 
|  | return dav_headers; | 
|  | } | 
|  |  | 
|  | static void finish_request(struct transfer_request *request); | 
|  | static void release_request(struct transfer_request *request); | 
|  |  | 
|  | static void process_response(void *callback_data) | 
|  | { | 
|  | struct transfer_request *request = | 
|  | (struct transfer_request *)callback_data; | 
|  |  | 
|  | finish_request(request); | 
|  | } | 
|  |  | 
|  | #ifdef USE_CURL_MULTI | 
|  |  | 
|  | static void start_fetch_loose(struct transfer_request *request) | 
|  | { | 
|  | struct active_request_slot *slot; | 
|  | struct http_object_request *obj_req; | 
|  |  | 
|  | obj_req = new_http_object_request(repo->url, &request->obj->oid); | 
|  | if (obj_req == NULL) { | 
|  | request->state = ABORTED; | 
|  | return; | 
|  | } | 
|  |  | 
|  | slot = obj_req->slot; | 
|  | slot->callback_func = process_response; | 
|  | slot->callback_data = request; | 
|  | request->slot = slot; | 
|  | request->userData = obj_req; | 
|  |  | 
|  | /* Try to get the request started, abort the request on error */ | 
|  | request->state = RUN_FETCH_LOOSE; | 
|  | if (!start_active_slot(slot)) { | 
|  | fprintf(stderr, "Unable to start GET request\n"); | 
|  | repo->can_update_info_refs = 0; | 
|  | release_http_object_request(obj_req); | 
|  | release_request(request); | 
|  | } | 
|  | } | 
|  |  | 
|  | static void start_mkcol(struct transfer_request *request) | 
|  | { | 
|  | char *hex = oid_to_hex(&request->obj->oid); | 
|  | struct active_request_slot *slot; | 
|  |  | 
|  | request->url = get_remote_object_url(repo->url, hex, 1); | 
|  |  | 
|  | slot = get_active_slot(); | 
|  | slot->callback_func = process_response; | 
|  | slot->callback_data = request; | 
|  | curl_setup_http_get(slot->curl, request->url, DAV_MKCOL); | 
|  | curl_easy_setopt(slot->curl, CURLOPT_ERRORBUFFER, request->errorstr); | 
|  |  | 
|  | if (start_active_slot(slot)) { | 
|  | request->slot = slot; | 
|  | request->state = RUN_MKCOL; | 
|  | } else { | 
|  | request->state = ABORTED; | 
|  | FREE_AND_NULL(request->url); | 
|  | } | 
|  | } | 
|  | #endif | 
|  |  | 
|  | static void start_fetch_packed(struct transfer_request *request) | 
|  | { | 
|  | struct packed_git *target; | 
|  |  | 
|  | struct transfer_request *check_request = request_queue_head; | 
|  | struct http_pack_request *preq; | 
|  |  | 
|  | target = find_sha1_pack(request->obj->oid.hash, repo->packs); | 
|  | if (!target) { | 
|  | fprintf(stderr, "Unable to fetch %s, will not be able to update server info refs\n", oid_to_hex(&request->obj->oid)); | 
|  | repo->can_update_info_refs = 0; | 
|  | release_request(request); | 
|  | return; | 
|  | } | 
|  |  | 
|  | fprintf(stderr,	"Fetching pack %s\n", sha1_to_hex(target->sha1)); | 
|  | fprintf(stderr, " which contains %s\n", oid_to_hex(&request->obj->oid)); | 
|  |  | 
|  | preq = new_http_pack_request(target, repo->url); | 
|  | if (preq == NULL) { | 
|  | repo->can_update_info_refs = 0; | 
|  | return; | 
|  | } | 
|  | preq->lst = &repo->packs; | 
|  |  | 
|  | /* Make sure there isn't another open request for this pack */ | 
|  | while (check_request) { | 
|  | if (check_request->state == RUN_FETCH_PACKED && | 
|  | !strcmp(check_request->url, preq->url)) { | 
|  | release_http_pack_request(preq); | 
|  | release_request(request); | 
|  | return; | 
|  | } | 
|  | check_request = check_request->next; | 
|  | } | 
|  |  | 
|  | preq->slot->callback_func = process_response; | 
|  | preq->slot->callback_data = request; | 
|  | request->slot = preq->slot; | 
|  | request->userData = preq; | 
|  |  | 
|  | /* Try to get the request started, abort the request on error */ | 
|  | request->state = RUN_FETCH_PACKED; | 
|  | if (!start_active_slot(preq->slot)) { | 
|  | fprintf(stderr, "Unable to start GET request\n"); | 
|  | release_http_pack_request(preq); | 
|  | repo->can_update_info_refs = 0; | 
|  | release_request(request); | 
|  | } | 
|  | } | 
|  |  | 
|  | static void start_put(struct transfer_request *request) | 
|  | { | 
|  | char *hex = oid_to_hex(&request->obj->oid); | 
|  | struct active_request_slot *slot; | 
|  | struct strbuf buf = STRBUF_INIT; | 
|  | enum object_type type; | 
|  | char hdr[50]; | 
|  | void *unpacked; | 
|  | unsigned long len; | 
|  | int hdrlen; | 
|  | ssize_t size; | 
|  | git_zstream stream; | 
|  |  | 
|  | unpacked = read_object_file(&request->obj->oid, &type, &len); | 
|  | hdrlen = xsnprintf(hdr, sizeof(hdr), "%s %"PRIuMAX , type_name(type), (uintmax_t)len) + 1; | 
|  |  | 
|  | /* Set it up */ | 
|  | git_deflate_init(&stream, zlib_compression_level); | 
|  | size = git_deflate_bound(&stream, len + hdrlen); | 
|  | strbuf_init(&request->buffer.buf, size); | 
|  | request->buffer.posn = 0; | 
|  |  | 
|  | /* Compress it */ | 
|  | stream.next_out = (unsigned char *)request->buffer.buf.buf; | 
|  | stream.avail_out = size; | 
|  |  | 
|  | /* First header.. */ | 
|  | stream.next_in = (void *)hdr; | 
|  | stream.avail_in = hdrlen; | 
|  | while (git_deflate(&stream, 0) == Z_OK) | 
|  | ; /* nothing */ | 
|  |  | 
|  | /* Then the data itself.. */ | 
|  | stream.next_in = unpacked; | 
|  | stream.avail_in = len; | 
|  | while (git_deflate(&stream, Z_FINISH) == Z_OK) | 
|  | ; /* nothing */ | 
|  | git_deflate_end(&stream); | 
|  | free(unpacked); | 
|  |  | 
|  | request->buffer.buf.len = stream.total_out; | 
|  |  | 
|  | strbuf_addstr(&buf, "Destination: "); | 
|  | append_remote_object_url(&buf, repo->url, hex, 0); | 
|  | request->dest = strbuf_detach(&buf, NULL); | 
|  |  | 
|  | append_remote_object_url(&buf, repo->url, hex, 0); | 
|  | strbuf_add(&buf, request->lock->tmpfile_suffix, 41); | 
|  | request->url = strbuf_detach(&buf, NULL); | 
|  |  | 
|  | slot = get_active_slot(); | 
|  | slot->callback_func = process_response; | 
|  | slot->callback_data = request; | 
|  | curl_setup_http(slot->curl, request->url, DAV_PUT, | 
|  | &request->buffer, fwrite_null); | 
|  |  | 
|  | if (start_active_slot(slot)) { | 
|  | request->slot = slot; | 
|  | request->state = RUN_PUT; | 
|  | } else { | 
|  | request->state = ABORTED; | 
|  | FREE_AND_NULL(request->url); | 
|  | } | 
|  | } | 
|  |  | 
|  | static void start_move(struct transfer_request *request) | 
|  | { | 
|  | struct active_request_slot *slot; | 
|  | struct curl_slist *dav_headers = http_copy_default_headers(); | 
|  |  | 
|  | slot = get_active_slot(); | 
|  | slot->callback_func = process_response; | 
|  | slot->callback_data = request; | 
|  | curl_setup_http_get(slot->curl, request->url, DAV_MOVE); | 
|  | dav_headers = curl_slist_append(dav_headers, request->dest); | 
|  | dav_headers = curl_slist_append(dav_headers, "Overwrite: T"); | 
|  | curl_easy_setopt(slot->curl, CURLOPT_HTTPHEADER, dav_headers); | 
|  |  | 
|  | if (start_active_slot(slot)) { | 
|  | request->slot = slot; | 
|  | request->state = RUN_MOVE; | 
|  | } else { | 
|  | request->state = ABORTED; | 
|  | FREE_AND_NULL(request->url); | 
|  | } | 
|  | } | 
|  |  | 
|  | static int refresh_lock(struct remote_lock *lock) | 
|  | { | 
|  | struct active_request_slot *slot; | 
|  | struct slot_results results; | 
|  | struct curl_slist *dav_headers; | 
|  | int rc = 0; | 
|  |  | 
|  | lock->refreshing = 1; | 
|  |  | 
|  | dav_headers = get_dav_token_headers(lock, DAV_HEADER_IF | DAV_HEADER_TIMEOUT); | 
|  |  | 
|  | slot = get_active_slot(); | 
|  | slot->results = &results; | 
|  | curl_setup_http_get(slot->curl, lock->url, DAV_LOCK); | 
|  | curl_easy_setopt(slot->curl, CURLOPT_HTTPHEADER, dav_headers); | 
|  |  | 
|  | if (start_active_slot(slot)) { | 
|  | run_active_slot(slot); | 
|  | if (results.curl_result != CURLE_OK) { | 
|  | fprintf(stderr, "LOCK HTTP error %ld\n", | 
|  | results.http_code); | 
|  | } else { | 
|  | lock->start_time = time(NULL); | 
|  | rc = 1; | 
|  | } | 
|  | } | 
|  |  | 
|  | lock->refreshing = 0; | 
|  | curl_slist_free_all(dav_headers); | 
|  |  | 
|  | return rc; | 
|  | } | 
|  |  | 
|  | static void check_locks(void) | 
|  | { | 
|  | struct remote_lock *lock = repo->locks; | 
|  | time_t current_time = time(NULL); | 
|  | int time_remaining; | 
|  |  | 
|  | while (lock) { | 
|  | time_remaining = lock->start_time + lock->timeout - | 
|  | current_time; | 
|  | if (!lock->refreshing && time_remaining < LOCK_REFRESH) { | 
|  | if (!refresh_lock(lock)) { | 
|  | fprintf(stderr, | 
|  | "Unable to refresh lock for %s\n", | 
|  | lock->url); | 
|  | aborted = 1; | 
|  | return; | 
|  | } | 
|  | } | 
|  | lock = lock->next; | 
|  | } | 
|  | } | 
|  |  | 
|  | static void release_request(struct transfer_request *request) | 
|  | { | 
|  | struct transfer_request *entry = request_queue_head; | 
|  |  | 
|  | if (request == request_queue_head) { | 
|  | request_queue_head = request->next; | 
|  | } else { | 
|  | while (entry->next != NULL && entry->next != request) | 
|  | entry = entry->next; | 
|  | if (entry->next == request) | 
|  | entry->next = entry->next->next; | 
|  | } | 
|  |  | 
|  | free(request->url); | 
|  | free(request); | 
|  | } | 
|  |  | 
|  | static void finish_request(struct transfer_request *request) | 
|  | { | 
|  | struct http_pack_request *preq; | 
|  | struct http_object_request *obj_req; | 
|  |  | 
|  | request->curl_result = request->slot->curl_result; | 
|  | request->http_code = request->slot->http_code; | 
|  | request->slot = NULL; | 
|  |  | 
|  | /* Keep locks active */ | 
|  | check_locks(); | 
|  |  | 
|  | if (request->headers != NULL) | 
|  | curl_slist_free_all(request->headers); | 
|  |  | 
|  | /* URL is reused for MOVE after PUT */ | 
|  | if (request->state != RUN_PUT) { | 
|  | FREE_AND_NULL(request->url); | 
|  | } | 
|  |  | 
|  | if (request->state == RUN_MKCOL) { | 
|  | if (request->curl_result == CURLE_OK || | 
|  | request->http_code == 405) { | 
|  | remote_dir_exists[request->obj->oid.hash[0]] = 1; | 
|  | start_put(request); | 
|  | } else { | 
|  | fprintf(stderr, "MKCOL %s failed, aborting (%d/%ld)\n", | 
|  | oid_to_hex(&request->obj->oid), | 
|  | request->curl_result, request->http_code); | 
|  | request->state = ABORTED; | 
|  | aborted = 1; | 
|  | } | 
|  | } else if (request->state == RUN_PUT) { | 
|  | if (request->curl_result == CURLE_OK) { | 
|  | start_move(request); | 
|  | } else { | 
|  | fprintf(stderr,	"PUT %s failed, aborting (%d/%ld)\n", | 
|  | oid_to_hex(&request->obj->oid), | 
|  | request->curl_result, request->http_code); | 
|  | request->state = ABORTED; | 
|  | aborted = 1; | 
|  | } | 
|  | } else if (request->state == RUN_MOVE) { | 
|  | if (request->curl_result == CURLE_OK) { | 
|  | if (push_verbosely) | 
|  | fprintf(stderr, "    sent %s\n", | 
|  | oid_to_hex(&request->obj->oid)); | 
|  | request->obj->flags |= REMOTE; | 
|  | release_request(request); | 
|  | } else { | 
|  | fprintf(stderr, "MOVE %s failed, aborting (%d/%ld)\n", | 
|  | oid_to_hex(&request->obj->oid), | 
|  | request->curl_result, request->http_code); | 
|  | request->state = ABORTED; | 
|  | aborted = 1; | 
|  | } | 
|  | } else if (request->state == RUN_FETCH_LOOSE) { | 
|  | obj_req = (struct http_object_request *)request->userData; | 
|  |  | 
|  | if (finish_http_object_request(obj_req) == 0) | 
|  | if (obj_req->rename == 0) | 
|  | request->obj->flags |= (LOCAL | REMOTE); | 
|  |  | 
|  | /* Try fetching packed if necessary */ | 
|  | if (request->obj->flags & LOCAL) { | 
|  | release_http_object_request(obj_req); | 
|  | release_request(request); | 
|  | } else | 
|  | start_fetch_packed(request); | 
|  |  | 
|  | } else if (request->state == RUN_FETCH_PACKED) { | 
|  | int fail = 1; | 
|  | if (request->curl_result != CURLE_OK) { | 
|  | fprintf(stderr, "Unable to get pack file %s\n%s", | 
|  | request->url, curl_errorstr); | 
|  | } else { | 
|  | preq = (struct http_pack_request *)request->userData; | 
|  |  | 
|  | if (preq) { | 
|  | if (finish_http_pack_request(preq) == 0) | 
|  | fail = 0; | 
|  | release_http_pack_request(preq); | 
|  | } | 
|  | } | 
|  | if (fail) | 
|  | repo->can_update_info_refs = 0; | 
|  | release_request(request); | 
|  | } | 
|  | } | 
|  |  | 
|  | #ifdef USE_CURL_MULTI | 
|  | static int is_running_queue; | 
|  | static int fill_active_slot(void *unused) | 
|  | { | 
|  | struct transfer_request *request; | 
|  |  | 
|  | if (aborted || !is_running_queue) | 
|  | return 0; | 
|  |  | 
|  | for (request = request_queue_head; request; request = request->next) { | 
|  | if (request->state == NEED_FETCH) { | 
|  | start_fetch_loose(request); | 
|  | return 1; | 
|  | } else if (pushing && request->state == NEED_PUSH) { | 
|  | if (remote_dir_exists[request->obj->oid.hash[0]] == 1) { | 
|  | start_put(request); | 
|  | } else { | 
|  | start_mkcol(request); | 
|  | } | 
|  | return 1; | 
|  | } | 
|  | } | 
|  | return 0; | 
|  | } | 
|  | #endif | 
|  |  | 
|  | static void get_remote_object_list(unsigned char parent); | 
|  |  | 
|  | static void add_fetch_request(struct object *obj) | 
|  | { | 
|  | struct transfer_request *request; | 
|  |  | 
|  | check_locks(); | 
|  |  | 
|  | /* | 
|  | * Don't fetch the object if it's known to exist locally | 
|  | * or is already in the request queue | 
|  | */ | 
|  | if (remote_dir_exists[obj->oid.hash[0]] == -1) | 
|  | get_remote_object_list(obj->oid.hash[0]); | 
|  | if (obj->flags & (LOCAL | FETCHING)) | 
|  | return; | 
|  |  | 
|  | obj->flags |= FETCHING; | 
|  | request = xmalloc(sizeof(*request)); | 
|  | request->obj = obj; | 
|  | request->url = NULL; | 
|  | request->lock = NULL; | 
|  | request->headers = NULL; | 
|  | request->state = NEED_FETCH; | 
|  | request->next = request_queue_head; | 
|  | request_queue_head = request; | 
|  |  | 
|  | #ifdef USE_CURL_MULTI | 
|  | fill_active_slots(); | 
|  | step_active_slots(); | 
|  | #endif | 
|  | } | 
|  |  | 
|  | static int add_send_request(struct object *obj, struct remote_lock *lock) | 
|  | { | 
|  | struct transfer_request *request; | 
|  | struct packed_git *target; | 
|  |  | 
|  | /* Keep locks active */ | 
|  | check_locks(); | 
|  |  | 
|  | /* | 
|  | * Don't push the object if it's known to exist on the remote | 
|  | * or is already in the request queue | 
|  | */ | 
|  | if (remote_dir_exists[obj->oid.hash[0]] == -1) | 
|  | get_remote_object_list(obj->oid.hash[0]); | 
|  | if (obj->flags & (REMOTE | PUSHING)) | 
|  | return 0; | 
|  | target = find_sha1_pack(obj->oid.hash, repo->packs); | 
|  | if (target) { | 
|  | obj->flags |= REMOTE; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | obj->flags |= PUSHING; | 
|  | request = xmalloc(sizeof(*request)); | 
|  | request->obj = obj; | 
|  | request->url = NULL; | 
|  | request->lock = lock; | 
|  | request->headers = NULL; | 
|  | request->state = NEED_PUSH; | 
|  | request->next = request_queue_head; | 
|  | request_queue_head = request; | 
|  |  | 
|  | #ifdef USE_CURL_MULTI | 
|  | fill_active_slots(); | 
|  | step_active_slots(); | 
|  | #endif | 
|  |  | 
|  | return 1; | 
|  | } | 
|  |  | 
|  | static int fetch_indices(void) | 
|  | { | 
|  | int ret; | 
|  |  | 
|  | if (push_verbosely) | 
|  | fprintf(stderr, "Getting pack list\n"); | 
|  |  | 
|  | switch (http_get_info_packs(repo->url, &repo->packs)) { | 
|  | case HTTP_OK: | 
|  | case HTTP_MISSING_TARGET: | 
|  | ret = 0; | 
|  | break; | 
|  | default: | 
|  | ret = -1; | 
|  | } | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void one_remote_object(const struct object_id *oid) | 
|  | { | 
|  | struct object *obj; | 
|  |  | 
|  | obj = lookup_object(the_repository, oid->hash); | 
|  | if (!obj) | 
|  | obj = parse_object(the_repository, oid); | 
|  |  | 
|  | /* Ignore remote objects that don't exist locally */ | 
|  | if (!obj) | 
|  | return; | 
|  |  | 
|  | obj->flags |= REMOTE; | 
|  | if (!object_list_contains(objects, obj)) | 
|  | object_list_insert(obj, &objects); | 
|  | } | 
|  |  | 
|  | static void handle_lockprop_ctx(struct xml_ctx *ctx, int tag_closed) | 
|  | { | 
|  | int *lock_flags = (int *)ctx->userData; | 
|  |  | 
|  | if (tag_closed) { | 
|  | if (!strcmp(ctx->name, DAV_CTX_LOCKENTRY)) { | 
|  | if ((*lock_flags & DAV_PROP_LOCKEX) && | 
|  | (*lock_flags & DAV_PROP_LOCKWR)) { | 
|  | *lock_flags |= DAV_LOCK_OK; | 
|  | } | 
|  | *lock_flags &= DAV_LOCK_OK; | 
|  | } else if (!strcmp(ctx->name, DAV_CTX_LOCKTYPE_WRITE)) { | 
|  | *lock_flags |= DAV_PROP_LOCKWR; | 
|  | } else if (!strcmp(ctx->name, DAV_CTX_LOCKTYPE_EXCLUSIVE)) { | 
|  | *lock_flags |= DAV_PROP_LOCKEX; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | static void handle_new_lock_ctx(struct xml_ctx *ctx, int tag_closed) | 
|  | { | 
|  | struct remote_lock *lock = (struct remote_lock *)ctx->userData; | 
|  | git_SHA_CTX sha_ctx; | 
|  | unsigned char lock_token_sha1[20]; | 
|  |  | 
|  | if (tag_closed && ctx->cdata) { | 
|  | if (!strcmp(ctx->name, DAV_ACTIVELOCK_OWNER)) { | 
|  | lock->owner = xstrdup(ctx->cdata); | 
|  | } else if (!strcmp(ctx->name, DAV_ACTIVELOCK_TIMEOUT)) { | 
|  | const char *arg; | 
|  | if (skip_prefix(ctx->cdata, "Second-", &arg)) | 
|  | lock->timeout = strtol(arg, NULL, 10); | 
|  | } else if (!strcmp(ctx->name, DAV_ACTIVELOCK_TOKEN)) { | 
|  | lock->token = xstrdup(ctx->cdata); | 
|  |  | 
|  | git_SHA1_Init(&sha_ctx); | 
|  | git_SHA1_Update(&sha_ctx, lock->token, strlen(lock->token)); | 
|  | git_SHA1_Final(lock_token_sha1, &sha_ctx); | 
|  |  | 
|  | lock->tmpfile_suffix[0] = '_'; | 
|  | memcpy(lock->tmpfile_suffix + 1, sha1_to_hex(lock_token_sha1), 40); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | static void one_remote_ref(const char *refname); | 
|  |  | 
|  | static void | 
|  | xml_start_tag(void *userData, const char *name, const char **atts) | 
|  | { | 
|  | struct xml_ctx *ctx = (struct xml_ctx *)userData; | 
|  | const char *c = strchr(name, ':'); | 
|  | int old_namelen, new_len; | 
|  |  | 
|  | if (c == NULL) | 
|  | c = name; | 
|  | else | 
|  | c++; | 
|  |  | 
|  | old_namelen = strlen(ctx->name); | 
|  | new_len = old_namelen + strlen(c) + 2; | 
|  |  | 
|  | if (new_len > ctx->len) { | 
|  | ctx->name = xrealloc(ctx->name, new_len); | 
|  | ctx->len = new_len; | 
|  | } | 
|  | xsnprintf(ctx->name + old_namelen, ctx->len - old_namelen, ".%s", c); | 
|  |  | 
|  | FREE_AND_NULL(ctx->cdata); | 
|  |  | 
|  | ctx->userFunc(ctx, 0); | 
|  | } | 
|  |  | 
|  | static void | 
|  | xml_end_tag(void *userData, const char *name) | 
|  | { | 
|  | struct xml_ctx *ctx = (struct xml_ctx *)userData; | 
|  | const char *c = strchr(name, ':'); | 
|  | char *ep; | 
|  |  | 
|  | ctx->userFunc(ctx, 1); | 
|  |  | 
|  | if (c == NULL) | 
|  | c = name; | 
|  | else | 
|  | c++; | 
|  |  | 
|  | ep = ctx->name + strlen(ctx->name) - strlen(c) - 1; | 
|  | *ep = 0; | 
|  | } | 
|  |  | 
|  | static void | 
|  | xml_cdata(void *userData, const XML_Char *s, int len) | 
|  | { | 
|  | struct xml_ctx *ctx = (struct xml_ctx *)userData; | 
|  | free(ctx->cdata); | 
|  | ctx->cdata = xmemdupz(s, len); | 
|  | } | 
|  |  | 
|  | static struct remote_lock *lock_remote(const char *path, long timeout) | 
|  | { | 
|  | struct active_request_slot *slot; | 
|  | struct slot_results results; | 
|  | struct buffer out_buffer = { STRBUF_INIT, 0 }; | 
|  | struct strbuf in_buffer = STRBUF_INIT; | 
|  | char *url; | 
|  | char *ep; | 
|  | char timeout_header[25]; | 
|  | struct remote_lock *lock = NULL; | 
|  | struct curl_slist *dav_headers = http_copy_default_headers(); | 
|  | struct xml_ctx ctx; | 
|  | char *escaped; | 
|  |  | 
|  | url = xstrfmt("%s%s", repo->url, path); | 
|  |  | 
|  | /* Make sure leading directories exist for the remote ref */ | 
|  | ep = strchr(url + strlen(repo->url) + 1, '/'); | 
|  | while (ep) { | 
|  | char saved_character = ep[1]; | 
|  | ep[1] = '\0'; | 
|  | slot = get_active_slot(); | 
|  | slot->results = &results; | 
|  | curl_setup_http_get(slot->curl, url, DAV_MKCOL); | 
|  | if (start_active_slot(slot)) { | 
|  | run_active_slot(slot); | 
|  | if (results.curl_result != CURLE_OK && | 
|  | results.http_code != 405) { | 
|  | fprintf(stderr, | 
|  | "Unable to create branch path %s\n", | 
|  | url); | 
|  | free(url); | 
|  | return NULL; | 
|  | } | 
|  | } else { | 
|  | fprintf(stderr, "Unable to start MKCOL request\n"); | 
|  | free(url); | 
|  | return NULL; | 
|  | } | 
|  | ep[1] = saved_character; | 
|  | ep = strchr(ep + 1, '/'); | 
|  | } | 
|  |  | 
|  | escaped = xml_entities(ident_default_email()); | 
|  | strbuf_addf(&out_buffer.buf, LOCK_REQUEST, escaped); | 
|  | free(escaped); | 
|  |  | 
|  | xsnprintf(timeout_header, sizeof(timeout_header), "Timeout: Second-%ld", timeout); | 
|  | dav_headers = curl_slist_append(dav_headers, timeout_header); | 
|  | dav_headers = curl_slist_append(dav_headers, "Content-Type: text/xml"); | 
|  |  | 
|  | slot = get_active_slot(); | 
|  | slot->results = &results; | 
|  | curl_setup_http(slot->curl, url, DAV_LOCK, &out_buffer, fwrite_buffer); | 
|  | curl_easy_setopt(slot->curl, CURLOPT_HTTPHEADER, dav_headers); | 
|  | curl_easy_setopt(slot->curl, CURLOPT_FILE, &in_buffer); | 
|  |  | 
|  | lock = xcalloc(1, sizeof(*lock)); | 
|  | lock->timeout = -1; | 
|  |  | 
|  | if (start_active_slot(slot)) { | 
|  | run_active_slot(slot); | 
|  | if (results.curl_result == CURLE_OK) { | 
|  | XML_Parser parser = XML_ParserCreate(NULL); | 
|  | enum XML_Status result; | 
|  | ctx.name = xcalloc(10, 1); | 
|  | ctx.len = 0; | 
|  | ctx.cdata = NULL; | 
|  | ctx.userFunc = handle_new_lock_ctx; | 
|  | ctx.userData = lock; | 
|  | XML_SetUserData(parser, &ctx); | 
|  | XML_SetElementHandler(parser, xml_start_tag, | 
|  | xml_end_tag); | 
|  | XML_SetCharacterDataHandler(parser, xml_cdata); | 
|  | result = XML_Parse(parser, in_buffer.buf, | 
|  | in_buffer.len, 1); | 
|  | free(ctx.name); | 
|  | if (result != XML_STATUS_OK) { | 
|  | fprintf(stderr, "XML error: %s\n", | 
|  | XML_ErrorString( | 
|  | XML_GetErrorCode(parser))); | 
|  | lock->timeout = -1; | 
|  | } | 
|  | XML_ParserFree(parser); | 
|  | } else { | 
|  | fprintf(stderr, | 
|  | "error: curl result=%d, HTTP code=%ld\n", | 
|  | results.curl_result, results.http_code); | 
|  | } | 
|  | } else { | 
|  | fprintf(stderr, "Unable to start LOCK request\n"); | 
|  | } | 
|  |  | 
|  | curl_slist_free_all(dav_headers); | 
|  | strbuf_release(&out_buffer.buf); | 
|  | strbuf_release(&in_buffer); | 
|  |  | 
|  | if (lock->token == NULL || lock->timeout <= 0) { | 
|  | free(lock->token); | 
|  | free(lock->owner); | 
|  | free(url); | 
|  | FREE_AND_NULL(lock); | 
|  | } else { | 
|  | lock->url = url; | 
|  | lock->start_time = time(NULL); | 
|  | lock->next = repo->locks; | 
|  | repo->locks = lock; | 
|  | } | 
|  |  | 
|  | return lock; | 
|  | } | 
|  |  | 
|  | static int unlock_remote(struct remote_lock *lock) | 
|  | { | 
|  | struct active_request_slot *slot; | 
|  | struct slot_results results; | 
|  | struct remote_lock *prev = repo->locks; | 
|  | struct curl_slist *dav_headers; | 
|  | int rc = 0; | 
|  |  | 
|  | dav_headers = get_dav_token_headers(lock, DAV_HEADER_LOCK); | 
|  |  | 
|  | slot = get_active_slot(); | 
|  | slot->results = &results; | 
|  | curl_setup_http_get(slot->curl, lock->url, DAV_UNLOCK); | 
|  | curl_easy_setopt(slot->curl, CURLOPT_HTTPHEADER, dav_headers); | 
|  |  | 
|  | if (start_active_slot(slot)) { | 
|  | run_active_slot(slot); | 
|  | if (results.curl_result == CURLE_OK) | 
|  | rc = 1; | 
|  | else | 
|  | fprintf(stderr, "UNLOCK HTTP error %ld\n", | 
|  | results.http_code); | 
|  | } else { | 
|  | fprintf(stderr, "Unable to start UNLOCK request\n"); | 
|  | } | 
|  |  | 
|  | curl_slist_free_all(dav_headers); | 
|  |  | 
|  | if (repo->locks == lock) { | 
|  | repo->locks = lock->next; | 
|  | } else { | 
|  | while (prev && prev->next != lock) | 
|  | prev = prev->next; | 
|  | if (prev) | 
|  | prev->next = prev->next->next; | 
|  | } | 
|  |  | 
|  | free(lock->owner); | 
|  | free(lock->url); | 
|  | free(lock->token); | 
|  | free(lock); | 
|  |  | 
|  | return rc; | 
|  | } | 
|  |  | 
|  | static void remove_locks(void) | 
|  | { | 
|  | struct remote_lock *lock = repo->locks; | 
|  |  | 
|  | fprintf(stderr, "Removing remote locks...\n"); | 
|  | while (lock) { | 
|  | struct remote_lock *next = lock->next; | 
|  | unlock_remote(lock); | 
|  | lock = next; | 
|  | } | 
|  | } | 
|  |  | 
|  | static void remove_locks_on_signal(int signo) | 
|  | { | 
|  | remove_locks(); | 
|  | sigchain_pop(signo); | 
|  | raise(signo); | 
|  | } | 
|  |  | 
|  | static void remote_ls(const char *path, int flags, | 
|  | void (*userFunc)(struct remote_ls_ctx *ls), | 
|  | void *userData); | 
|  |  | 
|  | /* extract hex from sharded "xx/x{38}" filename */ | 
|  | static int get_oid_hex_from_objpath(const char *path, struct object_id *oid) | 
|  | { | 
|  | if (strlen(path) != GIT_SHA1_HEXSZ + 1) | 
|  | return -1; | 
|  |  | 
|  | if (hex_to_bytes(oid->hash, path, 1)) | 
|  | return -1; | 
|  | path += 2; | 
|  | path++; /* skip '/' */ | 
|  |  | 
|  | return hex_to_bytes(oid->hash + 1, path, GIT_SHA1_RAWSZ - 1); | 
|  | } | 
|  |  | 
|  | static void process_ls_object(struct remote_ls_ctx *ls) | 
|  | { | 
|  | unsigned int *parent = (unsigned int *)ls->userData; | 
|  | const char *path = ls->dentry_name; | 
|  | struct object_id oid; | 
|  |  | 
|  | if (!strcmp(ls->path, ls->dentry_name) && (ls->flags & IS_DIR)) { | 
|  | remote_dir_exists[*parent] = 1; | 
|  | return; | 
|  | } | 
|  |  | 
|  | if (!skip_prefix(path, "objects/", &path) || | 
|  | get_oid_hex_from_objpath(path, &oid)) | 
|  | return; | 
|  |  | 
|  | one_remote_object(&oid); | 
|  | } | 
|  |  | 
|  | static void process_ls_ref(struct remote_ls_ctx *ls) | 
|  | { | 
|  | if (!strcmp(ls->path, ls->dentry_name) && (ls->dentry_flags & IS_DIR)) { | 
|  | fprintf(stderr, "  %s\n", ls->dentry_name); | 
|  | return; | 
|  | } | 
|  |  | 
|  | if (!(ls->dentry_flags & IS_DIR)) | 
|  | one_remote_ref(ls->dentry_name); | 
|  | } | 
|  |  | 
|  | static void handle_remote_ls_ctx(struct xml_ctx *ctx, int tag_closed) | 
|  | { | 
|  | struct remote_ls_ctx *ls = (struct remote_ls_ctx *)ctx->userData; | 
|  |  | 
|  | if (tag_closed) { | 
|  | if (!strcmp(ctx->name, DAV_PROPFIND_RESP) && ls->dentry_name) { | 
|  | if (ls->dentry_flags & IS_DIR) { | 
|  |  | 
|  | /* ensure collection names end with slash */ | 
|  | str_end_url_with_slash(ls->dentry_name, &ls->dentry_name); | 
|  |  | 
|  | if (ls->flags & PROCESS_DIRS) { | 
|  | ls->userFunc(ls); | 
|  | } | 
|  | if (strcmp(ls->dentry_name, ls->path) && | 
|  | ls->flags & RECURSIVE) { | 
|  | remote_ls(ls->dentry_name, | 
|  | ls->flags, | 
|  | ls->userFunc, | 
|  | ls->userData); | 
|  | } | 
|  | } else if (ls->flags & PROCESS_FILES) { | 
|  | ls->userFunc(ls); | 
|  | } | 
|  | } else if (!strcmp(ctx->name, DAV_PROPFIND_NAME) && ctx->cdata) { | 
|  | char *path = ctx->cdata; | 
|  | if (*ctx->cdata == 'h') { | 
|  | path = strstr(path, "//"); | 
|  | if (path) { | 
|  | path = strchr(path+2, '/'); | 
|  | } | 
|  | } | 
|  | if (path) { | 
|  | const char *url = repo->url; | 
|  | if (repo->path) | 
|  | url = repo->path; | 
|  | if (strncmp(path, url, repo->path_len)) | 
|  | error("Parsed path '%s' does not match url: '%s'", | 
|  | path, url); | 
|  | else { | 
|  | path += repo->path_len; | 
|  | ls->dentry_name = xstrdup(path); | 
|  | } | 
|  | } | 
|  | } else if (!strcmp(ctx->name, DAV_PROPFIND_COLLECTION)) { | 
|  | ls->dentry_flags |= IS_DIR; | 
|  | } | 
|  | } else if (!strcmp(ctx->name, DAV_PROPFIND_RESP)) { | 
|  | FREE_AND_NULL(ls->dentry_name); | 
|  | ls->dentry_flags = 0; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* | 
|  | * NEEDSWORK: remote_ls() ignores info/refs on the remote side.  But it | 
|  | * should _only_ heed the information from that file, instead of trying to | 
|  | * determine the refs from the remote file system (badly: it does not even | 
|  | * know about packed-refs). | 
|  | */ | 
|  | static void remote_ls(const char *path, int flags, | 
|  | void (*userFunc)(struct remote_ls_ctx *ls), | 
|  | void *userData) | 
|  | { | 
|  | char *url = xstrfmt("%s%s", repo->url, path); | 
|  | struct active_request_slot *slot; | 
|  | struct slot_results results; | 
|  | struct strbuf in_buffer = STRBUF_INIT; | 
|  | struct buffer out_buffer = { STRBUF_INIT, 0 }; | 
|  | struct curl_slist *dav_headers = http_copy_default_headers(); | 
|  | struct xml_ctx ctx; | 
|  | struct remote_ls_ctx ls; | 
|  |  | 
|  | ls.flags = flags; | 
|  | ls.path = xstrdup(path); | 
|  | ls.dentry_name = NULL; | 
|  | ls.dentry_flags = 0; | 
|  | ls.userData = userData; | 
|  | ls.userFunc = userFunc; | 
|  |  | 
|  | strbuf_addstr(&out_buffer.buf, PROPFIND_ALL_REQUEST); | 
|  |  | 
|  | dav_headers = curl_slist_append(dav_headers, "Depth: 1"); | 
|  | dav_headers = curl_slist_append(dav_headers, "Content-Type: text/xml"); | 
|  |  | 
|  | slot = get_active_slot(); | 
|  | slot->results = &results; | 
|  | curl_setup_http(slot->curl, url, DAV_PROPFIND, | 
|  | &out_buffer, fwrite_buffer); | 
|  | curl_easy_setopt(slot->curl, CURLOPT_HTTPHEADER, dav_headers); | 
|  | curl_easy_setopt(slot->curl, CURLOPT_FILE, &in_buffer); | 
|  |  | 
|  | if (start_active_slot(slot)) { | 
|  | run_active_slot(slot); | 
|  | if (results.curl_result == CURLE_OK) { | 
|  | XML_Parser parser = XML_ParserCreate(NULL); | 
|  | enum XML_Status result; | 
|  | ctx.name = xcalloc(10, 1); | 
|  | ctx.len = 0; | 
|  | ctx.cdata = NULL; | 
|  | ctx.userFunc = handle_remote_ls_ctx; | 
|  | ctx.userData = &ls; | 
|  | XML_SetUserData(parser, &ctx); | 
|  | XML_SetElementHandler(parser, xml_start_tag, | 
|  | xml_end_tag); | 
|  | XML_SetCharacterDataHandler(parser, xml_cdata); | 
|  | result = XML_Parse(parser, in_buffer.buf, | 
|  | in_buffer.len, 1); | 
|  | free(ctx.name); | 
|  |  | 
|  | if (result != XML_STATUS_OK) { | 
|  | fprintf(stderr, "XML error: %s\n", | 
|  | XML_ErrorString( | 
|  | XML_GetErrorCode(parser))); | 
|  | } | 
|  | XML_ParserFree(parser); | 
|  | } | 
|  | } else { | 
|  | fprintf(stderr, "Unable to start PROPFIND request\n"); | 
|  | } | 
|  |  | 
|  | free(ls.path); | 
|  | free(url); | 
|  | strbuf_release(&out_buffer.buf); | 
|  | strbuf_release(&in_buffer); | 
|  | curl_slist_free_all(dav_headers); | 
|  | } | 
|  |  | 
|  | static void get_remote_object_list(unsigned char parent) | 
|  | { | 
|  | char path[] = "objects/XX/"; | 
|  | static const char hex[] = "0123456789abcdef"; | 
|  | unsigned int val = parent; | 
|  |  | 
|  | path[8] = hex[val >> 4]; | 
|  | path[9] = hex[val & 0xf]; | 
|  | remote_dir_exists[val] = 0; | 
|  | remote_ls(path, (PROCESS_FILES | PROCESS_DIRS), | 
|  | process_ls_object, &val); | 
|  | } | 
|  |  | 
|  | static int locking_available(void) | 
|  | { | 
|  | struct active_request_slot *slot; | 
|  | struct slot_results results; | 
|  | struct strbuf in_buffer = STRBUF_INIT; | 
|  | struct buffer out_buffer = { STRBUF_INIT, 0 }; | 
|  | struct curl_slist *dav_headers = http_copy_default_headers(); | 
|  | struct xml_ctx ctx; | 
|  | int lock_flags = 0; | 
|  | char *escaped; | 
|  |  | 
|  | escaped = xml_entities(repo->url); | 
|  | strbuf_addf(&out_buffer.buf, PROPFIND_SUPPORTEDLOCK_REQUEST, escaped); | 
|  | free(escaped); | 
|  |  | 
|  | dav_headers = curl_slist_append(dav_headers, "Depth: 0"); | 
|  | dav_headers = curl_slist_append(dav_headers, "Content-Type: text/xml"); | 
|  |  | 
|  | slot = get_active_slot(); | 
|  | slot->results = &results; | 
|  | curl_setup_http(slot->curl, repo->url, DAV_PROPFIND, | 
|  | &out_buffer, fwrite_buffer); | 
|  | curl_easy_setopt(slot->curl, CURLOPT_HTTPHEADER, dav_headers); | 
|  | curl_easy_setopt(slot->curl, CURLOPT_FILE, &in_buffer); | 
|  |  | 
|  | if (start_active_slot(slot)) { | 
|  | run_active_slot(slot); | 
|  | if (results.curl_result == CURLE_OK) { | 
|  | XML_Parser parser = XML_ParserCreate(NULL); | 
|  | enum XML_Status result; | 
|  | ctx.name = xcalloc(10, 1); | 
|  | ctx.len = 0; | 
|  | ctx.cdata = NULL; | 
|  | ctx.userFunc = handle_lockprop_ctx; | 
|  | ctx.userData = &lock_flags; | 
|  | XML_SetUserData(parser, &ctx); | 
|  | XML_SetElementHandler(parser, xml_start_tag, | 
|  | xml_end_tag); | 
|  | result = XML_Parse(parser, in_buffer.buf, | 
|  | in_buffer.len, 1); | 
|  | free(ctx.name); | 
|  |  | 
|  | if (result != XML_STATUS_OK) { | 
|  | fprintf(stderr, "XML error: %s\n", | 
|  | XML_ErrorString( | 
|  | XML_GetErrorCode(parser))); | 
|  | lock_flags = 0; | 
|  | } | 
|  | XML_ParserFree(parser); | 
|  | if (!lock_flags) | 
|  | error("no DAV locking support on %s", | 
|  | repo->url); | 
|  |  | 
|  | } else { | 
|  | error("Cannot access URL %s, return code %d", | 
|  | repo->url, results.curl_result); | 
|  | lock_flags = 0; | 
|  | } | 
|  | } else { | 
|  | error("Unable to start PROPFIND request on %s", repo->url); | 
|  | } | 
|  |  | 
|  | strbuf_release(&out_buffer.buf); | 
|  | strbuf_release(&in_buffer); | 
|  | curl_slist_free_all(dav_headers); | 
|  |  | 
|  | return lock_flags; | 
|  | } | 
|  |  | 
|  | static struct object_list **add_one_object(struct object *obj, struct object_list **p) | 
|  | { | 
|  | struct object_list *entry = xmalloc(sizeof(struct object_list)); | 
|  | entry->item = obj; | 
|  | entry->next = *p; | 
|  | *p = entry; | 
|  | return &entry->next; | 
|  | } | 
|  |  | 
|  | static struct object_list **process_blob(struct blob *blob, | 
|  | struct object_list **p) | 
|  | { | 
|  | struct object *obj = &blob->object; | 
|  |  | 
|  | obj->flags |= LOCAL; | 
|  |  | 
|  | if (obj->flags & (UNINTERESTING | SEEN)) | 
|  | return p; | 
|  |  | 
|  | obj->flags |= SEEN; | 
|  | return add_one_object(obj, p); | 
|  | } | 
|  |  | 
|  | static struct object_list **process_tree(struct tree *tree, | 
|  | struct object_list **p) | 
|  | { | 
|  | struct object *obj = &tree->object; | 
|  | struct tree_desc desc; | 
|  | struct name_entry entry; | 
|  |  | 
|  | obj->flags |= LOCAL; | 
|  |  | 
|  | if (obj->flags & (UNINTERESTING | SEEN)) | 
|  | return p; | 
|  | if (parse_tree(tree) < 0) | 
|  | die("bad tree object %s", oid_to_hex(&obj->oid)); | 
|  |  | 
|  | obj->flags |= SEEN; | 
|  | p = add_one_object(obj, p); | 
|  |  | 
|  | init_tree_desc(&desc, tree->buffer, tree->size); | 
|  |  | 
|  | while (tree_entry(&desc, &entry)) | 
|  | switch (object_type(entry.mode)) { | 
|  | case OBJ_TREE: | 
|  | p = process_tree(lookup_tree(the_repository, &entry.oid), | 
|  | p); | 
|  | break; | 
|  | case OBJ_BLOB: | 
|  | p = process_blob(lookup_blob(the_repository, &entry.oid), | 
|  | p); | 
|  | break; | 
|  | default: | 
|  | /* Subproject commit - not in this repository */ | 
|  | break; | 
|  | } | 
|  |  | 
|  | free_tree_buffer(tree); | 
|  | return p; | 
|  | } | 
|  |  | 
|  | static int get_delta(struct rev_info *revs, struct remote_lock *lock) | 
|  | { | 
|  | int i; | 
|  | struct commit *commit; | 
|  | struct object_list **p = &objects; | 
|  | int count = 0; | 
|  |  | 
|  | while ((commit = get_revision(revs)) != NULL) { | 
|  | p = process_tree(get_commit_tree(commit), p); | 
|  | commit->object.flags |= LOCAL; | 
|  | if (!(commit->object.flags & UNINTERESTING)) | 
|  | count += add_send_request(&commit->object, lock); | 
|  | } | 
|  |  | 
|  | for (i = 0; i < revs->pending.nr; i++) { | 
|  | struct object_array_entry *entry = revs->pending.objects + i; | 
|  | struct object *obj = entry->item; | 
|  | const char *name = entry->name; | 
|  |  | 
|  | if (obj->flags & (UNINTERESTING | SEEN)) | 
|  | continue; | 
|  | if (obj->type == OBJ_TAG) { | 
|  | obj->flags |= SEEN; | 
|  | p = add_one_object(obj, p); | 
|  | continue; | 
|  | } | 
|  | if (obj->type == OBJ_TREE) { | 
|  | p = process_tree((struct tree *)obj, p); | 
|  | continue; | 
|  | } | 
|  | if (obj->type == OBJ_BLOB) { | 
|  | p = process_blob((struct blob *)obj, p); | 
|  | continue; | 
|  | } | 
|  | die("unknown pending object %s (%s)", oid_to_hex(&obj->oid), name); | 
|  | } | 
|  |  | 
|  | while (objects) { | 
|  | if (!(objects->item->flags & UNINTERESTING)) | 
|  | count += add_send_request(objects->item, lock); | 
|  | objects = objects->next; | 
|  | } | 
|  |  | 
|  | return count; | 
|  | } | 
|  |  | 
|  | static int update_remote(unsigned char *sha1, struct remote_lock *lock) | 
|  | { | 
|  | struct active_request_slot *slot; | 
|  | struct slot_results results; | 
|  | struct buffer out_buffer = { STRBUF_INIT, 0 }; | 
|  | struct curl_slist *dav_headers; | 
|  |  | 
|  | dav_headers = get_dav_token_headers(lock, DAV_HEADER_IF); | 
|  |  | 
|  | strbuf_addf(&out_buffer.buf, "%s\n", sha1_to_hex(sha1)); | 
|  |  | 
|  | slot = get_active_slot(); | 
|  | slot->results = &results; | 
|  | curl_setup_http(slot->curl, lock->url, DAV_PUT, | 
|  | &out_buffer, fwrite_null); | 
|  | curl_easy_setopt(slot->curl, CURLOPT_HTTPHEADER, dav_headers); | 
|  |  | 
|  | if (start_active_slot(slot)) { | 
|  | run_active_slot(slot); | 
|  | strbuf_release(&out_buffer.buf); | 
|  | if (results.curl_result != CURLE_OK) { | 
|  | fprintf(stderr, | 
|  | "PUT error: curl result=%d, HTTP code=%ld\n", | 
|  | results.curl_result, results.http_code); | 
|  | /* We should attempt recovery? */ | 
|  | return 0; | 
|  | } | 
|  | } else { | 
|  | strbuf_release(&out_buffer.buf); | 
|  | fprintf(stderr, "Unable to start PUT request\n"); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | return 1; | 
|  | } | 
|  |  | 
|  | static struct ref *remote_refs; | 
|  |  | 
|  | static void one_remote_ref(const char *refname) | 
|  | { | 
|  | struct ref *ref; | 
|  | struct object *obj; | 
|  |  | 
|  | ref = alloc_ref(refname); | 
|  |  | 
|  | if (http_fetch_ref(repo->url, ref) != 0) { | 
|  | fprintf(stderr, | 
|  | "Unable to fetch ref %s from %s\n", | 
|  | refname, repo->url); | 
|  | free(ref); | 
|  | return; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Fetch a copy of the object if it doesn't exist locally - it | 
|  | * may be required for updating server info later. | 
|  | */ | 
|  | if (repo->can_update_info_refs && !has_object_file(&ref->old_oid)) { | 
|  | obj = lookup_unknown_object(ref->old_oid.hash); | 
|  | fprintf(stderr,	"  fetch %s for %s\n", | 
|  | oid_to_hex(&ref->old_oid), refname); | 
|  | add_fetch_request(obj); | 
|  | } | 
|  |  | 
|  | ref->next = remote_refs; | 
|  | remote_refs = ref; | 
|  | } | 
|  |  | 
|  | static void get_dav_remote_heads(void) | 
|  | { | 
|  | remote_ls("refs/", (PROCESS_FILES | PROCESS_DIRS | RECURSIVE), process_ls_ref, NULL); | 
|  | } | 
|  |  | 
|  | static void add_remote_info_ref(struct remote_ls_ctx *ls) | 
|  | { | 
|  | struct strbuf *buf = (struct strbuf *)ls->userData; | 
|  | struct object *o; | 
|  | struct ref *ref; | 
|  |  | 
|  | ref = alloc_ref(ls->dentry_name); | 
|  |  | 
|  | if (http_fetch_ref(repo->url, ref) != 0) { | 
|  | fprintf(stderr, | 
|  | "Unable to fetch ref %s from %s\n", | 
|  | ls->dentry_name, repo->url); | 
|  | aborted = 1; | 
|  | free(ref); | 
|  | return; | 
|  | } | 
|  |  | 
|  | o = parse_object(the_repository, &ref->old_oid); | 
|  | if (!o) { | 
|  | fprintf(stderr, | 
|  | "Unable to parse object %s for remote ref %s\n", | 
|  | oid_to_hex(&ref->old_oid), ls->dentry_name); | 
|  | aborted = 1; | 
|  | free(ref); | 
|  | return; | 
|  | } | 
|  |  | 
|  | strbuf_addf(buf, "%s\t%s\n", | 
|  | oid_to_hex(&ref->old_oid), ls->dentry_name); | 
|  |  | 
|  | if (o->type == OBJ_TAG) { | 
|  | o = deref_tag(the_repository, o, ls->dentry_name, 0); | 
|  | if (o) | 
|  | strbuf_addf(buf, "%s\t%s^{}\n", | 
|  | oid_to_hex(&o->oid), ls->dentry_name); | 
|  | } | 
|  | free(ref); | 
|  | } | 
|  |  | 
|  | static void update_remote_info_refs(struct remote_lock *lock) | 
|  | { | 
|  | struct buffer buffer = { STRBUF_INIT, 0 }; | 
|  | struct active_request_slot *slot; | 
|  | struct slot_results results; | 
|  | struct curl_slist *dav_headers; | 
|  |  | 
|  | remote_ls("refs/", (PROCESS_FILES | RECURSIVE), | 
|  | add_remote_info_ref, &buffer.buf); | 
|  | if (!aborted) { | 
|  | dav_headers = get_dav_token_headers(lock, DAV_HEADER_IF); | 
|  |  | 
|  | slot = get_active_slot(); | 
|  | slot->results = &results; | 
|  | curl_setup_http(slot->curl, lock->url, DAV_PUT, | 
|  | &buffer, fwrite_null); | 
|  | curl_easy_setopt(slot->curl, CURLOPT_HTTPHEADER, dav_headers); | 
|  |  | 
|  | if (start_active_slot(slot)) { | 
|  | run_active_slot(slot); | 
|  | if (results.curl_result != CURLE_OK) { | 
|  | fprintf(stderr, | 
|  | "PUT error: curl result=%d, HTTP code=%ld\n", | 
|  | results.curl_result, results.http_code); | 
|  | } | 
|  | } | 
|  | } | 
|  | strbuf_release(&buffer.buf); | 
|  | } | 
|  |  | 
|  | static int remote_exists(const char *path) | 
|  | { | 
|  | char *url = xstrfmt("%s%s", repo->url, path); | 
|  | int ret; | 
|  |  | 
|  |  | 
|  | switch (http_get_strbuf(url, NULL, NULL)) { | 
|  | case HTTP_OK: | 
|  | ret = 1; | 
|  | break; | 
|  | case HTTP_MISSING_TARGET: | 
|  | ret = 0; | 
|  | break; | 
|  | case HTTP_ERROR: | 
|  | error("unable to access '%s': %s", url, curl_errorstr); | 
|  | /* fallthrough */ | 
|  | default: | 
|  | ret = -1; | 
|  | } | 
|  | free(url); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void fetch_symref(const char *path, char **symref, struct object_id *oid) | 
|  | { | 
|  | char *url = xstrfmt("%s%s", repo->url, path); | 
|  | struct strbuf buffer = STRBUF_INIT; | 
|  | const char *name; | 
|  |  | 
|  | if (http_get_strbuf(url, &buffer, NULL) != HTTP_OK) | 
|  | die("Couldn't get %s for remote symref\n%s", url, | 
|  | curl_errorstr); | 
|  | free(url); | 
|  |  | 
|  | FREE_AND_NULL(*symref); | 
|  | oidclr(oid); | 
|  |  | 
|  | if (buffer.len == 0) | 
|  | return; | 
|  |  | 
|  | /* Cut off trailing newline. */ | 
|  | strbuf_rtrim(&buffer); | 
|  |  | 
|  | /* If it's a symref, set the refname; otherwise try for a sha1 */ | 
|  | if (skip_prefix(buffer.buf, "ref: ", &name)) { | 
|  | *symref = xmemdupz(name, buffer.len - (name - buffer.buf)); | 
|  | } else { | 
|  | get_oid_hex(buffer.buf, oid); | 
|  | } | 
|  |  | 
|  | strbuf_release(&buffer); | 
|  | } | 
|  |  | 
|  | static int verify_merge_base(struct object_id *head_oid, struct ref *remote) | 
|  | { | 
|  | struct commit *head = lookup_commit_or_die(head_oid, "HEAD"); | 
|  | struct commit *branch = lookup_commit_or_die(&remote->old_oid, | 
|  | remote->name); | 
|  |  | 
|  | return in_merge_bases(branch, head); | 
|  | } | 
|  |  | 
|  | static int delete_remote_branch(const char *pattern, int force) | 
|  | { | 
|  | struct ref *refs = remote_refs; | 
|  | struct ref *remote_ref = NULL; | 
|  | struct object_id head_oid; | 
|  | char *symref = NULL; | 
|  | int match; | 
|  | int patlen = strlen(pattern); | 
|  | int i; | 
|  | struct active_request_slot *slot; | 
|  | struct slot_results results; | 
|  | char *url; | 
|  |  | 
|  | /* Find the remote branch(es) matching the specified branch name */ | 
|  | for (match = 0; refs; refs = refs->next) { | 
|  | char *name = refs->name; | 
|  | int namelen = strlen(name); | 
|  | if (namelen < patlen || | 
|  | memcmp(name + namelen - patlen, pattern, patlen)) | 
|  | continue; | 
|  | if (namelen != patlen && name[namelen - patlen - 1] != '/') | 
|  | continue; | 
|  | match++; | 
|  | remote_ref = refs; | 
|  | } | 
|  | if (match == 0) | 
|  | return error("No remote branch matches %s", pattern); | 
|  | if (match != 1) | 
|  | return error("More than one remote branch matches %s", | 
|  | pattern); | 
|  |  | 
|  | /* | 
|  | * Remote HEAD must be a symref (not exactly foolproof; a remote | 
|  | * symlink to a symref will look like a symref) | 
|  | */ | 
|  | fetch_symref("HEAD", &symref, &head_oid); | 
|  | if (!symref) | 
|  | return error("Remote HEAD is not a symref"); | 
|  |  | 
|  | /* Remote branch must not be the remote HEAD */ | 
|  | for (i = 0; symref && i < MAXDEPTH; i++) { | 
|  | if (!strcmp(remote_ref->name, symref)) | 
|  | return error("Remote branch %s is the current HEAD", | 
|  | remote_ref->name); | 
|  | fetch_symref(symref, &symref, &head_oid); | 
|  | } | 
|  |  | 
|  | /* Run extra sanity checks if delete is not forced */ | 
|  | if (!force) { | 
|  | /* Remote HEAD must resolve to a known object */ | 
|  | if (symref) | 
|  | return error("Remote HEAD symrefs too deep"); | 
|  | if (is_null_oid(&head_oid)) | 
|  | return error("Unable to resolve remote HEAD"); | 
|  | if (!has_object_file(&head_oid)) | 
|  | return error("Remote HEAD resolves to object %s\nwhich does not exist locally, perhaps you need to fetch?", oid_to_hex(&head_oid)); | 
|  |  | 
|  | /* Remote branch must resolve to a known object */ | 
|  | if (is_null_oid(&remote_ref->old_oid)) | 
|  | return error("Unable to resolve remote branch %s", | 
|  | remote_ref->name); | 
|  | if (!has_object_file(&remote_ref->old_oid)) | 
|  | return error("Remote branch %s resolves to object %s\nwhich does not exist locally, perhaps you need to fetch?", remote_ref->name, oid_to_hex(&remote_ref->old_oid)); | 
|  |  | 
|  | /* Remote branch must be an ancestor of remote HEAD */ | 
|  | if (!verify_merge_base(&head_oid, remote_ref)) { | 
|  | return error("The branch '%s' is not an ancestor " | 
|  | "of your current HEAD.\n" | 
|  | "If you are sure you want to delete it," | 
|  | " run:\n\t'git http-push -D %s %s'", | 
|  | remote_ref->name, repo->url, pattern); | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Send delete request */ | 
|  | fprintf(stderr, "Removing remote branch '%s'\n", remote_ref->name); | 
|  | if (dry_run) | 
|  | return 0; | 
|  | url = xstrfmt("%s%s", repo->url, remote_ref->name); | 
|  | slot = get_active_slot(); | 
|  | slot->results = &results; | 
|  | curl_setup_http_get(slot->curl, url, DAV_DELETE); | 
|  | if (start_active_slot(slot)) { | 
|  | run_active_slot(slot); | 
|  | free(url); | 
|  | if (results.curl_result != CURLE_OK) | 
|  | return error("DELETE request failed (%d/%ld)", | 
|  | results.curl_result, results.http_code); | 
|  | } else { | 
|  | free(url); | 
|  | return error("Unable to start DELETE request"); | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void run_request_queue(void) | 
|  | { | 
|  | #ifdef USE_CURL_MULTI | 
|  | is_running_queue = 1; | 
|  | fill_active_slots(); | 
|  | add_fill_function(NULL, fill_active_slot); | 
|  | #endif | 
|  | do { | 
|  | finish_all_active_slots(); | 
|  | #ifdef USE_CURL_MULTI | 
|  | fill_active_slots(); | 
|  | #endif | 
|  | } while (request_queue_head && !aborted); | 
|  |  | 
|  | #ifdef USE_CURL_MULTI | 
|  | is_running_queue = 0; | 
|  | #endif | 
|  | } | 
|  |  | 
|  | int cmd_main(int argc, const char **argv) | 
|  | { | 
|  | struct transfer_request *request; | 
|  | struct transfer_request *next_request; | 
|  | struct refspec rs = REFSPEC_INIT_PUSH; | 
|  | struct remote_lock *ref_lock = NULL; | 
|  | struct remote_lock *info_ref_lock = NULL; | 
|  | struct rev_info revs; | 
|  | int delete_branch = 0; | 
|  | int force_delete = 0; | 
|  | int objects_to_send; | 
|  | int rc = 0; | 
|  | int i; | 
|  | int new_refs; | 
|  | struct ref *ref, *local_refs; | 
|  |  | 
|  | repo = xcalloc(1, sizeof(*repo)); | 
|  |  | 
|  | argv++; | 
|  | for (i = 1; i < argc; i++, argv++) { | 
|  | const char *arg = *argv; | 
|  |  | 
|  | if (*arg == '-') { | 
|  | if (!strcmp(arg, "--all")) { | 
|  | push_all = MATCH_REFS_ALL; | 
|  | continue; | 
|  | } | 
|  | if (!strcmp(arg, "--force")) { | 
|  | force_all = 1; | 
|  | continue; | 
|  | } | 
|  | if (!strcmp(arg, "--dry-run")) { | 
|  | dry_run = 1; | 
|  | continue; | 
|  | } | 
|  | if (!strcmp(arg, "--helper-status")) { | 
|  | helper_status = 1; | 
|  | continue; | 
|  | } | 
|  | if (!strcmp(arg, "--verbose")) { | 
|  | push_verbosely = 1; | 
|  | http_is_verbose = 1; | 
|  | continue; | 
|  | } | 
|  | if (!strcmp(arg, "-d")) { | 
|  | delete_branch = 1; | 
|  | continue; | 
|  | } | 
|  | if (!strcmp(arg, "-D")) { | 
|  | delete_branch = 1; | 
|  | force_delete = 1; | 
|  | continue; | 
|  | } | 
|  | if (!strcmp(arg, "-h")) | 
|  | usage(http_push_usage); | 
|  | } | 
|  | if (!repo->url) { | 
|  | char *path = strstr(arg, "//"); | 
|  | str_end_url_with_slash(arg, &repo->url); | 
|  | repo->path_len = strlen(repo->url); | 
|  | if (path) { | 
|  | repo->path = strchr(path+2, '/'); | 
|  | if (repo->path) | 
|  | repo->path_len = strlen(repo->path); | 
|  | } | 
|  | continue; | 
|  | } | 
|  | refspec_appendn(&rs, argv, argc - i); | 
|  | break; | 
|  | } | 
|  |  | 
|  | #ifndef USE_CURL_MULTI | 
|  | die("git-push is not available for http/https repository when not compiled with USE_CURL_MULTI"); | 
|  | #endif | 
|  |  | 
|  | if (!repo->url) | 
|  | usage(http_push_usage); | 
|  |  | 
|  | if (delete_branch && rs.nr != 1) | 
|  | die("You must specify only one branch name when deleting a remote branch"); | 
|  |  | 
|  | setup_git_directory(); | 
|  |  | 
|  | memset(remote_dir_exists, -1, 256); | 
|  |  | 
|  | http_init(NULL, repo->url, 1); | 
|  |  | 
|  | #ifdef USE_CURL_MULTI | 
|  | is_running_queue = 0; | 
|  | #endif | 
|  |  | 
|  | /* Verify DAV compliance/lock support */ | 
|  | if (!locking_available()) { | 
|  | rc = 1; | 
|  | goto cleanup; | 
|  | } | 
|  |  | 
|  | sigchain_push_common(remove_locks_on_signal); | 
|  |  | 
|  | /* Check whether the remote has server info files */ | 
|  | repo->can_update_info_refs = 0; | 
|  | repo->has_info_refs = remote_exists("info/refs"); | 
|  | repo->has_info_packs = remote_exists("objects/info/packs"); | 
|  | if (repo->has_info_refs) { | 
|  | info_ref_lock = lock_remote("info/refs", LOCK_TIME); | 
|  | if (info_ref_lock) | 
|  | repo->can_update_info_refs = 1; | 
|  | else { | 
|  | error("cannot lock existing info/refs"); | 
|  | rc = 1; | 
|  | goto cleanup; | 
|  | } | 
|  | } | 
|  | if (repo->has_info_packs) | 
|  | fetch_indices(); | 
|  |  | 
|  | /* Get a list of all local and remote heads to validate refspecs */ | 
|  | local_refs = get_local_heads(); | 
|  | fprintf(stderr, "Fetching remote heads...\n"); | 
|  | get_dav_remote_heads(); | 
|  | run_request_queue(); | 
|  |  | 
|  | /* Remove a remote branch if -d or -D was specified */ | 
|  | if (delete_branch) { | 
|  | const char *branch = rs.items[i].src; | 
|  | if (delete_remote_branch(branch, force_delete) == -1) { | 
|  | fprintf(stderr, "Unable to delete remote branch %s\n", | 
|  | branch); | 
|  | if (helper_status) | 
|  | printf("error %s cannot remove\n", branch); | 
|  | } | 
|  | goto cleanup; | 
|  | } | 
|  |  | 
|  | /* match them up */ | 
|  | if (match_push_refs(local_refs, &remote_refs, &rs, push_all)) { | 
|  | rc = -1; | 
|  | goto cleanup; | 
|  | } | 
|  | if (!remote_refs) { | 
|  | fprintf(stderr, "No refs in common and none specified; doing nothing.\n"); | 
|  | if (helper_status) | 
|  | printf("error null no match\n"); | 
|  | rc = 0; | 
|  | goto cleanup; | 
|  | } | 
|  |  | 
|  | new_refs = 0; | 
|  | for (ref = remote_refs; ref; ref = ref->next) { | 
|  | struct argv_array commit_argv = ARGV_ARRAY_INIT; | 
|  |  | 
|  | if (!ref->peer_ref) | 
|  | continue; | 
|  |  | 
|  | if (is_null_oid(&ref->peer_ref->new_oid)) { | 
|  | if (delete_remote_branch(ref->name, 1) == -1) { | 
|  | error("Could not remove %s", ref->name); | 
|  | if (helper_status) | 
|  | printf("error %s cannot remove\n", ref->name); | 
|  | rc = -4; | 
|  | } | 
|  | else if (helper_status) | 
|  | printf("ok %s\n", ref->name); | 
|  | new_refs++; | 
|  | continue; | 
|  | } | 
|  |  | 
|  | if (oideq(&ref->old_oid, &ref->peer_ref->new_oid)) { | 
|  | if (push_verbosely) | 
|  | fprintf(stderr, "'%s': up-to-date\n", ref->name); | 
|  | if (helper_status) | 
|  | printf("ok %s up to date\n", ref->name); | 
|  | continue; | 
|  | } | 
|  |  | 
|  | if (!force_all && | 
|  | !is_null_oid(&ref->old_oid) && | 
|  | !ref->force) { | 
|  | if (!has_object_file(&ref->old_oid) || | 
|  | !ref_newer(&ref->peer_ref->new_oid, | 
|  | &ref->old_oid)) { | 
|  | /* | 
|  | * We do not have the remote ref, or | 
|  | * we know that the remote ref is not | 
|  | * an ancestor of what we are trying to | 
|  | * push.  Either way this can be losing | 
|  | * commits at the remote end and likely | 
|  | * we were not up to date to begin with. | 
|  | */ | 
|  | error("remote '%s' is not an ancestor of\n" | 
|  | "local '%s'.\n" | 
|  | "Maybe you are not up-to-date and " | 
|  | "need to pull first?", | 
|  | ref->name, | 
|  | ref->peer_ref->name); | 
|  | if (helper_status) | 
|  | printf("error %s non-fast forward\n", ref->name); | 
|  | rc = -2; | 
|  | continue; | 
|  | } | 
|  | } | 
|  | oidcpy(&ref->new_oid, &ref->peer_ref->new_oid); | 
|  | new_refs++; | 
|  |  | 
|  | fprintf(stderr, "updating '%s'", ref->name); | 
|  | if (strcmp(ref->name, ref->peer_ref->name)) | 
|  | fprintf(stderr, " using '%s'", ref->peer_ref->name); | 
|  | fprintf(stderr, "\n  from %s\n  to   %s\n", | 
|  | oid_to_hex(&ref->old_oid), oid_to_hex(&ref->new_oid)); | 
|  | if (dry_run) { | 
|  | if (helper_status) | 
|  | printf("ok %s\n", ref->name); | 
|  | continue; | 
|  | } | 
|  |  | 
|  | /* Lock remote branch ref */ | 
|  | ref_lock = lock_remote(ref->name, LOCK_TIME); | 
|  | if (ref_lock == NULL) { | 
|  | fprintf(stderr, "Unable to lock remote branch %s\n", | 
|  | ref->name); | 
|  | if (helper_status) | 
|  | printf("error %s lock error\n", ref->name); | 
|  | rc = 1; | 
|  | continue; | 
|  | } | 
|  |  | 
|  | /* Set up revision info for this refspec */ | 
|  | argv_array_push(&commit_argv, ""); /* ignored */ | 
|  | argv_array_push(&commit_argv, "--objects"); | 
|  | argv_array_push(&commit_argv, oid_to_hex(&ref->new_oid)); | 
|  | if (!push_all && !is_null_oid(&ref->old_oid)) | 
|  | argv_array_pushf(&commit_argv, "^%s", | 
|  | oid_to_hex(&ref->old_oid)); | 
|  | repo_init_revisions(the_repository, &revs, setup_git_directory()); | 
|  | setup_revisions(commit_argv.argc, commit_argv.argv, &revs, NULL); | 
|  | revs.edge_hint = 0; /* just in case */ | 
|  |  | 
|  | /* Generate a list of objects that need to be pushed */ | 
|  | pushing = 0; | 
|  | if (prepare_revision_walk(&revs)) | 
|  | die("revision walk setup failed"); | 
|  | mark_edges_uninteresting(&revs, NULL, 0); | 
|  | objects_to_send = get_delta(&revs, ref_lock); | 
|  | finish_all_active_slots(); | 
|  |  | 
|  | /* Push missing objects to remote, this would be a | 
|  | convenient time to pack them first if appropriate. */ | 
|  | pushing = 1; | 
|  | if (objects_to_send) | 
|  | fprintf(stderr, "    sending %d objects\n", | 
|  | objects_to_send); | 
|  |  | 
|  | run_request_queue(); | 
|  |  | 
|  | /* Update the remote branch if all went well */ | 
|  | if (aborted || !update_remote(ref->new_oid.hash, ref_lock)) | 
|  | rc = 1; | 
|  |  | 
|  | if (!rc) | 
|  | fprintf(stderr, "    done\n"); | 
|  | if (helper_status) | 
|  | printf("%s %s\n", !rc ? "ok" : "error", ref->name); | 
|  | unlock_remote(ref_lock); | 
|  | check_locks(); | 
|  | argv_array_clear(&commit_argv); | 
|  | } | 
|  |  | 
|  | /* Update remote server info if appropriate */ | 
|  | if (repo->has_info_refs && new_refs) { | 
|  | if (info_ref_lock && repo->can_update_info_refs) { | 
|  | fprintf(stderr, "Updating remote server info\n"); | 
|  | if (!dry_run) | 
|  | update_remote_info_refs(info_ref_lock); | 
|  | } else { | 
|  | fprintf(stderr, "Unable to update server info\n"); | 
|  | } | 
|  | } | 
|  |  | 
|  | cleanup: | 
|  | if (info_ref_lock) | 
|  | unlock_remote(info_ref_lock); | 
|  | free(repo); | 
|  |  | 
|  | http_cleanup(); | 
|  |  | 
|  | request = request_queue_head; | 
|  | while (request != NULL) { | 
|  | next_request = request->next; | 
|  | release_request(request); | 
|  | request = next_request; | 
|  | } | 
|  |  | 
|  | return rc; | 
|  | } |