|  | #ifndef READ_CACHE_LL_H | 
|  | #define READ_CACHE_LL_H | 
|  |  | 
|  | #include "hash.h" | 
|  | #include "hashmap.h" | 
|  | #include "statinfo.h" | 
|  |  | 
|  | /* | 
|  | * Basic data structures for the directory cache | 
|  | */ | 
|  |  | 
|  | #define CACHE_SIGNATURE 0x44495243	/* "DIRC" */ | 
|  | struct cache_header { | 
|  | uint32_t hdr_signature; | 
|  | uint32_t hdr_version; | 
|  | uint32_t hdr_entries; | 
|  | }; | 
|  |  | 
|  | #define INDEX_FORMAT_LB 2 | 
|  | #define INDEX_FORMAT_UB 4 | 
|  |  | 
|  | struct cache_entry { | 
|  | struct hashmap_entry ent; | 
|  | struct stat_data ce_stat_data; | 
|  | unsigned int ce_mode; | 
|  | unsigned int ce_flags; | 
|  | unsigned int mem_pool_allocated; | 
|  | unsigned int ce_namelen; | 
|  | unsigned int index;	/* for link extension */ | 
|  | struct object_id oid; | 
|  | char name[FLEX_ARRAY]; /* more */ | 
|  | }; | 
|  |  | 
|  | #define CE_STAGEMASK (0x3000) | 
|  | #define CE_EXTENDED  (0x4000) | 
|  | #define CE_VALID     (0x8000) | 
|  | #define CE_STAGESHIFT 12 | 
|  |  | 
|  | /* | 
|  | * Range 0xFFFF0FFF in ce_flags is divided into | 
|  | * two parts: in-memory flags and on-disk ones. | 
|  | * Flags in CE_EXTENDED_FLAGS will get saved on-disk | 
|  | * if you want to save a new flag, add it in | 
|  | * CE_EXTENDED_FLAGS | 
|  | * | 
|  | * In-memory only flags | 
|  | */ | 
|  | #define CE_UPDATE            (1 << 16) | 
|  | #define CE_REMOVE            (1 << 17) | 
|  | #define CE_UPTODATE          (1 << 18) | 
|  | #define CE_ADDED             (1 << 19) | 
|  |  | 
|  | #define CE_HASHED            (1 << 20) | 
|  | #define CE_FSMONITOR_VALID   (1 << 21) | 
|  | #define CE_WT_REMOVE         (1 << 22) /* remove in work directory */ | 
|  | #define CE_CONFLICTED        (1 << 23) | 
|  |  | 
|  | #define CE_UNPACKED          (1 << 24) | 
|  | #define CE_NEW_SKIP_WORKTREE (1 << 25) | 
|  |  | 
|  | /* used to temporarily mark paths matched by pathspecs */ | 
|  | #define CE_MATCHED           (1 << 26) | 
|  |  | 
|  | #define CE_UPDATE_IN_BASE    (1 << 27) | 
|  | #define CE_STRIP_NAME        (1 << 28) | 
|  |  | 
|  | /* | 
|  | * Extended on-disk flags | 
|  | */ | 
|  | #define CE_INTENT_TO_ADD     (1 << 29) | 
|  | #define CE_SKIP_WORKTREE     (1 << 30) | 
|  | /* CE_EXTENDED2 is for future extension */ | 
|  | #define CE_EXTENDED2         (1U << 31) | 
|  |  | 
|  | #define CE_EXTENDED_FLAGS (CE_INTENT_TO_ADD | CE_SKIP_WORKTREE) | 
|  |  | 
|  | /* | 
|  | * Safeguard to avoid saving wrong flags: | 
|  | *  - CE_EXTENDED2 won't get saved until its semantic is known | 
|  | *  - Bits in 0x0000FFFF have been saved in ce_flags already | 
|  | *  - Bits in 0x003F0000 are currently in-memory flags | 
|  | */ | 
|  | #if CE_EXTENDED_FLAGS & 0x803FFFFF | 
|  | #error "CE_EXTENDED_FLAGS out of range" | 
|  | #endif | 
|  |  | 
|  | /* Forward structure decls */ | 
|  | struct pathspec; | 
|  | struct tree; | 
|  |  | 
|  | /* | 
|  | * Copy the sha1 and stat state of a cache entry from one to | 
|  | * another. But we never change the name, or the hash state! | 
|  | */ | 
|  | static inline void copy_cache_entry(struct cache_entry *dst, | 
|  | const struct cache_entry *src) | 
|  | { | 
|  | unsigned int state = dst->ce_flags & CE_HASHED; | 
|  | int mem_pool_allocated = dst->mem_pool_allocated; | 
|  |  | 
|  | /* Don't copy hash chain and name */ | 
|  | memcpy(&dst->ce_stat_data, &src->ce_stat_data, | 
|  | offsetof(struct cache_entry, name) - | 
|  | offsetof(struct cache_entry, ce_stat_data)); | 
|  |  | 
|  | /* Restore the hash state */ | 
|  | dst->ce_flags = (dst->ce_flags & ~CE_HASHED) | state; | 
|  |  | 
|  | /* Restore the mem_pool_allocated flag */ | 
|  | dst->mem_pool_allocated = mem_pool_allocated; | 
|  | } | 
|  |  | 
|  | static inline unsigned create_ce_flags(unsigned stage) | 
|  | { | 
|  | return (stage << CE_STAGESHIFT); | 
|  | } | 
|  |  | 
|  | #define ce_namelen(ce) ((ce)->ce_namelen) | 
|  | #define ce_size(ce) cache_entry_size(ce_namelen(ce)) | 
|  | #define ce_stage(ce) ((CE_STAGEMASK & (ce)->ce_flags) >> CE_STAGESHIFT) | 
|  | #define ce_uptodate(ce) ((ce)->ce_flags & CE_UPTODATE) | 
|  | #define ce_skip_worktree(ce) ((ce)->ce_flags & CE_SKIP_WORKTREE) | 
|  | #define ce_mark_uptodate(ce) ((ce)->ce_flags |= CE_UPTODATE) | 
|  | #define ce_intent_to_add(ce) ((ce)->ce_flags & CE_INTENT_TO_ADD) | 
|  |  | 
|  | #define cache_entry_size(len) (offsetof(struct cache_entry,name) + (len) + 1) | 
|  |  | 
|  | #define SOMETHING_CHANGED	(1 << 0) /* unclassified changes go here */ | 
|  | #define CE_ENTRY_CHANGED	(1 << 1) | 
|  | #define CE_ENTRY_REMOVED	(1 << 2) | 
|  | #define CE_ENTRY_ADDED		(1 << 3) | 
|  | #define RESOLVE_UNDO_CHANGED	(1 << 4) | 
|  | #define CACHE_TREE_CHANGED	(1 << 5) | 
|  | #define SPLIT_INDEX_ORDERED	(1 << 6) | 
|  | #define UNTRACKED_CHANGED	(1 << 7) | 
|  | #define FSMONITOR_CHANGED	(1 << 8) | 
|  |  | 
|  | struct split_index; | 
|  | struct untracked_cache; | 
|  | struct progress; | 
|  | struct pattern_list; | 
|  |  | 
|  | enum sparse_index_mode { | 
|  | /* | 
|  | * There are no sparse directories in the index at all. | 
|  | * | 
|  | * Repositories that don't use cone-mode sparse-checkout will | 
|  | * always have their indexes in this mode. | 
|  | */ | 
|  | INDEX_EXPANDED = 0, | 
|  |  | 
|  | /* | 
|  | * The index has already been collapsed to sparse directories | 
|  | * wherever possible. | 
|  | */ | 
|  | INDEX_COLLAPSED, | 
|  |  | 
|  | /* | 
|  | * The sparse directories that exist are outside the | 
|  | * sparse-checkout boundary, but it is possible that some file | 
|  | * entries could collapse to sparse directory entries. | 
|  | */ | 
|  | INDEX_PARTIALLY_SPARSE, | 
|  | }; | 
|  |  | 
|  | struct index_state { | 
|  | struct cache_entry **cache; | 
|  | unsigned int version; | 
|  | unsigned int cache_nr, cache_alloc, cache_changed; | 
|  | struct string_list *resolve_undo; | 
|  | struct cache_tree *cache_tree; | 
|  | struct split_index *split_index; | 
|  | struct cache_time timestamp; | 
|  | unsigned name_hash_initialized : 1, | 
|  | initialized : 1, | 
|  | drop_cache_tree : 1, | 
|  | updated_workdir : 1, | 
|  | updated_skipworktree : 1, | 
|  | fsmonitor_has_run_once : 1; | 
|  | enum sparse_index_mode sparse_index; | 
|  | struct hashmap name_hash; | 
|  | struct hashmap dir_hash; | 
|  | struct object_id oid; | 
|  | struct untracked_cache *untracked; | 
|  | char *fsmonitor_last_update; | 
|  | struct ewah_bitmap *fsmonitor_dirty; | 
|  | struct mem_pool *ce_mem_pool; | 
|  | struct progress *progress; | 
|  | struct repository *repo; | 
|  | struct pattern_list *sparse_checkout_patterns; | 
|  | }; | 
|  |  | 
|  | /** | 
|  | * A "struct index_state istate" must be initialized with | 
|  | * INDEX_STATE_INIT or the corresponding index_state_init(). | 
|  | * | 
|  | * If the variable won't be used again, use release_index() to free() | 
|  | * its resources. If it needs to be used again use discard_index(), | 
|  | * which does the same thing, but will use index_state_init() at | 
|  | * the end. The discard_index() will use its own "istate->repo" as the | 
|  | * "r" argument to index_state_init() in that case. | 
|  | */ | 
|  | #define INDEX_STATE_INIT(r) { \ | 
|  | .repo = (r), \ | 
|  | } | 
|  | void index_state_init(struct index_state *istate, struct repository *r); | 
|  | void release_index(struct index_state *istate); | 
|  |  | 
|  | /* Cache entry creation and cleanup */ | 
|  |  | 
|  | /* | 
|  | * Create cache_entry intended for use in the specified index. Caller | 
|  | * is responsible for discarding the cache_entry with | 
|  | * `discard_cache_entry`. | 
|  | */ | 
|  | struct cache_entry *make_cache_entry(struct index_state *istate, | 
|  | unsigned int mode, | 
|  | const struct object_id *oid, | 
|  | const char *path, | 
|  | int stage, | 
|  | unsigned int refresh_options); | 
|  |  | 
|  | struct cache_entry *make_empty_cache_entry(struct index_state *istate, | 
|  | size_t name_len); | 
|  |  | 
|  | /* | 
|  | * Create a cache_entry that is not intended to be added to an index. If | 
|  | * `ce_mem_pool` is not NULL, the entry is allocated within the given memory | 
|  | * pool. Caller is responsible for discarding "loose" entries with | 
|  | * `discard_cache_entry()` and the memory pool with | 
|  | * `mem_pool_discard(ce_mem_pool, should_validate_cache_entries())`. | 
|  | */ | 
|  | struct cache_entry *make_transient_cache_entry(unsigned int mode, | 
|  | const struct object_id *oid, | 
|  | const char *path, | 
|  | int stage, | 
|  | struct mem_pool *ce_mem_pool); | 
|  |  | 
|  | struct cache_entry *make_empty_transient_cache_entry(size_t len, | 
|  | struct mem_pool *ce_mem_pool); | 
|  |  | 
|  | /* | 
|  | * Discard cache entry. | 
|  | */ | 
|  | void discard_cache_entry(struct cache_entry *ce); | 
|  |  | 
|  | /* | 
|  | * Check configuration if we should perform extra validation on cache | 
|  | * entries. | 
|  | */ | 
|  | int should_validate_cache_entries(void); | 
|  |  | 
|  | /* | 
|  | * Duplicate a cache_entry. Allocate memory for the new entry from a | 
|  | * memory_pool. Takes into account cache_entry fields that are meant | 
|  | * for managing the underlying memory allocation of the cache_entry. | 
|  | */ | 
|  | struct cache_entry *dup_cache_entry(const struct cache_entry *ce, struct index_state *istate); | 
|  |  | 
|  | /* | 
|  | * Validate the cache entries in the index.  This is an internal | 
|  | * consistency check that the cache_entry structs are allocated from | 
|  | * the expected memory pool. | 
|  | */ | 
|  | void validate_cache_entries(const struct index_state *istate); | 
|  |  | 
|  | /* | 
|  | * Bulk prefetch all missing cache entries that are not GITLINKs and that match | 
|  | * the given predicate. This function should only be called if | 
|  | * repo_has_promisor_remote() returns true. | 
|  | */ | 
|  | typedef int (*must_prefetch_predicate)(const struct cache_entry *); | 
|  | void prefetch_cache_entries(const struct index_state *istate, | 
|  | must_prefetch_predicate must_prefetch); | 
|  |  | 
|  | /* Initialize and use the cache information */ | 
|  | struct lock_file; | 
|  | int do_read_index(struct index_state *istate, const char *path, | 
|  | int must_exist); /* for testting only! */ | 
|  | int read_index_from(struct index_state *, const char *path, | 
|  | const char *gitdir); | 
|  | int is_index_unborn(struct index_state *); | 
|  |  | 
|  | /* For use with `write_locked_index()`. */ | 
|  | #define COMMIT_LOCK		(1 << 0) | 
|  | #define SKIP_IF_UNCHANGED	(1 << 1) | 
|  |  | 
|  | /* | 
|  | * Write the index while holding an already-taken lock. Close the lock, | 
|  | * and if `COMMIT_LOCK` is given, commit it. | 
|  | * | 
|  | * Unless a split index is in use, write the index into the lockfile. | 
|  | * | 
|  | * With a split index, write the shared index to a temporary file, | 
|  | * adjust its permissions and rename it into place, then write the | 
|  | * split index to the lockfile. If the temporary file for the shared | 
|  | * index cannot be created, fall back to the behavior described in | 
|  | * the previous paragraph. | 
|  | * | 
|  | * With `COMMIT_LOCK`, the lock is always committed or rolled back. | 
|  | * Without it, the lock is closed, but neither committed nor rolled | 
|  | * back. | 
|  | * | 
|  | * If `SKIP_IF_UNCHANGED` is given and the index is unchanged, nothing | 
|  | * is written (and the lock is rolled back if `COMMIT_LOCK` is given). | 
|  | */ | 
|  | int write_locked_index(struct index_state *, struct lock_file *lock, unsigned flags); | 
|  |  | 
|  | void discard_index(struct index_state *); | 
|  | void move_index_extensions(struct index_state *dst, struct index_state *src); | 
|  | int unmerged_index(const struct index_state *); | 
|  |  | 
|  | /** | 
|  | * Returns 1 if istate differs from tree, 0 otherwise.  If tree is NULL, | 
|  | * compares istate to HEAD.  If tree is NULL and on an unborn branch, | 
|  | * returns 1 if there are entries in istate, 0 otherwise.  If an strbuf is | 
|  | * provided, the space-separated list of files that differ will be appended | 
|  | * to it. | 
|  | */ | 
|  | int repo_index_has_changes(struct repository *repo, | 
|  | struct tree *tree, | 
|  | struct strbuf *sb); | 
|  |  | 
|  | int verify_path(const char *path, unsigned mode); | 
|  | int strcmp_offset(const char *s1, const char *s2, size_t *first_change); | 
|  |  | 
|  | /* | 
|  | * Searches for an entry defined by name and namelen in the given index. | 
|  | * If the return value is positive (including 0) it is the position of an | 
|  | * exact match. If the return value is negative, the negated value minus 1 | 
|  | * is the position where the entry would be inserted. | 
|  | * Example: The current index consists of these files and its stages: | 
|  | * | 
|  | *   b#0, d#0, f#1, f#3 | 
|  | * | 
|  | * index_name_pos(&index, "a", 1) -> -1 | 
|  | * index_name_pos(&index, "b", 1) ->  0 | 
|  | * index_name_pos(&index, "c", 1) -> -2 | 
|  | * index_name_pos(&index, "d", 1) ->  1 | 
|  | * index_name_pos(&index, "e", 1) -> -3 | 
|  | * index_name_pos(&index, "f", 1) -> -3 | 
|  | * index_name_pos(&index, "g", 1) -> -5 | 
|  | */ | 
|  | int index_name_pos(struct index_state *, const char *name, int namelen); | 
|  |  | 
|  | /* | 
|  | * Like index_name_pos, returns the position of an entry of the given name in | 
|  | * the index if one exists, otherwise returns a negative value where the negated | 
|  | * value minus 1 is the position where the index entry would be inserted. Unlike | 
|  | * index_name_pos, however, a sparse index is not expanded to find an entry | 
|  | * inside a sparse directory. | 
|  | */ | 
|  | int index_name_pos_sparse(struct index_state *, const char *name, int namelen); | 
|  |  | 
|  | /* | 
|  | * Determines whether an entry with the given name exists within the | 
|  | * given index. The return value is 1 if an exact match is found, otherwise | 
|  | * it is 0. Note that, unlike index_name_pos, this function does not expand | 
|  | * the index if it is sparse. If an item exists within the full index but it | 
|  | * is contained within a sparse directory (and not in the sparse index), 0 is | 
|  | * returned. | 
|  | */ | 
|  | int index_entry_exists(struct index_state *, const char *name, int namelen); | 
|  |  | 
|  | /* | 
|  | * Some functions return the negative complement of an insert position when a | 
|  | * precise match was not found but a position was found where the entry would | 
|  | * need to be inserted. This helper protects that logic from any integer | 
|  | * underflow. | 
|  | */ | 
|  | static inline int index_pos_to_insert_pos(uintmax_t pos) | 
|  | { | 
|  | if (pos > INT_MAX) | 
|  | die("overflow: -1 - %"PRIuMAX, pos); | 
|  | return -1 - (int)pos; | 
|  | } | 
|  |  | 
|  | #define ADD_CACHE_OK_TO_ADD 1		/* Ok to add */ | 
|  | #define ADD_CACHE_OK_TO_REPLACE 2	/* Ok to replace file/directory */ | 
|  | #define ADD_CACHE_SKIP_DFCHECK 4	/* Ok to skip DF conflict checks */ | 
|  | #define ADD_CACHE_JUST_APPEND 8		/* Append only */ | 
|  | #define ADD_CACHE_NEW_ONLY 16		/* Do not replace existing ones */ | 
|  | #define ADD_CACHE_KEEP_CACHE_TREE 32	/* Do not invalidate cache-tree */ | 
|  | #define ADD_CACHE_RENORMALIZE 64        /* Pass along HASH_RENORMALIZE */ | 
|  | int add_index_entry(struct index_state *, struct cache_entry *ce, int option); | 
|  | void rename_index_entry_at(struct index_state *, int pos, const char *new_name); | 
|  |  | 
|  | /* Remove entry, return true if there are more entries to go. */ | 
|  | int remove_index_entry_at(struct index_state *, int pos); | 
|  |  | 
|  | void remove_marked_cache_entries(struct index_state *istate, int invalidate); | 
|  | int remove_file_from_index(struct index_state *, const char *path); | 
|  | #define ADD_CACHE_VERBOSE 1 | 
|  | #define ADD_CACHE_PRETEND 2 | 
|  | #define ADD_CACHE_IGNORE_ERRORS	4 | 
|  | #define ADD_CACHE_IGNORE_REMOVAL 8 | 
|  | #define ADD_CACHE_INTENT 16 | 
|  | /* | 
|  | * These two are used to add the contents of the file at path | 
|  | * to the index, marking the working tree up-to-date by storing | 
|  | * the cached stat info in the resulting cache entry.  A caller | 
|  | * that has already run lstat(2) on the path can call | 
|  | * add_to_index(), and all others can call add_file_to_index(); | 
|  | * the latter will do necessary lstat(2) internally before | 
|  | * calling the former. | 
|  | */ | 
|  | int add_to_index(struct index_state *, const char *path, struct stat *, int flags); | 
|  | int add_file_to_index(struct index_state *, const char *path, int flags); | 
|  |  | 
|  | int chmod_index_entry(struct index_state *, struct cache_entry *ce, char flip); | 
|  | int ce_same_name(const struct cache_entry *a, const struct cache_entry *b); | 
|  | void set_object_name_for_intent_to_add_entry(struct cache_entry *ce); | 
|  | int index_name_is_other(struct index_state *, const char *, int); | 
|  | void *read_blob_data_from_index(struct index_state *, const char *, unsigned long *); | 
|  |  | 
|  | /* do stat comparison even if CE_VALID is true */ | 
|  | #define CE_MATCH_IGNORE_VALID		01 | 
|  | /* do not check the contents but report dirty on racily-clean entries */ | 
|  | #define CE_MATCH_RACY_IS_DIRTY		02 | 
|  | /* do stat comparison even if CE_SKIP_WORKTREE is true */ | 
|  | #define CE_MATCH_IGNORE_SKIP_WORKTREE	04 | 
|  | /* ignore non-existent files during stat update  */ | 
|  | #define CE_MATCH_IGNORE_MISSING		0x08 | 
|  | /* enable stat refresh */ | 
|  | #define CE_MATCH_REFRESH		0x10 | 
|  | /* don't refresh_fsmonitor state or do stat comparison even if CE_FSMONITOR_VALID is true */ | 
|  | #define CE_MATCH_IGNORE_FSMONITOR 0X20 | 
|  | int is_racy_timestamp(const struct index_state *istate, | 
|  | const struct cache_entry *ce); | 
|  | int has_racy_timestamp(struct index_state *istate); | 
|  | int ie_match_stat(struct index_state *, const struct cache_entry *, struct stat *, unsigned int); | 
|  | int ie_modified(struct index_state *, const struct cache_entry *, struct stat *, unsigned int); | 
|  |  | 
|  | int match_stat_data_racy(const struct index_state *istate, | 
|  | const struct stat_data *sd, struct stat *st); | 
|  |  | 
|  | void fill_stat_cache_info(struct index_state *istate, struct cache_entry *ce, struct stat *st); | 
|  |  | 
|  | /* | 
|  | * Fill members of st by members of sd enough to convince match_stat() | 
|  | * to consider that they match.  It should be usable as a replacement | 
|  | * for lstat() for a tracked path that is known to be up-to-date via | 
|  | * some out-of-line means (like fsmonitor). | 
|  | */ | 
|  | int fake_lstat(const struct cache_entry *ce, struct stat *st); | 
|  |  | 
|  | #define REFRESH_REALLY                   (1 << 0) /* ignore_valid */ | 
|  | #define REFRESH_UNMERGED                 (1 << 1) /* allow unmerged */ | 
|  | #define REFRESH_QUIET                    (1 << 2) /* be quiet about it */ | 
|  | #define REFRESH_IGNORE_MISSING           (1 << 3) /* ignore non-existent */ | 
|  | #define REFRESH_IGNORE_SUBMODULES        (1 << 4) /* ignore submodules */ | 
|  | #define REFRESH_IN_PORCELAIN             (1 << 5) /* user friendly output, not "needs update" */ | 
|  | #define REFRESH_PROGRESS                 (1 << 6) /* show progress bar if stderr is tty */ | 
|  | #define REFRESH_IGNORE_SKIP_WORKTREE     (1 << 7) /* ignore skip_worktree entries */ | 
|  | int refresh_index(struct index_state *, unsigned int flags, const struct pathspec *pathspec, char *seen, const char *header_msg); | 
|  | /* | 
|  | * Refresh the index and write it to disk. | 
|  | * | 
|  | * 'refresh_flags' is passed directly to 'refresh_index()', while | 
|  | * 'COMMIT_LOCK | write_flags' is passed to 'write_locked_index()', so | 
|  | * the lockfile is always either committed or rolled back. | 
|  | * | 
|  | * If 'gentle' is passed, errors locking the index are ignored. | 
|  | * | 
|  | * Return 1 if refreshing the index returns an error, -1 if writing | 
|  | * the index to disk fails, 0 on success. | 
|  | * | 
|  | * Note that if refreshing the index returns an error, we still write | 
|  | * out the index (unless locking fails). | 
|  | */ | 
|  | int repo_refresh_and_write_index(struct repository*, unsigned int refresh_flags, unsigned int write_flags, int gentle, const struct pathspec *, char *seen, const char *header_msg); | 
|  |  | 
|  | struct cache_entry *refresh_cache_entry(struct index_state *, struct cache_entry *, unsigned int); | 
|  |  | 
|  | void set_alternate_index_output(const char *); | 
|  |  | 
|  | extern int verify_index_checksum; | 
|  | extern int verify_ce_order; | 
|  |  | 
|  | int cmp_cache_name_compare(const void *a_, const void *b_); | 
|  |  | 
|  | int add_files_to_cache(struct repository *repo, const char *prefix, | 
|  | const struct pathspec *pathspec, char *ps_matched, | 
|  | int include_sparse, int flags); | 
|  |  | 
|  | void overlay_tree_on_index(struct index_state *istate, | 
|  | const char *tree_name, const char *prefix); | 
|  |  | 
|  | #endif /* READ_CACHE_LL_H */ |