]> git.ipfire.org Git - thirdparty/git.git/blame - bulk-checkin.c
t5351: avoid relying on `core.fsyncMethod = batch` to be supported
[thirdparty/git.git] / bulk-checkin.c
CommitLineData
568508e7
JH
1/*
2 * Copyright (c) 2011, Google Inc.
3 */
1c4b6604 4#include "cache.h"
568508e7 5#include "bulk-checkin.h"
c0f4752e 6#include "lockfile.h"
a49d2834 7#include "repository.h"
568508e7
JH
8#include "csum-file.h"
9#include "pack.h"
58892711 10#include "strbuf.h"
c0f4752e
NS
11#include "string-list.h"
12#include "tmp-objdir.h"
0abe14f6 13#include "packfile.h"
cbd53a21 14#include "object-store.h"
568508e7 15
2c23d1b4 16static int odb_transaction_nesting;
568508e7 17
c0f4752e
NS
18static struct tmp_objdir *bulk_fsync_objdir;
19
897c9e25 20static struct bulk_checkin_packfile {
568508e7 21 char *pack_tmp_name;
98a3beab 22 struct hashfile *f;
568508e7
JH
23 off_t offset;
24 struct pack_idx_option pack_idx_opts;
25
26 struct pack_idx_entry **written;
27 uint32_t alloc_written;
28 uint32_t nr_written;
897c9e25 29} bulk_checkin_packfile;
568508e7 30
2ec02dd5
ÆAB
31static void finish_tmp_packfile(struct strbuf *basename,
32 const char *pack_tmp_name,
33 struct pack_idx_entry **written_list,
34 uint32_t nr_written,
35 struct pack_idx_option *pack_idx_opts,
36 unsigned char hash[])
37{
38 char *idx_tmp_name = NULL;
39
40 stage_tmp_packfiles(basename, pack_tmp_name, written_list, nr_written,
1c573cdd 41 NULL, pack_idx_opts, hash, &idx_tmp_name);
2ec02dd5
ÆAB
42 rename_tmp_packfile_idx(basename, &idx_tmp_name);
43
44 free(idx_tmp_name);
45}
46
897c9e25 47static void flush_bulk_checkin_packfile(struct bulk_checkin_packfile *state)
568508e7 48{
ae44b5a4 49 unsigned char hash[GIT_MAX_RAWSZ];
58892711 50 struct strbuf packname = STRBUF_INIT;
568508e7
JH
51 int i;
52
53 if (!state->f)
54 return;
55
56 if (state->nr_written == 0) {
57 close(state->f->fd);
58 unlink(state->pack_tmp_name);
59 goto clear_exit;
60 } else if (state->nr_written == 1) {
020406ea
NS
61 finalize_hashfile(state->f, hash, FSYNC_COMPONENT_PACK,
62 CSUM_HASH_IN_STREAM | CSUM_FSYNC | CSUM_CLOSE);
568508e7 63 } else {
020406ea 64 int fd = finalize_hashfile(state->f, hash, FSYNC_COMPONENT_PACK, 0);
ae44b5a4
TB
65 fixup_pack_header_footer(fd, hash, state->pack_tmp_name,
66 state->nr_written, hash,
568508e7
JH
67 state->offset);
68 close(fd);
69 }
70
66833f0e
ÆAB
71 strbuf_addf(&packname, "%s/pack/pack-%s.", get_object_directory(),
72 hash_to_hex(hash));
58892711 73 finish_tmp_packfile(&packname, state->pack_tmp_name,
568508e7 74 state->written, state->nr_written,
ae44b5a4 75 &state->pack_idx_opts, hash);
568508e7
JH
76 for (i = 0; i < state->nr_written; i++)
77 free(state->written[i]);
78
79clear_exit:
80 free(state->written);
81 memset(state, 0, sizeof(*state));
82
58892711 83 strbuf_release(&packname);
568508e7 84 /* Make objects we just wrote available to ourselves */
a49d2834 85 reprepare_packed_git(the_repository);
568508e7
JH
86}
87
c0f4752e
NS
88/*
89 * Cleanup after batch-mode fsync_object_files.
90 */
91static void flush_batch_fsync(void)
92{
93 struct strbuf temp_path = STRBUF_INIT;
94 struct tempfile *temp;
95
96 if (!bulk_fsync_objdir)
97 return;
98
99 /*
100 * Issue a full hardware flush against a temporary file to ensure
101 * that all objects are durable before any renames occur. The code in
102 * fsync_loose_object_bulk_checkin has already issued a writeout
103 * request, but it has not flushed any writeback cache in the storage
104 * hardware or any filesystem logs. This fsync call acts as a barrier
105 * to ensure that the data in each new object file is durable before
106 * the final name is visible.
107 */
108 strbuf_addf(&temp_path, "%s/bulk_fsync_XXXXXX", get_object_directory());
109 temp = xmks_tempfile(temp_path.buf);
110 fsync_or_die(get_tempfile_fd(temp), get_tempfile_path(temp));
111 delete_tempfile(&temp);
112 strbuf_release(&temp_path);
113
114 /*
115 * Make the object files visible in the primary ODB after their data is
116 * fully durable.
117 */
118 tmp_objdir_migrate(bulk_fsync_objdir);
119 bulk_fsync_objdir = NULL;
120}
121
897c9e25 122static int already_written(struct bulk_checkin_packfile *state, struct object_id *oid)
568508e7
JH
123{
124 int i;
125
126 /* The object may already exist in the repository */
98374a07 127 if (has_object_file(oid))
568508e7
JH
128 return 1;
129
130 /* Might want to keep the list sorted */
131 for (i = 0; i < state->nr_written; i++)
4a7e27e9 132 if (oideq(&state->written[i]->oid, oid))
568508e7
JH
133 return 1;
134
135 /* This is a new object we need to keep */
136 return 0;
137}
138
139/*
140 * Read the contents from fd for size bytes, streaming it to the
141 * packfile in state while updating the hash in ctx. Signal a failure
142 * by returning a negative value when the resulting pack would exceed
143 * the pack size limit and this is not the first object in the pack,
144 * so that the caller can discard what we wrote from the current pack
145 * by truncating it and opening a new one. The caller will then call
146 * us again after rewinding the input fd.
147 *
148 * The already_hashed_to pointer is kept untouched by the caller to
149 * make sure we do not hash the same byte when we are called
150 * again. This way, the caller does not have to checkpoint its hash
151 * status before calling us just in case we ask it to call us again
152 * with a new pack.
153 */
897c9e25 154static int stream_to_pack(struct bulk_checkin_packfile *state,
f87e8137 155 git_hash_ctx *ctx, off_t *already_hashed_to,
568508e7
JH
156 int fd, size_t size, enum object_type type,
157 const char *path, unsigned flags)
158{
159 git_zstream s;
f96178c5 160 unsigned char ibuf[16384];
568508e7
JH
161 unsigned char obuf[16384];
162 unsigned hdrlen;
163 int status = Z_OK;
164 int write_object = (flags & HASH_WRITE_OBJECT);
165 off_t offset = 0;
166
568508e7
JH
167 git_deflate_init(&s, pack_compression_level);
168
7202a6fa 169 hdrlen = encode_in_pack_object_header(obuf, sizeof(obuf), type, size);
568508e7
JH
170 s.next_out = obuf + hdrlen;
171 s.avail_out = sizeof(obuf) - hdrlen;
172
173 while (status != Z_STREAM_END) {
568508e7
JH
174 if (size && !s.avail_in) {
175 ssize_t rsize = size < sizeof(ibuf) ? size : sizeof(ibuf);
41dcc4dc
JK
176 ssize_t read_result = read_in_full(fd, ibuf, rsize);
177 if (read_result < 0)
178 die_errno("failed to read from '%s'", path);
179 if (read_result != rsize)
568508e7
JH
180 die("failed to read %d bytes from '%s'",
181 (int)rsize, path);
182 offset += rsize;
183 if (*already_hashed_to < offset) {
184 size_t hsize = offset - *already_hashed_to;
185 if (rsize < hsize)
186 hsize = rsize;
187 if (hsize)
f87e8137 188 the_hash_algo->update_fn(ctx, ibuf, hsize);
568508e7
JH
189 *already_hashed_to = offset;
190 }
191 s.next_in = ibuf;
192 s.avail_in = rsize;
193 size -= rsize;
194 }
195
196 status = git_deflate(&s, size ? 0 : Z_FINISH);
197
198 if (!s.avail_out || status == Z_STREAM_END) {
199 if (write_object) {
200 size_t written = s.next_out - obuf;
201
202 /* would we bust the size limit? */
203 if (state->nr_written &&
204 pack_size_limit_cfg &&
205 pack_size_limit_cfg < state->offset + written) {
206 git_deflate_abort(&s);
207 return -1;
208 }
209
98a3beab 210 hashwrite(state->f, obuf, written);
568508e7
JH
211 state->offset += written;
212 }
213 s.next_out = obuf;
214 s.avail_out = sizeof(obuf);
215 }
216
217 switch (status) {
218 case Z_OK:
219 case Z_BUF_ERROR:
220 case Z_STREAM_END:
221 continue;
222 default:
223 die("unexpected deflate failure: %d", status);
224 }
225 }
226 git_deflate_end(&s);
227 return 0;
228}
229
230/* Lazily create backing packfile for the state */
897c9e25 231static void prepare_to_stream(struct bulk_checkin_packfile *state,
568508e7
JH
232 unsigned flags)
233{
234 if (!(flags & HASH_WRITE_OBJECT) || state->f)
235 return;
236
237 state->f = create_tmp_packfile(&state->pack_tmp_name);
238 reset_pack_idx_option(&state->pack_idx_opts);
239
240 /* Pretend we are going to write only one object */
241 state->offset = write_pack_header(state->f, 1);
242 if (!state->offset)
243 die_errno("unable to write pack header");
244}
245
897c9e25 246static int deflate_to_pack(struct bulk_checkin_packfile *state,
68ee6dfc 247 struct object_id *result_oid,
568508e7
JH
248 int fd, size_t size,
249 enum object_type type, const char *path,
250 unsigned flags)
251{
252 off_t seekback, already_hashed_to;
f87e8137 253 git_hash_ctx ctx;
568508e7
JH
254 unsigned char obuf[16384];
255 unsigned header_len;
7140414d 256 struct hashfile_checkpoint checkpoint = {0};
568508e7
JH
257 struct pack_idx_entry *idx = NULL;
258
259 seekback = lseek(fd, 0, SEEK_CUR);
260 if (seekback == (off_t) -1)
261 return error("cannot find the current offset");
262
b04cdea4
ÆAB
263 header_len = format_object_header((char *)obuf, sizeof(obuf),
264 type, size);
f87e8137 265 the_hash_algo->init_fn(&ctx);
266 the_hash_algo->update_fn(&ctx, obuf, header_len);
568508e7
JH
267
268 /* Note: idx is non-NULL when we are writing */
269 if ((flags & HASH_WRITE_OBJECT) != 0)
ca56dadb 270 CALLOC_ARRAY(idx, 1);
568508e7
JH
271
272 already_hashed_to = 0;
273
274 while (1) {
275 prepare_to_stream(state, flags);
276 if (idx) {
98a3beab 277 hashfile_checkpoint(state->f, &checkpoint);
568508e7
JH
278 idx->offset = state->offset;
279 crc32_begin(state->f);
280 }
281 if (!stream_to_pack(state, &ctx, &already_hashed_to,
282 fd, size, type, path, flags))
283 break;
284 /*
285 * Writing this object to the current pack will make
286 * it too big; we need to truncate it, start a new
287 * pack, and write into it.
288 */
289 if (!idx)
033abf97 290 BUG("should not happen");
98a3beab 291 hashfile_truncate(state->f, &checkpoint);
568508e7 292 state->offset = checkpoint.offset;
897c9e25 293 flush_bulk_checkin_packfile(state);
568508e7
JH
294 if (lseek(fd, seekback, SEEK_SET) == (off_t) -1)
295 return error("cannot seek back");
296 }
5951bf46 297 the_hash_algo->final_oid_fn(result_oid, &ctx);
568508e7
JH
298 if (!idx)
299 return 0;
300
301 idx->crc32 = crc32_end(state->f);
68ee6dfc 302 if (already_written(state, result_oid)) {
98a3beab 303 hashfile_truncate(state->f, &checkpoint);
568508e7
JH
304 state->offset = checkpoint.offset;
305 free(idx);
306 } else {
68ee6dfc 307 oidcpy(&idx->oid, result_oid);
568508e7
JH
308 ALLOC_GROW(state->written,
309 state->nr_written + 1,
310 state->alloc_written);
311 state->written[state->nr_written++] = idx;
312 }
313 return 0;
314}
315
c0f4752e
NS
316void prepare_loose_object_bulk_checkin(void)
317{
318 /*
319 * We lazily create the temporary object directory
320 * the first time an object might be added, since
321 * callers may not know whether any objects will be
322 * added at the time they call begin_odb_transaction.
323 */
324 if (!odb_transaction_nesting || bulk_fsync_objdir)
325 return;
326
327 bulk_fsync_objdir = tmp_objdir_create("bulk-fsync");
328 if (bulk_fsync_objdir)
329 tmp_objdir_replace_primary_odb(bulk_fsync_objdir, 0);
330}
331
332void fsync_loose_object_bulk_checkin(int fd, const char *filename)
333{
334 /*
335 * If we have an active ODB transaction, we issue a call that
336 * cleans the filesystem page cache but avoids a hardware flush
337 * command. Later on we will issue a single hardware flush
338 * before renaming the objects to their final names as part of
339 * flush_batch_fsync.
340 */
341 if (!bulk_fsync_objdir ||
342 git_fsync(fd, FSYNC_WRITEOUT_ONLY) < 0) {
ce50f1f3
JS
343 if (errno == ENOSYS)
344 warning(_("core.fsyncMethod = batch is unsupported on this platform"));
c0f4752e
NS
345 fsync_or_die(fd, filename);
346 }
347}
348
68ee6dfc 349int index_bulk_checkin(struct object_id *oid,
568508e7
JH
350 int fd, size_t size, enum object_type type,
351 const char *path, unsigned flags)
352{
897c9e25 353 int status = deflate_to_pack(&bulk_checkin_packfile, oid, fd, size, type,
568508e7 354 path, flags);
2c23d1b4 355 if (!odb_transaction_nesting)
897c9e25 356 flush_bulk_checkin_packfile(&bulk_checkin_packfile);
568508e7
JH
357 return status;
358}
359
2c23d1b4 360void begin_odb_transaction(void)
568508e7 361{
2c23d1b4 362 odb_transaction_nesting += 1;
568508e7
JH
363}
364
2c23d1b4 365void flush_odb_transaction(void)
568508e7 366{
c0f4752e 367 flush_batch_fsync();
897c9e25 368 flush_bulk_checkin_packfile(&bulk_checkin_packfile);
568508e7 369}
2c23d1b4
NS
370
371void end_odb_transaction(void)
372{
373 odb_transaction_nesting -= 1;
374 if (odb_transaction_nesting < 0)
375 BUG("Unbalanced ODB transaction nesting");
376
377 if (odb_transaction_nesting)
378 return;
379
380 flush_odb_transaction();
568508e7 381}