]> git.ipfire.org Git - thirdparty/git.git/blob - reftable/stack.c
Merge branch 'es/test-cron-safety'
[thirdparty/git.git] / reftable / stack.c
1 /*
2 Copyright 2020 Google LLC
3
4 Use of this source code is governed by a BSD-style
5 license that can be found in the LICENSE file or at
6 https://developers.google.com/open-source/licenses/bsd
7 */
8
9 #include "stack.h"
10
11 #include "../write-or-die.h"
12 #include "system.h"
13 #include "merged.h"
14 #include "reader.h"
15 #include "refname.h"
16 #include "reftable-error.h"
17 #include "reftable-record.h"
18 #include "reftable-merged.h"
19 #include "writer.h"
20 #include "tempfile.h"
21
22 static int stack_try_add(struct reftable_stack *st,
23 int (*write_table)(struct reftable_writer *wr,
24 void *arg),
25 void *arg);
26 static int stack_write_compact(struct reftable_stack *st,
27 struct reftable_writer *wr,
28 size_t first, size_t last,
29 struct reftable_log_expiry_config *config);
30 static int stack_check_addition(struct reftable_stack *st,
31 const char *new_tab_name);
32 static void reftable_addition_close(struct reftable_addition *add);
33 static int reftable_stack_reload_maybe_reuse(struct reftable_stack *st,
34 int reuse_open);
35
36 static void stack_filename(struct strbuf *dest, struct reftable_stack *st,
37 const char *name)
38 {
39 strbuf_reset(dest);
40 strbuf_addstr(dest, st->reftable_dir);
41 strbuf_addstr(dest, "/");
42 strbuf_addstr(dest, name);
43 }
44
45 static ssize_t reftable_fd_write(void *arg, const void *data, size_t sz)
46 {
47 int *fdp = (int *)arg;
48 return write_in_full(*fdp, data, sz);
49 }
50
51 static int reftable_fd_flush(void *arg)
52 {
53 int *fdp = (int *)arg;
54
55 return fsync_component(FSYNC_COMPONENT_REFERENCE, *fdp);
56 }
57
58 int reftable_new_stack(struct reftable_stack **dest, const char *dir,
59 struct reftable_write_options config)
60 {
61 struct reftable_stack *p = reftable_calloc(1, sizeof(*p));
62 struct strbuf list_file_name = STRBUF_INIT;
63 int err = 0;
64
65 if (config.hash_id == 0) {
66 config.hash_id = GIT_SHA1_FORMAT_ID;
67 }
68
69 *dest = NULL;
70
71 strbuf_reset(&list_file_name);
72 strbuf_addstr(&list_file_name, dir);
73 strbuf_addstr(&list_file_name, "/tables.list");
74
75 p->list_file = strbuf_detach(&list_file_name, NULL);
76 p->list_fd = -1;
77 p->reftable_dir = xstrdup(dir);
78 p->config = config;
79
80 err = reftable_stack_reload_maybe_reuse(p, 1);
81 if (err < 0) {
82 reftable_stack_destroy(p);
83 } else {
84 *dest = p;
85 }
86 return err;
87 }
88
89 static int fd_read_lines(int fd, char ***namesp)
90 {
91 off_t size = lseek(fd, 0, SEEK_END);
92 char *buf = NULL;
93 int err = 0;
94 if (size < 0) {
95 err = REFTABLE_IO_ERROR;
96 goto done;
97 }
98 err = lseek(fd, 0, SEEK_SET);
99 if (err < 0) {
100 err = REFTABLE_IO_ERROR;
101 goto done;
102 }
103
104 REFTABLE_ALLOC_ARRAY(buf, size + 1);
105 if (read_in_full(fd, buf, size) != size) {
106 err = REFTABLE_IO_ERROR;
107 goto done;
108 }
109 buf[size] = 0;
110
111 parse_names(buf, size, namesp);
112
113 done:
114 reftable_free(buf);
115 return err;
116 }
117
118 int read_lines(const char *filename, char ***namesp)
119 {
120 int fd = open(filename, O_RDONLY);
121 int err = 0;
122 if (fd < 0) {
123 if (errno == ENOENT) {
124 REFTABLE_CALLOC_ARRAY(*namesp, 1);
125 return 0;
126 }
127
128 return REFTABLE_IO_ERROR;
129 }
130 err = fd_read_lines(fd, namesp);
131 close(fd);
132 return err;
133 }
134
135 struct reftable_merged_table *
136 reftable_stack_merged_table(struct reftable_stack *st)
137 {
138 return st->merged;
139 }
140
141 static int has_name(char **names, const char *name)
142 {
143 while (*names) {
144 if (!strcmp(*names, name))
145 return 1;
146 names++;
147 }
148 return 0;
149 }
150
151 /* Close and free the stack */
152 void reftable_stack_destroy(struct reftable_stack *st)
153 {
154 char **names = NULL;
155 int err = 0;
156 if (st->merged) {
157 reftable_merged_table_free(st->merged);
158 st->merged = NULL;
159 }
160
161 err = read_lines(st->list_file, &names);
162 if (err < 0) {
163 FREE_AND_NULL(names);
164 }
165
166 if (st->readers) {
167 int i = 0;
168 struct strbuf filename = STRBUF_INIT;
169 for (i = 0; i < st->readers_len; i++) {
170 const char *name = reader_name(st->readers[i]);
171 strbuf_reset(&filename);
172 if (names && !has_name(names, name)) {
173 stack_filename(&filename, st, name);
174 }
175 reftable_reader_free(st->readers[i]);
176
177 if (filename.len) {
178 /* On Windows, can only unlink after closing. */
179 unlink(filename.buf);
180 }
181 }
182 strbuf_release(&filename);
183 st->readers_len = 0;
184 FREE_AND_NULL(st->readers);
185 }
186
187 if (st->list_fd >= 0) {
188 close(st->list_fd);
189 st->list_fd = -1;
190 }
191
192 FREE_AND_NULL(st->list_file);
193 FREE_AND_NULL(st->reftable_dir);
194 reftable_free(st);
195 free_names(names);
196 }
197
198 static struct reftable_reader **stack_copy_readers(struct reftable_stack *st,
199 int cur_len)
200 {
201 struct reftable_reader **cur = reftable_calloc(cur_len, sizeof(*cur));
202 int i = 0;
203 for (i = 0; i < cur_len; i++) {
204 cur[i] = st->readers[i];
205 }
206 return cur;
207 }
208
209 static int reftable_stack_reload_once(struct reftable_stack *st, char **names,
210 int reuse_open)
211 {
212 size_t cur_len = !st->merged ? 0 : st->merged->stack_len;
213 struct reftable_reader **cur = stack_copy_readers(st, cur_len);
214 size_t names_len = names_length(names);
215 struct reftable_reader **new_readers =
216 reftable_calloc(names_len, sizeof(*new_readers));
217 struct reftable_table *new_tables =
218 reftable_calloc(names_len, sizeof(*new_tables));
219 size_t new_readers_len = 0;
220 struct reftable_merged_table *new_merged = NULL;
221 struct strbuf table_path = STRBUF_INIT;
222 int err = 0;
223 size_t i;
224
225 while (*names) {
226 struct reftable_reader *rd = NULL;
227 char *name = *names++;
228
229 /* this is linear; we assume compaction keeps the number of
230 tables under control so this is not quadratic. */
231 for (i = 0; reuse_open && i < cur_len; i++) {
232 if (cur[i] && 0 == strcmp(cur[i]->name, name)) {
233 rd = cur[i];
234 cur[i] = NULL;
235 break;
236 }
237 }
238
239 if (!rd) {
240 struct reftable_block_source src = { NULL };
241 stack_filename(&table_path, st, name);
242
243 err = reftable_block_source_from_file(&src,
244 table_path.buf);
245 if (err < 0)
246 goto done;
247
248 err = reftable_new_reader(&rd, &src, name);
249 if (err < 0)
250 goto done;
251 }
252
253 new_readers[new_readers_len] = rd;
254 reftable_table_from_reader(&new_tables[new_readers_len], rd);
255 new_readers_len++;
256 }
257
258 /* success! */
259 err = reftable_new_merged_table(&new_merged, new_tables,
260 new_readers_len, st->config.hash_id);
261 if (err < 0)
262 goto done;
263
264 new_tables = NULL;
265 st->readers_len = new_readers_len;
266 if (st->merged) {
267 merged_table_release(st->merged);
268 reftable_merged_table_free(st->merged);
269 }
270 if (st->readers) {
271 reftable_free(st->readers);
272 }
273 st->readers = new_readers;
274 new_readers = NULL;
275 new_readers_len = 0;
276
277 new_merged->suppress_deletions = 1;
278 st->merged = new_merged;
279 for (i = 0; i < cur_len; i++) {
280 if (cur[i]) {
281 const char *name = reader_name(cur[i]);
282 stack_filename(&table_path, st, name);
283
284 reader_close(cur[i]);
285 reftable_reader_free(cur[i]);
286
287 /* On Windows, can only unlink after closing. */
288 unlink(table_path.buf);
289 }
290 }
291
292 done:
293 for (i = 0; i < new_readers_len; i++) {
294 reader_close(new_readers[i]);
295 reftable_reader_free(new_readers[i]);
296 }
297 reftable_free(new_readers);
298 reftable_free(new_tables);
299 reftable_free(cur);
300 strbuf_release(&table_path);
301 return err;
302 }
303
304 /* return negative if a before b. */
305 static int tv_cmp(struct timeval *a, struct timeval *b)
306 {
307 time_t diff = a->tv_sec - b->tv_sec;
308 int udiff = a->tv_usec - b->tv_usec;
309
310 if (diff != 0)
311 return diff;
312
313 return udiff;
314 }
315
316 static int reftable_stack_reload_maybe_reuse(struct reftable_stack *st,
317 int reuse_open)
318 {
319 char **names = NULL, **names_after = NULL;
320 struct timeval deadline;
321 int64_t delay = 0;
322 int tries = 0, err;
323 int fd = -1;
324
325 err = gettimeofday(&deadline, NULL);
326 if (err < 0)
327 goto out;
328 deadline.tv_sec += 3;
329
330 while (1) {
331 struct timeval now;
332
333 err = gettimeofday(&now, NULL);
334 if (err < 0)
335 goto out;
336
337 /*
338 * Only look at deadlines after the first few times. This
339 * simplifies debugging in GDB.
340 */
341 tries++;
342 if (tries > 3 && tv_cmp(&now, &deadline) >= 0)
343 goto out;
344
345 fd = open(st->list_file, O_RDONLY);
346 if (fd < 0) {
347 if (errno != ENOENT) {
348 err = REFTABLE_IO_ERROR;
349 goto out;
350 }
351
352 REFTABLE_CALLOC_ARRAY(names, 1);
353 } else {
354 err = fd_read_lines(fd, &names);
355 if (err < 0)
356 goto out;
357 }
358
359 err = reftable_stack_reload_once(st, names, reuse_open);
360 if (!err)
361 break;
362 if (err != REFTABLE_NOT_EXIST_ERROR)
363 goto out;
364
365 /*
366 * REFTABLE_NOT_EXIST_ERROR can be caused by a concurrent
367 * writer. Check if there was one by checking if the name list
368 * changed.
369 */
370 err = read_lines(st->list_file, &names_after);
371 if (err < 0)
372 goto out;
373 if (names_equal(names_after, names)) {
374 err = REFTABLE_NOT_EXIST_ERROR;
375 goto out;
376 }
377
378 free_names(names);
379 names = NULL;
380 free_names(names_after);
381 names_after = NULL;
382 close(fd);
383 fd = -1;
384
385 delay = delay + (delay * rand()) / RAND_MAX + 1;
386 sleep_millisec(delay);
387 }
388
389 out:
390 /*
391 * Invalidate the stat cache. It is sufficient to only close the file
392 * descriptor and keep the cached stat info because we never use the
393 * latter when the former is negative.
394 */
395 if (st->list_fd >= 0) {
396 close(st->list_fd);
397 st->list_fd = -1;
398 }
399
400 /*
401 * Cache stat information in case it provides a useful signal to us.
402 * According to POSIX, "The st_ino and st_dev fields taken together
403 * uniquely identify the file within the system." That being said,
404 * Windows is not POSIX compliant and we do not have these fields
405 * available. So the information we have there is insufficient to
406 * determine whether two file descriptors point to the same file.
407 *
408 * While we could fall back to using other signals like the file's
409 * mtime, those are not sufficient to avoid races. We thus refrain from
410 * using the stat cache on such systems and fall back to the secondary
411 * caching mechanism, which is to check whether contents of the file
412 * have changed.
413 *
414 * On other systems which are POSIX compliant we must keep the file
415 * descriptor open. This is to avoid a race condition where two
416 * processes access the reftable stack at the same point in time:
417 *
418 * 1. A reads the reftable stack and caches its stat info.
419 *
420 * 2. B updates the stack, appending a new table to "tables.list".
421 * This will both use a new inode and result in a different file
422 * size, thus invalidating A's cache in theory.
423 *
424 * 3. B decides to auto-compact the stack and merges two tables. The
425 * file size now matches what A has cached again. Furthermore, the
426 * filesystem may decide to recycle the inode number of the file
427 * we have replaced in (2) because it is not in use anymore.
428 *
429 * 4. A reloads the reftable stack. Neither the inode number nor the
430 * file size changed. If the timestamps did not change either then
431 * we think the cached copy of our stack is up-to-date.
432 *
433 * By keeping the file descriptor open the inode number cannot be
434 * recycled, mitigating the race.
435 */
436 if (!err && fd >= 0 && !fstat(fd, &st->list_st) &&
437 st->list_st.st_dev && st->list_st.st_ino) {
438 st->list_fd = fd;
439 fd = -1;
440 }
441
442 if (fd >= 0)
443 close(fd);
444 free_names(names);
445 free_names(names_after);
446 return err;
447 }
448
449 /* -1 = error
450 0 = up to date
451 1 = changed. */
452 static int stack_uptodate(struct reftable_stack *st)
453 {
454 char **names = NULL;
455 int err;
456 int i = 0;
457
458 /*
459 * When we have cached stat information available then we use it to
460 * verify whether the file has been rewritten.
461 *
462 * Note that we explicitly do not want to use `stat_validity_check()`
463 * and friends here because they may end up not comparing the `st_dev`
464 * and `st_ino` fields. These functions thus cannot guarantee that we
465 * indeed still have the same file.
466 */
467 if (st->list_fd >= 0) {
468 struct stat list_st;
469
470 if (stat(st->list_file, &list_st) < 0) {
471 /*
472 * It's fine for "tables.list" to not exist. In that
473 * case, we have to refresh when the loaded stack has
474 * any readers.
475 */
476 if (errno == ENOENT)
477 return !!st->readers_len;
478 return REFTABLE_IO_ERROR;
479 }
480
481 /*
482 * When "tables.list" refers to the same file we can assume
483 * that it didn't change. This is because we always use
484 * rename(3P) to update the file and never write to it
485 * directly.
486 */
487 if (st->list_st.st_dev == list_st.st_dev &&
488 st->list_st.st_ino == list_st.st_ino)
489 return 0;
490 }
491
492 err = read_lines(st->list_file, &names);
493 if (err < 0)
494 return err;
495
496 for (i = 0; i < st->readers_len; i++) {
497 if (!names[i]) {
498 err = 1;
499 goto done;
500 }
501
502 if (strcmp(st->readers[i]->name, names[i])) {
503 err = 1;
504 goto done;
505 }
506 }
507
508 if (names[st->merged->stack_len]) {
509 err = 1;
510 goto done;
511 }
512
513 done:
514 free_names(names);
515 return err;
516 }
517
518 int reftable_stack_reload(struct reftable_stack *st)
519 {
520 int err = stack_uptodate(st);
521 if (err > 0)
522 return reftable_stack_reload_maybe_reuse(st, 1);
523 return err;
524 }
525
526 int reftable_stack_add(struct reftable_stack *st,
527 int (*write)(struct reftable_writer *wr, void *arg),
528 void *arg)
529 {
530 int err = stack_try_add(st, write, arg);
531 if (err < 0) {
532 if (err == REFTABLE_LOCK_ERROR) {
533 /* Ignore error return, we want to propagate
534 REFTABLE_LOCK_ERROR.
535 */
536 reftable_stack_reload(st);
537 }
538 return err;
539 }
540
541 return 0;
542 }
543
544 static void format_name(struct strbuf *dest, uint64_t min, uint64_t max)
545 {
546 char buf[100];
547 uint32_t rnd = (uint32_t)git_rand();
548 snprintf(buf, sizeof(buf), "0x%012" PRIx64 "-0x%012" PRIx64 "-%08x",
549 min, max, rnd);
550 strbuf_reset(dest);
551 strbuf_addstr(dest, buf);
552 }
553
554 struct reftable_addition {
555 struct tempfile *lock_file;
556 struct reftable_stack *stack;
557
558 char **new_tables;
559 size_t new_tables_len, new_tables_cap;
560 uint64_t next_update_index;
561 };
562
563 #define REFTABLE_ADDITION_INIT {0}
564
565 static int reftable_stack_init_addition(struct reftable_addition *add,
566 struct reftable_stack *st)
567 {
568 struct strbuf lock_file_name = STRBUF_INIT;
569 int err = 0;
570 add->stack = st;
571
572 strbuf_addf(&lock_file_name, "%s.lock", st->list_file);
573
574 add->lock_file = create_tempfile(lock_file_name.buf);
575 if (!add->lock_file) {
576 if (errno == EEXIST) {
577 err = REFTABLE_LOCK_ERROR;
578 } else {
579 err = REFTABLE_IO_ERROR;
580 }
581 goto done;
582 }
583 if (st->config.default_permissions) {
584 if (chmod(add->lock_file->filename.buf, st->config.default_permissions) < 0) {
585 err = REFTABLE_IO_ERROR;
586 goto done;
587 }
588 }
589
590 err = stack_uptodate(st);
591 if (err < 0)
592 goto done;
593
594 if (err > 1) {
595 err = REFTABLE_LOCK_ERROR;
596 goto done;
597 }
598
599 add->next_update_index = reftable_stack_next_update_index(st);
600 done:
601 if (err) {
602 reftable_addition_close(add);
603 }
604 strbuf_release(&lock_file_name);
605 return err;
606 }
607
608 static void reftable_addition_close(struct reftable_addition *add)
609 {
610 struct strbuf nm = STRBUF_INIT;
611 size_t i;
612
613 for (i = 0; i < add->new_tables_len; i++) {
614 stack_filename(&nm, add->stack, add->new_tables[i]);
615 unlink(nm.buf);
616 reftable_free(add->new_tables[i]);
617 add->new_tables[i] = NULL;
618 }
619 reftable_free(add->new_tables);
620 add->new_tables = NULL;
621 add->new_tables_len = 0;
622 add->new_tables_cap = 0;
623
624 delete_tempfile(&add->lock_file);
625 strbuf_release(&nm);
626 }
627
628 void reftable_addition_destroy(struct reftable_addition *add)
629 {
630 if (!add) {
631 return;
632 }
633 reftable_addition_close(add);
634 reftable_free(add);
635 }
636
637 int reftable_addition_commit(struct reftable_addition *add)
638 {
639 struct strbuf table_list = STRBUF_INIT;
640 int lock_file_fd = get_tempfile_fd(add->lock_file);
641 int err = 0;
642 size_t i;
643
644 if (add->new_tables_len == 0)
645 goto done;
646
647 for (i = 0; i < add->stack->merged->stack_len; i++) {
648 strbuf_addstr(&table_list, add->stack->readers[i]->name);
649 strbuf_addstr(&table_list, "\n");
650 }
651 for (i = 0; i < add->new_tables_len; i++) {
652 strbuf_addstr(&table_list, add->new_tables[i]);
653 strbuf_addstr(&table_list, "\n");
654 }
655
656 err = write_in_full(lock_file_fd, table_list.buf, table_list.len);
657 strbuf_release(&table_list);
658 if (err < 0) {
659 err = REFTABLE_IO_ERROR;
660 goto done;
661 }
662
663 fsync_component_or_die(FSYNC_COMPONENT_REFERENCE, lock_file_fd,
664 get_tempfile_path(add->lock_file));
665
666 err = rename_tempfile(&add->lock_file, add->stack->list_file);
667 if (err < 0) {
668 err = REFTABLE_IO_ERROR;
669 goto done;
670 }
671
672 /* success, no more state to clean up. */
673 for (i = 0; i < add->new_tables_len; i++)
674 reftable_free(add->new_tables[i]);
675 reftable_free(add->new_tables);
676 add->new_tables = NULL;
677 add->new_tables_len = 0;
678 add->new_tables_cap = 0;
679
680 err = reftable_stack_reload_maybe_reuse(add->stack, 1);
681 if (err)
682 goto done;
683
684 if (!add->stack->disable_auto_compact)
685 err = reftable_stack_auto_compact(add->stack);
686
687 done:
688 reftable_addition_close(add);
689 return err;
690 }
691
692 int reftable_stack_new_addition(struct reftable_addition **dest,
693 struct reftable_stack *st)
694 {
695 int err = 0;
696 struct reftable_addition empty = REFTABLE_ADDITION_INIT;
697 REFTABLE_CALLOC_ARRAY(*dest, 1);
698 **dest = empty;
699 err = reftable_stack_init_addition(*dest, st);
700 if (err) {
701 reftable_free(*dest);
702 *dest = NULL;
703 }
704 return err;
705 }
706
707 static int stack_try_add(struct reftable_stack *st,
708 int (*write_table)(struct reftable_writer *wr,
709 void *arg),
710 void *arg)
711 {
712 struct reftable_addition add = REFTABLE_ADDITION_INIT;
713 int err = reftable_stack_init_addition(&add, st);
714 if (err < 0)
715 goto done;
716 if (err > 0) {
717 err = REFTABLE_LOCK_ERROR;
718 goto done;
719 }
720
721 err = reftable_addition_add(&add, write_table, arg);
722 if (err < 0)
723 goto done;
724
725 err = reftable_addition_commit(&add);
726 done:
727 reftable_addition_close(&add);
728 return err;
729 }
730
731 int reftable_addition_add(struct reftable_addition *add,
732 int (*write_table)(struct reftable_writer *wr,
733 void *arg),
734 void *arg)
735 {
736 struct strbuf temp_tab_file_name = STRBUF_INIT;
737 struct strbuf tab_file_name = STRBUF_INIT;
738 struct strbuf next_name = STRBUF_INIT;
739 struct reftable_writer *wr = NULL;
740 struct tempfile *tab_file = NULL;
741 int err = 0;
742 int tab_fd;
743
744 strbuf_reset(&next_name);
745 format_name(&next_name, add->next_update_index, add->next_update_index);
746
747 stack_filename(&temp_tab_file_name, add->stack, next_name.buf);
748 strbuf_addstr(&temp_tab_file_name, ".temp.XXXXXX");
749
750 tab_file = mks_tempfile(temp_tab_file_name.buf);
751 if (!tab_file) {
752 err = REFTABLE_IO_ERROR;
753 goto done;
754 }
755 if (add->stack->config.default_permissions) {
756 if (chmod(get_tempfile_path(tab_file),
757 add->stack->config.default_permissions)) {
758 err = REFTABLE_IO_ERROR;
759 goto done;
760 }
761 }
762 tab_fd = get_tempfile_fd(tab_file);
763
764 wr = reftable_new_writer(reftable_fd_write, reftable_fd_flush, &tab_fd,
765 &add->stack->config);
766 err = write_table(wr, arg);
767 if (err < 0)
768 goto done;
769
770 err = reftable_writer_close(wr);
771 if (err == REFTABLE_EMPTY_TABLE_ERROR) {
772 err = 0;
773 goto done;
774 }
775 if (err < 0)
776 goto done;
777
778 err = close_tempfile_gently(tab_file);
779 if (err < 0) {
780 err = REFTABLE_IO_ERROR;
781 goto done;
782 }
783
784 err = stack_check_addition(add->stack, get_tempfile_path(tab_file));
785 if (err < 0)
786 goto done;
787
788 if (wr->min_update_index < add->next_update_index) {
789 err = REFTABLE_API_ERROR;
790 goto done;
791 }
792
793 format_name(&next_name, wr->min_update_index, wr->max_update_index);
794 strbuf_addstr(&next_name, ".ref");
795 stack_filename(&tab_file_name, add->stack, next_name.buf);
796
797 /*
798 On windows, this relies on rand() picking a unique destination name.
799 Maybe we should do retry loop as well?
800 */
801 err = rename_tempfile(&tab_file, tab_file_name.buf);
802 if (err < 0) {
803 err = REFTABLE_IO_ERROR;
804 goto done;
805 }
806
807 REFTABLE_ALLOC_GROW(add->new_tables, add->new_tables_len + 1,
808 add->new_tables_cap);
809 add->new_tables[add->new_tables_len++] = strbuf_detach(&next_name, NULL);
810 done:
811 delete_tempfile(&tab_file);
812 strbuf_release(&temp_tab_file_name);
813 strbuf_release(&tab_file_name);
814 strbuf_release(&next_name);
815 reftable_writer_free(wr);
816 return err;
817 }
818
819 uint64_t reftable_stack_next_update_index(struct reftable_stack *st)
820 {
821 int sz = st->merged->stack_len;
822 if (sz > 0)
823 return reftable_reader_max_update_index(st->readers[sz - 1]) +
824 1;
825 return 1;
826 }
827
828 static int stack_compact_locked(struct reftable_stack *st,
829 size_t first, size_t last,
830 struct reftable_log_expiry_config *config,
831 struct tempfile **tab_file_out)
832 {
833 struct strbuf next_name = STRBUF_INIT;
834 struct strbuf tab_file_path = STRBUF_INIT;
835 struct reftable_writer *wr = NULL;
836 struct tempfile *tab_file;
837 int tab_fd, err = 0;
838
839 format_name(&next_name,
840 reftable_reader_min_update_index(st->readers[first]),
841 reftable_reader_max_update_index(st->readers[last]));
842 stack_filename(&tab_file_path, st, next_name.buf);
843 strbuf_addstr(&tab_file_path, ".temp.XXXXXX");
844
845 tab_file = mks_tempfile(tab_file_path.buf);
846 if (!tab_file) {
847 err = REFTABLE_IO_ERROR;
848 goto done;
849 }
850 tab_fd = get_tempfile_fd(tab_file);
851
852 if (st->config.default_permissions &&
853 chmod(get_tempfile_path(tab_file), st->config.default_permissions) < 0) {
854 err = REFTABLE_IO_ERROR;
855 goto done;
856 }
857
858 wr = reftable_new_writer(reftable_fd_write, reftable_fd_flush,
859 &tab_fd, &st->config);
860 err = stack_write_compact(st, wr, first, last, config);
861 if (err < 0)
862 goto done;
863
864 err = reftable_writer_close(wr);
865 if (err < 0)
866 goto done;
867
868 err = close_tempfile_gently(tab_file);
869 if (err < 0)
870 goto done;
871
872 *tab_file_out = tab_file;
873 tab_file = NULL;
874
875 done:
876 delete_tempfile(&tab_file);
877 reftable_writer_free(wr);
878 strbuf_release(&next_name);
879 strbuf_release(&tab_file_path);
880 return err;
881 }
882
883 static int stack_write_compact(struct reftable_stack *st,
884 struct reftable_writer *wr,
885 size_t first, size_t last,
886 struct reftable_log_expiry_config *config)
887 {
888 size_t subtabs_len = last - first + 1;
889 struct reftable_table *subtabs = reftable_calloc(
890 last - first + 1, sizeof(*subtabs));
891 struct reftable_merged_table *mt = NULL;
892 struct reftable_iterator it = { NULL };
893 struct reftable_ref_record ref = { NULL };
894 struct reftable_log_record log = { NULL };
895 uint64_t entries = 0;
896 int err = 0;
897
898 for (size_t i = first, j = 0; i <= last; i++) {
899 struct reftable_reader *t = st->readers[i];
900 reftable_table_from_reader(&subtabs[j++], t);
901 st->stats.bytes += t->size;
902 }
903 reftable_writer_set_limits(wr, st->readers[first]->min_update_index,
904 st->readers[last]->max_update_index);
905
906 err = reftable_new_merged_table(&mt, subtabs, subtabs_len,
907 st->config.hash_id);
908 if (err < 0) {
909 reftable_free(subtabs);
910 goto done;
911 }
912
913 err = reftable_merged_table_seek_ref(mt, &it, "");
914 if (err < 0)
915 goto done;
916
917 while (1) {
918 err = reftable_iterator_next_ref(&it, &ref);
919 if (err > 0) {
920 err = 0;
921 break;
922 }
923 if (err < 0)
924 goto done;
925
926 if (first == 0 && reftable_ref_record_is_deletion(&ref)) {
927 continue;
928 }
929
930 err = reftable_writer_add_ref(wr, &ref);
931 if (err < 0)
932 goto done;
933 entries++;
934 }
935 reftable_iterator_destroy(&it);
936
937 err = reftable_merged_table_seek_log(mt, &it, "");
938 if (err < 0)
939 goto done;
940
941 while (1) {
942 err = reftable_iterator_next_log(&it, &log);
943 if (err > 0) {
944 err = 0;
945 break;
946 }
947 if (err < 0)
948 goto done;
949 if (first == 0 && reftable_log_record_is_deletion(&log)) {
950 continue;
951 }
952
953 if (config && config->min_update_index > 0 &&
954 log.update_index < config->min_update_index) {
955 continue;
956 }
957
958 if (config && config->time > 0 &&
959 log.value.update.time < config->time) {
960 continue;
961 }
962
963 err = reftable_writer_add_log(wr, &log);
964 if (err < 0)
965 goto done;
966 entries++;
967 }
968
969 done:
970 reftable_iterator_destroy(&it);
971 if (mt) {
972 merged_table_release(mt);
973 reftable_merged_table_free(mt);
974 }
975 reftable_ref_record_release(&ref);
976 reftable_log_record_release(&log);
977 st->stats.entries_written += entries;
978 return err;
979 }
980
981 /* < 0: error. 0 == OK, > 0 attempt failed; could retry. */
982 static int stack_compact_range(struct reftable_stack *st,
983 size_t first, size_t last,
984 struct reftable_log_expiry_config *expiry)
985 {
986 struct strbuf tables_list_buf = STRBUF_INIT;
987 struct strbuf new_table_name = STRBUF_INIT;
988 struct strbuf new_table_path = STRBUF_INIT;
989 struct strbuf table_name = STRBUF_INIT;
990 struct lock_file tables_list_lock = LOCK_INIT;
991 struct lock_file *table_locks = NULL;
992 struct tempfile *new_table = NULL;
993 int is_empty_table = 0, err = 0;
994 size_t i;
995
996 if (first > last || (!expiry && first == last)) {
997 err = 0;
998 goto done;
999 }
1000
1001 st->stats.attempts++;
1002
1003 /*
1004 * Hold the lock so that we can read "tables.list" and lock all tables
1005 * which are part of the user-specified range.
1006 */
1007 err = hold_lock_file_for_update(&tables_list_lock, st->list_file,
1008 LOCK_NO_DEREF);
1009 if (err < 0) {
1010 if (errno == EEXIST)
1011 err = 1;
1012 else
1013 err = REFTABLE_IO_ERROR;
1014 goto done;
1015 }
1016
1017 err = stack_uptodate(st);
1018 if (err)
1019 goto done;
1020
1021 /*
1022 * Lock all tables in the user-provided range. This is the slice of our
1023 * stack which we'll compact.
1024 */
1025 REFTABLE_CALLOC_ARRAY(table_locks, last - first + 1);
1026 for (i = first; i <= last; i++) {
1027 stack_filename(&table_name, st, reader_name(st->readers[i]));
1028
1029 err = hold_lock_file_for_update(&table_locks[i - first],
1030 table_name.buf, LOCK_NO_DEREF);
1031 if (err < 0) {
1032 if (errno == EEXIST)
1033 err = 1;
1034 else
1035 err = REFTABLE_IO_ERROR;
1036 goto done;
1037 }
1038
1039 /*
1040 * We need to close the lockfiles as we might otherwise easily
1041 * run into file descriptor exhaustion when we compress a lot
1042 * of tables.
1043 */
1044 err = close_lock_file_gently(&table_locks[i - first]);
1045 if (err < 0) {
1046 err = REFTABLE_IO_ERROR;
1047 goto done;
1048 }
1049 }
1050
1051 /*
1052 * We have locked all tables in our range and can thus release the
1053 * "tables.list" lock while compacting the locked tables. This allows
1054 * concurrent updates to the stack to proceed.
1055 */
1056 err = rollback_lock_file(&tables_list_lock);
1057 if (err < 0) {
1058 err = REFTABLE_IO_ERROR;
1059 goto done;
1060 }
1061
1062 /*
1063 * Compact the now-locked tables into a new table. Note that compacting
1064 * these tables may end up with an empty new table in case tombstones
1065 * end up cancelling out all refs in that range.
1066 */
1067 err = stack_compact_locked(st, first, last, expiry, &new_table);
1068 if (err < 0) {
1069 if (err != REFTABLE_EMPTY_TABLE_ERROR)
1070 goto done;
1071 is_empty_table = 1;
1072 }
1073
1074 /*
1075 * Now that we have written the new, compacted table we need to re-lock
1076 * "tables.list". We'll then replace the compacted range of tables with
1077 * the new table.
1078 */
1079 err = hold_lock_file_for_update(&tables_list_lock, st->list_file,
1080 LOCK_NO_DEREF);
1081 if (err < 0) {
1082 if (errno == EEXIST)
1083 err = 1;
1084 else
1085 err = REFTABLE_IO_ERROR;
1086 goto done;
1087 }
1088
1089 if (st->config.default_permissions) {
1090 if (chmod(get_lock_file_path(&tables_list_lock),
1091 st->config.default_permissions) < 0) {
1092 err = REFTABLE_IO_ERROR;
1093 goto done;
1094 }
1095 }
1096
1097 /*
1098 * If the resulting compacted table is not empty, then we need to move
1099 * it into place now.
1100 */
1101 if (!is_empty_table) {
1102 format_name(&new_table_name, st->readers[first]->min_update_index,
1103 st->readers[last]->max_update_index);
1104 strbuf_addstr(&new_table_name, ".ref");
1105 stack_filename(&new_table_path, st, new_table_name.buf);
1106
1107 err = rename_tempfile(&new_table, new_table_path.buf);
1108 if (err < 0) {
1109 err = REFTABLE_IO_ERROR;
1110 goto done;
1111 }
1112 }
1113
1114 /*
1115 * Write the new "tables.list" contents with the compacted table we
1116 * have just written. In case the compacted table became empty we
1117 * simply skip writing it.
1118 */
1119 for (i = 0; i < first; i++)
1120 strbuf_addf(&tables_list_buf, "%s\n", st->readers[i]->name);
1121 if (!is_empty_table)
1122 strbuf_addf(&tables_list_buf, "%s\n", new_table_name.buf);
1123 for (i = last + 1; i < st->merged->stack_len; i++)
1124 strbuf_addf(&tables_list_buf, "%s\n", st->readers[i]->name);
1125
1126 err = write_in_full(get_lock_file_fd(&tables_list_lock),
1127 tables_list_buf.buf, tables_list_buf.len);
1128 if (err < 0) {
1129 err = REFTABLE_IO_ERROR;
1130 unlink(new_table_path.buf);
1131 goto done;
1132 }
1133
1134 err = fsync_component(FSYNC_COMPONENT_REFERENCE, get_lock_file_fd(&tables_list_lock));
1135 if (err < 0) {
1136 err = REFTABLE_IO_ERROR;
1137 unlink(new_table_path.buf);
1138 goto done;
1139 }
1140
1141 err = commit_lock_file(&tables_list_lock);
1142 if (err < 0) {
1143 err = REFTABLE_IO_ERROR;
1144 unlink(new_table_path.buf);
1145 goto done;
1146 }
1147
1148 /*
1149 * Reload the stack before deleting the compacted tables. We can only
1150 * delete the files after we closed them on Windows, so this needs to
1151 * happen first.
1152 */
1153 err = reftable_stack_reload_maybe_reuse(st, first < last);
1154 if (err < 0)
1155 goto done;
1156
1157 /*
1158 * Delete the old tables. They may still be in use by concurrent
1159 * readers, so it is expected that unlinking tables may fail.
1160 */
1161 for (i = first; i <= last; i++) {
1162 struct lock_file *table_lock = &table_locks[i - first];
1163 char *table_path = get_locked_file_path(table_lock);
1164 unlink(table_path);
1165 free(table_path);
1166 }
1167
1168 done:
1169 rollback_lock_file(&tables_list_lock);
1170 for (i = first; table_locks && i <= last; i++)
1171 rollback_lock_file(&table_locks[i - first]);
1172 reftable_free(table_locks);
1173
1174 delete_tempfile(&new_table);
1175 strbuf_release(&new_table_name);
1176 strbuf_release(&new_table_path);
1177
1178 strbuf_release(&tables_list_buf);
1179 strbuf_release(&table_name);
1180 return err;
1181 }
1182
1183 int reftable_stack_compact_all(struct reftable_stack *st,
1184 struct reftable_log_expiry_config *config)
1185 {
1186 return stack_compact_range(st, 0, st->merged->stack_len ?
1187 st->merged->stack_len - 1 : 0, config);
1188 }
1189
1190 static int stack_compact_range_stats(struct reftable_stack *st,
1191 size_t first, size_t last,
1192 struct reftable_log_expiry_config *config)
1193 {
1194 int err = stack_compact_range(st, first, last, config);
1195 if (err > 0)
1196 st->stats.failures++;
1197 return err;
1198 }
1199
1200 static int segment_size(struct segment *s)
1201 {
1202 return s->end - s->start;
1203 }
1204
1205 int fastlog2(uint64_t sz)
1206 {
1207 int l = 0;
1208 if (sz == 0)
1209 return 0;
1210 for (; sz; sz /= 2) {
1211 l++;
1212 }
1213 return l - 1;
1214 }
1215
1216 struct segment *sizes_to_segments(size_t *seglen, uint64_t *sizes, size_t n)
1217 {
1218 struct segment *segs = reftable_calloc(n, sizeof(*segs));
1219 struct segment cur = { 0 };
1220 size_t next = 0, i;
1221
1222 if (n == 0) {
1223 *seglen = 0;
1224 return segs;
1225 }
1226 for (i = 0; i < n; i++) {
1227 int log = fastlog2(sizes[i]);
1228 if (cur.log != log && cur.bytes > 0) {
1229 struct segment fresh = {
1230 .start = i,
1231 };
1232
1233 segs[next++] = cur;
1234 cur = fresh;
1235 }
1236
1237 cur.log = log;
1238 cur.end = i + 1;
1239 cur.bytes += sizes[i];
1240 }
1241 segs[next++] = cur;
1242 *seglen = next;
1243 return segs;
1244 }
1245
1246 struct segment suggest_compaction_segment(uint64_t *sizes, size_t n)
1247 {
1248 struct segment min_seg = {
1249 .log = 64,
1250 };
1251 struct segment *segs;
1252 size_t seglen = 0, i;
1253
1254 segs = sizes_to_segments(&seglen, sizes, n);
1255 for (i = 0; i < seglen; i++) {
1256 if (segment_size(&segs[i]) == 1)
1257 continue;
1258
1259 if (segs[i].log < min_seg.log)
1260 min_seg = segs[i];
1261 }
1262
1263 while (min_seg.start > 0) {
1264 size_t prev = min_seg.start - 1;
1265 if (fastlog2(min_seg.bytes) < fastlog2(sizes[prev]))
1266 break;
1267
1268 min_seg.start = prev;
1269 min_seg.bytes += sizes[prev];
1270 }
1271
1272 reftable_free(segs);
1273 return min_seg;
1274 }
1275
1276 static uint64_t *stack_table_sizes_for_compaction(struct reftable_stack *st)
1277 {
1278 uint64_t *sizes =
1279 reftable_calloc(st->merged->stack_len, sizeof(*sizes));
1280 int version = (st->config.hash_id == GIT_SHA1_FORMAT_ID) ? 1 : 2;
1281 int overhead = header_size(version) - 1;
1282 int i = 0;
1283 for (i = 0; i < st->merged->stack_len; i++) {
1284 sizes[i] = st->readers[i]->size - overhead;
1285 }
1286 return sizes;
1287 }
1288
1289 int reftable_stack_auto_compact(struct reftable_stack *st)
1290 {
1291 uint64_t *sizes = stack_table_sizes_for_compaction(st);
1292 struct segment seg =
1293 suggest_compaction_segment(sizes, st->merged->stack_len);
1294 reftable_free(sizes);
1295 if (segment_size(&seg) > 0)
1296 return stack_compact_range_stats(st, seg.start, seg.end - 1,
1297 NULL);
1298
1299 return 0;
1300 }
1301
1302 struct reftable_compaction_stats *
1303 reftable_stack_compaction_stats(struct reftable_stack *st)
1304 {
1305 return &st->stats;
1306 }
1307
1308 int reftable_stack_read_ref(struct reftable_stack *st, const char *refname,
1309 struct reftable_ref_record *ref)
1310 {
1311 struct reftable_table tab = { NULL };
1312 reftable_table_from_merged_table(&tab, reftable_stack_merged_table(st));
1313 return reftable_table_read_ref(&tab, refname, ref);
1314 }
1315
1316 int reftable_stack_read_log(struct reftable_stack *st, const char *refname,
1317 struct reftable_log_record *log)
1318 {
1319 struct reftable_iterator it = { NULL };
1320 struct reftable_merged_table *mt = reftable_stack_merged_table(st);
1321 int err = reftable_merged_table_seek_log(mt, &it, refname);
1322 if (err)
1323 goto done;
1324
1325 err = reftable_iterator_next_log(&it, log);
1326 if (err)
1327 goto done;
1328
1329 if (strcmp(log->refname, refname) ||
1330 reftable_log_record_is_deletion(log)) {
1331 err = 1;
1332 goto done;
1333 }
1334
1335 done:
1336 if (err) {
1337 reftable_log_record_release(log);
1338 }
1339 reftable_iterator_destroy(&it);
1340 return err;
1341 }
1342
1343 static int stack_check_addition(struct reftable_stack *st,
1344 const char *new_tab_name)
1345 {
1346 int err = 0;
1347 struct reftable_block_source src = { NULL };
1348 struct reftable_reader *rd = NULL;
1349 struct reftable_table tab = { NULL };
1350 struct reftable_ref_record *refs = NULL;
1351 struct reftable_iterator it = { NULL };
1352 int cap = 0;
1353 int len = 0;
1354 int i = 0;
1355
1356 if (st->config.skip_name_check)
1357 return 0;
1358
1359 err = reftable_block_source_from_file(&src, new_tab_name);
1360 if (err < 0)
1361 goto done;
1362
1363 err = reftable_new_reader(&rd, &src, new_tab_name);
1364 if (err < 0)
1365 goto done;
1366
1367 err = reftable_reader_seek_ref(rd, &it, "");
1368 if (err > 0) {
1369 err = 0;
1370 goto done;
1371 }
1372 if (err < 0)
1373 goto done;
1374
1375 while (1) {
1376 struct reftable_ref_record ref = { NULL };
1377 err = reftable_iterator_next_ref(&it, &ref);
1378 if (err > 0)
1379 break;
1380 if (err < 0)
1381 goto done;
1382
1383 REFTABLE_ALLOC_GROW(refs, len + 1, cap);
1384 refs[len++] = ref;
1385 }
1386
1387 reftable_table_from_merged_table(&tab, reftable_stack_merged_table(st));
1388
1389 err = validate_ref_record_addition(tab, refs, len);
1390
1391 done:
1392 for (i = 0; i < len; i++) {
1393 reftable_ref_record_release(&refs[i]);
1394 }
1395
1396 free(refs);
1397 reftable_iterator_destroy(&it);
1398 reftable_reader_free(rd);
1399 return err;
1400 }
1401
1402 static int is_table_name(const char *s)
1403 {
1404 const char *dot = strrchr(s, '.');
1405 return dot && !strcmp(dot, ".ref");
1406 }
1407
1408 static void remove_maybe_stale_table(struct reftable_stack *st, uint64_t max,
1409 const char *name)
1410 {
1411 int err = 0;
1412 uint64_t update_idx = 0;
1413 struct reftable_block_source src = { NULL };
1414 struct reftable_reader *rd = NULL;
1415 struct strbuf table_path = STRBUF_INIT;
1416 stack_filename(&table_path, st, name);
1417
1418 err = reftable_block_source_from_file(&src, table_path.buf);
1419 if (err < 0)
1420 goto done;
1421
1422 err = reftable_new_reader(&rd, &src, name);
1423 if (err < 0)
1424 goto done;
1425
1426 update_idx = reftable_reader_max_update_index(rd);
1427 reftable_reader_free(rd);
1428
1429 if (update_idx <= max) {
1430 unlink(table_path.buf);
1431 }
1432 done:
1433 strbuf_release(&table_path);
1434 }
1435
1436 static int reftable_stack_clean_locked(struct reftable_stack *st)
1437 {
1438 uint64_t max = reftable_merged_table_max_update_index(
1439 reftable_stack_merged_table(st));
1440 DIR *dir = opendir(st->reftable_dir);
1441 struct dirent *d = NULL;
1442 if (!dir) {
1443 return REFTABLE_IO_ERROR;
1444 }
1445
1446 while ((d = readdir(dir))) {
1447 int i = 0;
1448 int found = 0;
1449 if (!is_table_name(d->d_name))
1450 continue;
1451
1452 for (i = 0; !found && i < st->readers_len; i++) {
1453 found = !strcmp(reader_name(st->readers[i]), d->d_name);
1454 }
1455 if (found)
1456 continue;
1457
1458 remove_maybe_stale_table(st, max, d->d_name);
1459 }
1460
1461 closedir(dir);
1462 return 0;
1463 }
1464
1465 int reftable_stack_clean(struct reftable_stack *st)
1466 {
1467 struct reftable_addition *add = NULL;
1468 int err = reftable_stack_new_addition(&add, st);
1469 if (err < 0) {
1470 goto done;
1471 }
1472
1473 err = reftable_stack_reload(st);
1474 if (err < 0) {
1475 goto done;
1476 }
1477
1478 err = reftable_stack_clean_locked(st);
1479
1480 done:
1481 reftable_addition_destroy(add);
1482 return err;
1483 }
1484
1485 int reftable_stack_print_directory(const char *stackdir, uint32_t hash_id)
1486 {
1487 struct reftable_stack *stack = NULL;
1488 struct reftable_write_options cfg = { .hash_id = hash_id };
1489 struct reftable_merged_table *merged = NULL;
1490 struct reftable_table table = { NULL };
1491
1492 int err = reftable_new_stack(&stack, stackdir, cfg);
1493 if (err < 0)
1494 goto done;
1495
1496 merged = reftable_stack_merged_table(stack);
1497 reftable_table_from_merged_table(&table, merged);
1498 err = reftable_table_print(&table);
1499 done:
1500 if (stack)
1501 reftable_stack_destroy(stack);
1502 return err;
1503 }