1 /* SPDX-License-Identifier: LGPL-2.1-or-later */
6 #include <linux/magic.h>
9 #include <sys/inotify.h>
13 #include "sd-journal.h"
15 #include "alloc-util.h"
18 #include "dirent-util.h"
23 #include "format-util.h"
26 #include "hostname-util.h"
27 #include "id128-util.h"
28 #include "inotify-util.h"
30 #include "journal-def.h"
31 #include "journal-file.h"
32 #include "journal-internal.h"
35 #include "nulstr-util.h"
36 #include "path-util.h"
37 #include "process-util.h"
38 #include "replace-var.h"
39 #include "stat-util.h"
40 #include "stdio-util.h"
41 #include "string-util.h"
43 #include "syslog-util.h"
45 #define JOURNAL_FILES_MAX 7168
47 #define JOURNAL_FILES_RECHECK_USEC (2 * USEC_PER_SEC)
49 /* The maximum size of variable values we'll expand in catalog entries. We bind this to PATH_MAX for now, as
50 * we want to be able to show all officially valid paths at least */
51 #define REPLACE_VAR_MAX PATH_MAX
53 #define DEFAULT_DATA_THRESHOLD (64*1024)
55 static void remove_file_real(sd_journal
*j
, JournalFile
*f
);
57 static bool journal_pid_changed(sd_journal
*j
) {
60 /* We don't support people creating a journal object and
61 * keeping it around over a fork(). Let's complain. */
63 return j
->original_pid
!= getpid_cached();
66 static int journal_put_error(sd_journal
*j
, int r
, const char *path
) {
67 _cleanup_free_
char *copy
= NULL
;
70 /* Memorize an error we encountered, and store which
71 * file/directory it was generated from. Note that we store
72 * only *one* path per error code, as the error code is the
73 * key into the hashmap, and the path is the value. This means
74 * we keep track only of all error kinds, but not of all error
75 * locations. This has the benefit that the hashmap cannot
78 * We return an error here only if we didn't manage to
79 * memorize the real error. */
90 k
= hashmap_ensure_put(&j
->errors
, NULL
, INT_TO_PTR(r
), copy
);
102 static void detach_location(sd_journal
*j
) {
107 j
->current_file
= NULL
;
108 j
->current_field
= 0;
110 ORDERED_HASHMAP_FOREACH(f
, j
->files
)
111 journal_file_reset_location(f
);
114 static void init_location(Location
*l
, LocationType type
, JournalFile
*f
, Object
*o
) {
116 assert(IN_SET(type
, LOCATION_DISCRETE
, LOCATION_SEEK
));
121 .seqnum
= le64toh(o
->entry
.seqnum
),
122 .seqnum_id
= f
->header
->seqnum_id
,
123 .realtime
= le64toh(o
->entry
.realtime
),
124 .monotonic
= le64toh(o
->entry
.monotonic
),
125 .boot_id
= o
->entry
.boot_id
,
126 .xor_hash
= le64toh(o
->entry
.xor_hash
),
128 .realtime_set
= true,
129 .monotonic_set
= true,
130 .xor_hash_set
= true,
134 static void set_location(sd_journal
*j
, JournalFile
*f
, Object
*o
) {
139 init_location(&j
->current_location
, LOCATION_DISCRETE
, f
, o
);
142 j
->current_field
= 0;
144 /* Let f know its candidate entry was picked. */
145 assert(f
->location_type
== LOCATION_SEEK
);
146 f
->location_type
= LOCATION_DISCRETE
;
149 static int match_is_valid(const void *data
, size_t size
) {
157 if (((char*) data
)[0] == '_' && ((char*) data
)[1] == '_')
161 for (p
= b
; p
< b
+ size
; p
++) {
169 if (*p
>= 'A' && *p
<= 'Z')
172 if (*p
>= '0' && *p
<= '9')
181 static bool same_field(const void *_a
, size_t s
, const void *_b
, size_t t
) {
182 const uint8_t *a
= _a
, *b
= _b
;
185 for (j
= 0; j
< s
&& j
< t
; j
++) {
194 assert_not_reached();
197 static Match
*match_new(Match
*p
, MatchType t
) {
210 LIST_PREPEND(matches
, p
->matches
, m
);
215 static Match
*match_free(Match
*m
) {
219 match_free(m
->matches
);
222 LIST_REMOVE(matches
, m
->parent
->matches
, m
);
228 static Match
*match_free_if_empty(Match
*m
) {
229 if (!m
|| m
->matches
)
232 return match_free(m
);
235 _public_
int sd_journal_add_match(sd_journal
*j
, const void *data
, size_t size
) {
236 Match
*l3
, *l4
, *add_here
= NULL
, *m
= NULL
;
239 assert_return(j
, -EINVAL
);
240 assert_return(!journal_pid_changed(j
), -ECHILD
);
241 assert_return(data
, -EINVAL
);
246 assert_return(match_is_valid(data
, size
), -EINVAL
);
252 * level 4: concrete matches */
255 j
->level0
= match_new(NULL
, MATCH_AND_TERM
);
261 j
->level1
= match_new(j
->level0
, MATCH_OR_TERM
);
267 j
->level2
= match_new(j
->level1
, MATCH_AND_TERM
);
272 assert(j
->level0
->type
== MATCH_AND_TERM
);
273 assert(j
->level1
->type
== MATCH_OR_TERM
);
274 assert(j
->level2
->type
== MATCH_AND_TERM
);
276 /* Old-style Jenkins (unkeyed) hashing only here. We do not cover new-style siphash (keyed) hashing
277 * here, since it's different for each file, and thus can't be pre-calculated in the Match object. */
278 hash
= jenkins_hash64(data
, size
);
280 LIST_FOREACH(matches
, l3
, j
->level2
->matches
) {
281 assert(l3
->type
== MATCH_OR_TERM
);
283 LIST_FOREACH(matches
, l4
, l3
->matches
) {
284 assert(l4
->type
== MATCH_DISCRETE
);
286 /* Exactly the same match already? Then ignore
288 if (l4
->hash
== hash
&&
290 memcmp(l4
->data
, data
, size
) == 0)
293 /* Same field? Then let's add this to this OR term */
294 if (same_field(data
, size
, l4
->data
, l4
->size
)) {
305 add_here
= match_new(j
->level2
, MATCH_OR_TERM
);
310 m
= match_new(add_here
, MATCH_DISCRETE
);
316 m
->data
= memdup(data
, size
);
326 match_free_if_empty(add_here
);
327 j
->level2
= match_free_if_empty(j
->level2
);
328 j
->level1
= match_free_if_empty(j
->level1
);
329 j
->level0
= match_free_if_empty(j
->level0
);
334 _public_
int sd_journal_add_conjunction(sd_journal
*j
) {
335 assert_return(j
, -EINVAL
);
336 assert_return(!journal_pid_changed(j
), -ECHILD
);
344 if (!j
->level1
->matches
)
353 _public_
int sd_journal_add_disjunction(sd_journal
*j
) {
354 assert_return(j
, -EINVAL
);
355 assert_return(!journal_pid_changed(j
), -ECHILD
);
366 if (!j
->level2
->matches
)
373 static char *match_make_string(Match
*m
) {
376 bool enclose
= false;
379 return strdup("none");
381 if (m
->type
== MATCH_DISCRETE
)
382 return cescape_length(m
->data
, m
->size
);
384 LIST_FOREACH(matches
, i
, m
->matches
) {
387 t
= match_make_string(i
);
392 k
= strjoin(p
, m
->type
== MATCH_OR_TERM
? " OR " : " AND ", t
);
407 r
= strjoin("(", p
, ")");
415 char *journal_make_match_string(sd_journal
*j
) {
418 return match_make_string(j
->level0
);
421 _public_
void sd_journal_flush_matches(sd_journal
*j
) {
426 match_free(j
->level0
);
428 j
->level0
= j
->level1
= j
->level2
= NULL
;
433 _pure_
static int compare_with_location(const JournalFile
*f
, const Location
*l
, const JournalFile
*current_file
) {
438 assert(f
->location_type
== LOCATION_SEEK
);
439 assert(IN_SET(l
->type
, LOCATION_DISCRETE
, LOCATION_SEEK
));
441 if (l
->monotonic_set
&&
442 sd_id128_equal(f
->current_boot_id
, l
->boot_id
) &&
444 f
->current_realtime
== l
->realtime
&&
446 f
->current_xor_hash
== l
->xor_hash
&&
448 sd_id128_equal(f
->header
->seqnum_id
, l
->seqnum_id
) &&
449 f
->current_seqnum
== l
->seqnum
&&
454 sd_id128_equal(f
->header
->seqnum_id
, l
->seqnum_id
)) {
456 r
= CMP(f
->current_seqnum
, l
->seqnum
);
461 if (l
->monotonic_set
&&
462 sd_id128_equal(f
->current_boot_id
, l
->boot_id
)) {
464 r
= CMP(f
->current_monotonic
, l
->monotonic
);
469 if (l
->realtime_set
) {
471 r
= CMP(f
->current_realtime
, l
->realtime
);
476 if (l
->xor_hash_set
) {
478 r
= CMP(f
->current_xor_hash
, l
->xor_hash
);
486 static int next_for_match(
490 uint64_t after_offset
,
491 direction_t direction
,
503 if (m
->type
== MATCH_DISCRETE
) {
506 /* If the keyed hash logic is used, we need to calculate the hash fresh per file. Otherwise
507 * we can use what we pre-calculated. */
508 if (JOURNAL_HEADER_KEYED_HASH(f
->header
))
509 hash
= journal_file_hash_data(f
, m
->data
, m
->size
);
513 r
= journal_file_find_data_object_with_hash(f
, m
->data
, m
->size
, hash
, NULL
, &dp
);
517 return journal_file_move_to_entry_by_offset_for_data(f
, dp
, after_offset
, direction
, ret
, offset
);
519 } else if (m
->type
== MATCH_OR_TERM
) {
522 /* Find the earliest match beyond after_offset */
524 LIST_FOREACH(matches
, i
, m
->matches
) {
527 r
= next_for_match(j
, i
, f
, after_offset
, direction
, NULL
, &cp
);
531 if (np
== 0 || (direction
== DIRECTION_DOWN
? cp
< np
: cp
> np
))
539 } else if (m
->type
== MATCH_AND_TERM
) {
540 Match
*i
, *last_moved
;
542 /* Always jump to the next matching entry and repeat
543 * this until we find an offset that matches for all
549 r
= next_for_match(j
, m
->matches
, f
, after_offset
, direction
, NULL
, &np
);
553 assert(direction
== DIRECTION_DOWN
? np
>= after_offset
: np
<= after_offset
);
554 last_moved
= m
->matches
;
556 LIST_LOOP_BUT_ONE(matches
, i
, m
->matches
, last_moved
) {
559 r
= next_for_match(j
, i
, f
, np
, direction
, NULL
, &cp
);
563 assert(direction
== DIRECTION_DOWN
? cp
>= np
: cp
<= np
);
564 if (direction
== DIRECTION_DOWN
? cp
> np
: cp
< np
) {
573 r
= journal_file_move_to_object(f
, OBJECT_ENTRY
, np
, &n
);
585 static int find_location_for_match(
589 direction_t direction
,
599 if (m
->type
== MATCH_DISCRETE
) {
602 if (JOURNAL_HEADER_KEYED_HASH(f
->header
))
603 hash
= journal_file_hash_data(f
, m
->data
, m
->size
);
607 r
= journal_file_find_data_object_with_hash(f
, m
->data
, m
->size
, hash
, NULL
, &dp
);
611 /* FIXME: missing: find by monotonic */
613 if (j
->current_location
.type
== LOCATION_HEAD
)
614 return journal_file_next_entry_for_data(f
, dp
, DIRECTION_DOWN
, ret
, offset
);
615 if (j
->current_location
.type
== LOCATION_TAIL
)
616 return journal_file_next_entry_for_data(f
, dp
, DIRECTION_UP
, ret
, offset
);
617 if (j
->current_location
.seqnum_set
&& sd_id128_equal(j
->current_location
.seqnum_id
, f
->header
->seqnum_id
))
618 return journal_file_move_to_entry_by_seqnum_for_data(f
, dp
, j
->current_location
.seqnum
, direction
, ret
, offset
);
619 if (j
->current_location
.monotonic_set
) {
620 r
= journal_file_move_to_entry_by_monotonic_for_data(f
, dp
, j
->current_location
.boot_id
, j
->current_location
.monotonic
, direction
, ret
, offset
);
624 if (j
->current_location
.realtime_set
)
625 return journal_file_move_to_entry_by_realtime_for_data(f
, dp
, j
->current_location
.realtime
, direction
, ret
, offset
);
627 return journal_file_next_entry_for_data(f
, dp
, direction
, ret
, offset
);
629 } else if (m
->type
== MATCH_OR_TERM
) {
634 /* Find the earliest match */
636 LIST_FOREACH(matches
, i
, m
->matches
) {
639 r
= find_location_for_match(j
, i
, f
, direction
, NULL
, &cp
);
643 if (np
== 0 || (direction
== DIRECTION_DOWN
? np
> cp
: np
< cp
))
651 r
= journal_file_move_to_object(f
, OBJECT_ENTRY
, np
, &n
);
666 assert(m
->type
== MATCH_AND_TERM
);
668 /* First jump to the last match, and then find the
669 * next one where all matches match */
674 LIST_FOREACH(matches
, i
, m
->matches
) {
677 r
= find_location_for_match(j
, i
, f
, direction
, NULL
, &cp
);
681 if (np
== 0 || (direction
== DIRECTION_DOWN
? cp
> np
: cp
< np
))
685 return next_for_match(j
, m
, f
, np
, direction
, ret
, offset
);
689 static int find_location_with_matches(
692 direction_t direction
,
704 /* No matches is simple */
706 if (j
->current_location
.type
== LOCATION_HEAD
)
707 return journal_file_next_entry(f
, 0, DIRECTION_DOWN
, ret
, offset
);
708 if (j
->current_location
.type
== LOCATION_TAIL
)
709 return journal_file_next_entry(f
, 0, DIRECTION_UP
, ret
, offset
);
710 if (j
->current_location
.seqnum_set
&& sd_id128_equal(j
->current_location
.seqnum_id
, f
->header
->seqnum_id
))
711 return journal_file_move_to_entry_by_seqnum(f
, j
->current_location
.seqnum
, direction
, ret
, offset
);
712 if (j
->current_location
.monotonic_set
) {
713 r
= journal_file_move_to_entry_by_monotonic(f
, j
->current_location
.boot_id
, j
->current_location
.monotonic
, direction
, ret
, offset
);
717 if (j
->current_location
.realtime_set
)
718 return journal_file_move_to_entry_by_realtime(f
, j
->current_location
.realtime
, direction
, ret
, offset
);
720 return journal_file_next_entry(f
, 0, direction
, ret
, offset
);
722 return find_location_for_match(j
, j
->level0
, f
, direction
, ret
, offset
);
725 static int next_with_matches(
728 direction_t direction
,
737 /* No matches is easy. We simple advance the file
740 return journal_file_next_entry(f
, f
->current_offset
, direction
, ret
, offset
);
742 /* If we have a match then we look for the next matching entry
743 * with an offset at least one step larger */
744 return next_for_match(j
, j
->level0
, f
,
745 direction
== DIRECTION_DOWN
? f
->current_offset
+ 1
746 : f
->current_offset
- 1,
747 direction
, ret
, offset
);
750 static int next_beyond_location(sd_journal
*j
, JournalFile
*f
, direction_t direction
) {
752 uint64_t cp
, n_entries
;
758 n_entries
= le64toh(f
->header
->n_entries
);
760 /* If we hit EOF before, we don't need to look into this file again
761 * unless direction changed or new entries appeared. */
762 if (f
->last_direction
== direction
&& f
->location_type
== LOCATION_TAIL
&&
763 n_entries
== f
->last_n_entries
)
766 f
->last_n_entries
= n_entries
;
768 if (f
->last_direction
== direction
&& f
->current_offset
> 0) {
769 /* LOCATION_SEEK here means we did the work in a previous
770 * iteration and the current location already points to a
771 * candidate entry. */
772 if (f
->location_type
!= LOCATION_SEEK
) {
773 r
= next_with_matches(j
, f
, direction
, &c
, &cp
);
777 journal_file_save_location(f
, c
, cp
);
780 f
->last_direction
= direction
;
782 r
= find_location_with_matches(j
, f
, direction
, &c
, &cp
);
786 journal_file_save_location(f
, c
, cp
);
789 /* OK, we found the spot, now let's advance until an entry
790 * that is actually different from what we were previously
791 * looking at. This is necessary to handle entries which exist
792 * in two (or more) journal files, and which shall all be
793 * suppressed but one. */
798 if (j
->current_location
.type
== LOCATION_DISCRETE
) {
801 k
= compare_with_location(f
, &j
->current_location
, j
->current_file
);
803 found
= direction
== DIRECTION_DOWN
? k
> 0 : k
< 0;
810 r
= next_with_matches(j
, f
, direction
, &c
, &cp
);
814 journal_file_save_location(f
, c
, cp
);
818 static int real_journal_next(sd_journal
*j
, direction_t direction
) {
819 JournalFile
*new_file
= NULL
;
825 assert_return(j
, -EINVAL
);
826 assert_return(!journal_pid_changed(j
), -ECHILD
);
828 r
= iterated_cache_get(j
->files_cache
, NULL
, &files
, &n_files
);
832 for (i
= 0; i
< n_files
; i
++) {
833 JournalFile
*f
= (JournalFile
*)files
[i
];
836 r
= next_beyond_location(j
, f
, direction
);
838 log_debug_errno(r
, "Can't iterate through %s, ignoring: %m", f
->path
);
839 remove_file_real(j
, f
);
842 f
->location_type
= LOCATION_TAIL
;
851 k
= journal_file_compare_locations(f
, new_file
);
853 found
= direction
== DIRECTION_DOWN
? k
< 0 : k
> 0;
863 r
= journal_file_move_to_object(new_file
, OBJECT_ENTRY
, new_file
->current_offset
, &o
);
867 set_location(j
, new_file
, o
);
872 _public_
int sd_journal_next(sd_journal
*j
) {
873 return real_journal_next(j
, DIRECTION_DOWN
);
876 _public_
int sd_journal_previous(sd_journal
*j
) {
877 return real_journal_next(j
, DIRECTION_UP
);
880 static int real_journal_next_skip(sd_journal
*j
, direction_t direction
, uint64_t skip
) {
883 assert_return(j
, -EINVAL
);
884 assert_return(!journal_pid_changed(j
), -ECHILD
);
885 assert_return(skip
<= INT_MAX
, -ERANGE
);
888 /* If this is not a discrete skip, then at least
889 * resolve the current location */
890 if (j
->current_location
.type
!= LOCATION_DISCRETE
) {
891 r
= real_journal_next(j
, direction
);
900 r
= real_journal_next(j
, direction
);
914 _public_
int sd_journal_next_skip(sd_journal
*j
, uint64_t skip
) {
915 return real_journal_next_skip(j
, DIRECTION_DOWN
, skip
);
918 _public_
int sd_journal_previous_skip(sd_journal
*j
, uint64_t skip
) {
919 return real_journal_next_skip(j
, DIRECTION_UP
, skip
);
922 _public_
int sd_journal_get_cursor(sd_journal
*j
, char **cursor
) {
926 assert_return(j
, -EINVAL
);
927 assert_return(!journal_pid_changed(j
), -ECHILD
);
928 assert_return(cursor
, -EINVAL
);
930 if (!j
->current_file
|| j
->current_file
->current_offset
<= 0)
931 return -EADDRNOTAVAIL
;
933 r
= journal_file_move_to_object(j
->current_file
, OBJECT_ENTRY
, j
->current_file
->current_offset
, &o
);
938 "s=%s;i=%"PRIx64
";b=%s;m=%"PRIx64
";t=%"PRIx64
";x=%"PRIx64
,
939 SD_ID128_TO_STRING(j
->current_file
->header
->seqnum_id
), le64toh(o
->entry
.seqnum
),
940 SD_ID128_TO_STRING(o
->entry
.boot_id
), le64toh(o
->entry
.monotonic
),
941 le64toh(o
->entry
.realtime
),
942 le64toh(o
->entry
.xor_hash
)) < 0)
948 _public_
int sd_journal_seek_cursor(sd_journal
*j
, const char *cursor
) {
949 unsigned long long seqnum
, monotonic
, realtime
, xor_hash
;
950 bool seqnum_id_set
= false,
953 monotonic_set
= false,
954 realtime_set
= false,
955 xor_hash_set
= false;
956 sd_id128_t seqnum_id
, boot_id
;
959 assert_return(j
, -EINVAL
);
960 assert_return(!journal_pid_changed(j
), -ECHILD
);
961 assert_return(!isempty(cursor
), -EINVAL
);
963 for (const char *p
= cursor
;;) {
964 _cleanup_free_
char *word
= NULL
;
966 r
= extract_first_word(&p
, &word
, ";", EXTRACT_DONT_COALESCE_SEPARATORS
);
972 if (word
[0] == '\0' || word
[1] != '=')
977 seqnum_id_set
= true;
978 r
= sd_id128_from_string(word
+ 2, &seqnum_id
);
985 if (sscanf(word
+ 2, "%llx", &seqnum
) != 1)
991 r
= sd_id128_from_string(word
+ 2, &boot_id
);
995 monotonic_set
= true;
996 if (sscanf(word
+ 2, "%llx", &monotonic
) != 1)
1001 realtime_set
= true;
1002 if (sscanf(word
+ 2, "%llx", &realtime
) != 1)
1007 xor_hash_set
= true;
1008 if (sscanf(word
+ 2, "%llx", &xor_hash
) != 1)
1014 if ((!seqnum_set
|| !seqnum_id_set
) &&
1015 (!monotonic_set
|| !boot_id_set
) &&
1020 j
->current_location
= (Location
) {
1021 .type
= LOCATION_SEEK
,
1025 j
->current_location
.realtime
= (uint64_t) realtime
;
1026 j
->current_location
.realtime_set
= true;
1029 if (seqnum_set
&& seqnum_id_set
) {
1030 j
->current_location
.seqnum
= (uint64_t) seqnum
;
1031 j
->current_location
.seqnum_id
= seqnum_id
;
1032 j
->current_location
.seqnum_set
= true;
1035 if (monotonic_set
&& boot_id_set
) {
1036 j
->current_location
.monotonic
= (uint64_t) monotonic
;
1037 j
->current_location
.boot_id
= boot_id
;
1038 j
->current_location
.monotonic_set
= true;
1042 j
->current_location
.xor_hash
= (uint64_t) xor_hash
;
1043 j
->current_location
.xor_hash_set
= true;
1049 _public_
int sd_journal_test_cursor(sd_journal
*j
, const char *cursor
) {
1053 assert_return(j
, -EINVAL
);
1054 assert_return(!journal_pid_changed(j
), -ECHILD
);
1055 assert_return(!isempty(cursor
), -EINVAL
);
1057 if (!j
->current_file
|| j
->current_file
->current_offset
<= 0)
1058 return -EADDRNOTAVAIL
;
1060 r
= journal_file_move_to_object(j
->current_file
, OBJECT_ENTRY
, j
->current_file
->current_offset
, &o
);
1065 _cleanup_free_
char *item
= NULL
;
1066 unsigned long long ll
;
1070 r
= extract_first_word(&cursor
, &item
, ";", EXTRACT_DONT_COALESCE_SEPARATORS
);
1077 if (strlen(item
) < 2 || item
[1] != '=')
1083 k
= sd_id128_from_string(item
+2, &id
);
1086 if (!sd_id128_equal(id
, j
->current_file
->header
->seqnum_id
))
1091 if (sscanf(item
+2, "%llx", &ll
) != 1)
1093 if (ll
!= le64toh(o
->entry
.seqnum
))
1098 k
= sd_id128_from_string(item
+2, &id
);
1101 if (!sd_id128_equal(id
, o
->entry
.boot_id
))
1106 if (sscanf(item
+2, "%llx", &ll
) != 1)
1108 if (ll
!= le64toh(o
->entry
.monotonic
))
1113 if (sscanf(item
+2, "%llx", &ll
) != 1)
1115 if (ll
!= le64toh(o
->entry
.realtime
))
1120 if (sscanf(item
+2, "%llx", &ll
) != 1)
1122 if (ll
!= le64toh(o
->entry
.xor_hash
))
1131 _public_
int sd_journal_seek_monotonic_usec(sd_journal
*j
, sd_id128_t boot_id
, uint64_t usec
) {
1132 assert_return(j
, -EINVAL
);
1133 assert_return(!journal_pid_changed(j
), -ECHILD
);
1137 j
->current_location
= (Location
) {
1138 .type
= LOCATION_SEEK
,
1141 .monotonic_set
= true,
1147 _public_
int sd_journal_seek_realtime_usec(sd_journal
*j
, uint64_t usec
) {
1148 assert_return(j
, -EINVAL
);
1149 assert_return(!journal_pid_changed(j
), -ECHILD
);
1153 j
->current_location
= (Location
) {
1154 .type
= LOCATION_SEEK
,
1156 .realtime_set
= true,
1162 _public_
int sd_journal_seek_head(sd_journal
*j
) {
1163 assert_return(j
, -EINVAL
);
1164 assert_return(!journal_pid_changed(j
), -ECHILD
);
1168 j
->current_location
= (Location
) {
1169 .type
= LOCATION_HEAD
,
1175 _public_
int sd_journal_seek_tail(sd_journal
*j
) {
1176 assert_return(j
, -EINVAL
);
1177 assert_return(!journal_pid_changed(j
), -ECHILD
);
1181 j
->current_location
= (Location
) {
1182 .type
= LOCATION_TAIL
,
1188 static void check_network(sd_journal
*j
, int fd
) {
1194 j
->on_network
= fd_is_network_fs(fd
);
1197 static bool file_has_type_prefix(const char *prefix
, const char *filename
) {
1198 const char *full
, *tilded
, *atted
;
1200 full
= strjoina(prefix
, ".journal");
1201 tilded
= strjoina(full
, "~");
1202 atted
= strjoina(prefix
, "@");
1204 return STR_IN_SET(filename
, full
, tilded
) ||
1205 startswith(filename
, atted
);
1208 static bool file_type_wanted(int flags
, const char *filename
) {
1211 if (!endswith(filename
, ".journal") && !endswith(filename
, ".journal~"))
1214 /* no flags set → every type is OK */
1215 if (!(flags
& (SD_JOURNAL_SYSTEM
| SD_JOURNAL_CURRENT_USER
)))
1218 if (flags
& SD_JOURNAL_SYSTEM
&& file_has_type_prefix("system", filename
))
1221 if (flags
& SD_JOURNAL_CURRENT_USER
) {
1222 char prefix
[5 + DECIMAL_STR_MAX(uid_t
) + 1];
1224 xsprintf(prefix
, "user-"UID_FMT
, getuid());
1226 if (file_has_type_prefix(prefix
, filename
))
1233 static bool path_has_prefix(sd_journal
*j
, const char *path
, const char *prefix
) {
1238 if (j
->toplevel_fd
>= 0)
1241 return path_startswith(path
, prefix
);
1244 static void track_file_disposition(sd_journal
*j
, JournalFile
*f
) {
1248 if (!j
->has_runtime_files
&& path_has_prefix(j
, f
->path
, "/run"))
1249 j
->has_runtime_files
= true;
1250 else if (!j
->has_persistent_files
&& path_has_prefix(j
, f
->path
, "/var"))
1251 j
->has_persistent_files
= true;
1254 static const char *skip_slash(const char *p
) {
1265 static int add_any_file(
1270 bool close_fd
= false;
1276 assert(fd
>= 0 || path
);
1279 if (j
->toplevel_fd
>= 0)
1280 /* If there's a top-level fd defined make the path relative, explicitly, since otherwise
1281 * openat() ignores the first argument. */
1283 fd
= openat(j
->toplevel_fd
, skip_slash(path
), O_RDONLY
|O_CLOEXEC
|O_NONBLOCK
);
1285 fd
= open(path
, O_RDONLY
|O_CLOEXEC
|O_NONBLOCK
);
1287 r
= log_debug_errno(errno
, "Failed to open journal file %s: %m", path
);
1293 r
= fd_nonblock(fd
, false);
1295 r
= log_debug_errno(errno
, "Failed to turn off O_NONBLOCK for %s: %m", path
);
1300 if (fstat(fd
, &st
) < 0) {
1301 r
= log_debug_errno(errno
, "Failed to fstat file '%s': %m", path
);
1305 r
= stat_verify_regular(&st
);
1307 log_debug_errno(r
, "Refusing to open '%s', as it is not a regular file.", path
);
1311 f
= ordered_hashmap_get(j
->files
, path
);
1313 if (f
->last_stat
.st_dev
== st
.st_dev
&&
1314 f
->last_stat
.st_ino
== st
.st_ino
) {
1316 /* We already track this file, under the same path and with the same device/inode numbers, it's
1317 * hence really the same. Mark this file as seen in this generation. This is used to GC old
1318 * files in process_q_overflow() to detect journal files that are still there and discern them
1319 * from those which are gone. */
1321 f
->last_seen_generation
= j
->generation
;
1326 /* So we tracked a file under this name, but it has a different inode/device. In that case, it got
1327 * replaced (probably due to rotation?), let's drop it hence from our list. */
1328 remove_file_real(j
, f
);
1332 if (ordered_hashmap_size(j
->files
) >= JOURNAL_FILES_MAX
) {
1333 log_debug("Too many open journal files, not adding %s.", path
);
1338 r
= journal_file_open(fd
, path
, O_RDONLY
, 0, false, 0, false, NULL
, j
->mmap
, NULL
, &f
);
1340 log_debug_errno(r
, "Failed to open journal file %s: %m", path
);
1344 /* journal_file_dump(f); */
1346 r
= ordered_hashmap_put(j
->files
, f
->path
, f
);
1348 f
->close_fd
= false; /* make sure journal_file_close() doesn't close the caller's fd (or our own). We'll let the caller do that, or ourselves */
1349 (void) journal_file_close(f
);
1353 close_fd
= false; /* the fd is now owned by the JournalFile object */
1355 f
->last_seen_generation
= j
->generation
;
1357 track_file_disposition(j
, f
);
1358 check_network(j
, f
->fd
);
1360 j
->current_invalidate_counter
++;
1362 log_debug("File %s added.", f
->path
);
1371 k
= journal_put_error(j
, r
, path
);
1379 static int add_file_by_name(
1382 const char *filename
) {
1390 if (j
->no_new_files
)
1393 if (!file_type_wanted(j
->flags
, filename
))
1396 path
= prefix_roota(prefix
, filename
);
1397 return add_any_file(j
, -1, path
);
1400 static void remove_file_by_name(
1403 const char *filename
) {
1412 path
= prefix_roota(prefix
, filename
);
1413 f
= ordered_hashmap_get(j
->files
, path
);
1417 remove_file_real(j
, f
);
1420 static void remove_file_real(sd_journal
*j
, JournalFile
*f
) {
1424 (void) ordered_hashmap_remove(j
->files
, f
->path
);
1426 log_debug("File %s removed.", f
->path
);
1428 if (j
->current_file
== f
) {
1429 j
->current_file
= NULL
;
1430 j
->current_field
= 0;
1433 if (j
->unique_file
== f
) {
1434 /* Jump to the next unique_file or NULL if that one was last */
1435 j
->unique_file
= ordered_hashmap_next(j
->files
, j
->unique_file
->path
);
1436 j
->unique_offset
= 0;
1437 if (!j
->unique_file
)
1438 j
->unique_file_lost
= true;
1441 if (j
->fields_file
== f
) {
1442 j
->fields_file
= ordered_hashmap_next(j
->files
, j
->fields_file
->path
);
1443 j
->fields_offset
= 0;
1444 if (!j
->fields_file
)
1445 j
->fields_file_lost
= true;
1448 (void) journal_file_close(f
);
1450 j
->current_invalidate_counter
++;
1453 static int dirname_is_machine_id(const char *fn
) {
1454 sd_id128_t id
, machine
;
1458 /* Returns true if the specified directory name matches the local machine ID */
1460 r
= sd_id128_get_machine(&machine
);
1464 e
= strchr(fn
, '.');
1468 /* Looks like it has a namespace suffix. Verify that. */
1469 if (!log_namespace_name_valid(e
+ 1))
1472 k
= strndupa_safe(fn
, e
- fn
);
1473 r
= sd_id128_from_string(k
, &id
);
1475 r
= sd_id128_from_string(fn
, &id
);
1479 return sd_id128_equal(id
, machine
);
1482 static int dirname_has_namespace(const char *fn
, const char *namespace) {
1485 /* Returns true if the specified directory name matches the specified namespace */
1487 e
= strchr(fn
, '.');
1494 if (!streq(e
+ 1, namespace))
1497 k
= strndupa_safe(fn
, e
- fn
);
1498 return id128_is_valid(k
);
1504 return id128_is_valid(fn
);
1507 static bool dirent_is_journal_file(const struct dirent
*de
) {
1510 /* Returns true if the specified directory entry looks like a journal file we might be interested in */
1512 if (!IN_SET(de
->d_type
, DT_REG
, DT_LNK
, DT_UNKNOWN
))
1515 return endswith(de
->d_name
, ".journal") ||
1516 endswith(de
->d_name
, ".journal~");
1519 static bool dirent_is_journal_subdir(const struct dirent
*de
) {
1523 /* returns true if the specified directory entry looks like a directory that might contain journal
1524 * files we might be interested in, i.e. is either a 128bit ID or a 128bit ID suffixed by a
1527 if (!IN_SET(de
->d_type
, DT_DIR
, DT_LNK
, DT_UNKNOWN
))
1530 e
= strchr(de
->d_name
, '.');
1532 return id128_is_valid(de
->d_name
); /* No namespace */
1534 n
= strndupa_safe(de
->d_name
, e
- de
->d_name
);
1535 if (!id128_is_valid(n
))
1538 return log_namespace_name_valid(e
+ 1);
1541 static int directory_open(sd_journal
*j
, const char *path
, DIR **ret
) {
1548 if (j
->toplevel_fd
< 0)
1551 /* Open the specified directory relative to the toplevel fd. Enforce that the path specified is
1552 * relative, by dropping the initial slash */
1553 d
= xopendirat(j
->toplevel_fd
, skip_slash(path
), 0);
1561 static int add_directory(sd_journal
*j
, const char *prefix
, const char *dirname
);
1563 static void directory_enumerate(sd_journal
*j
, Directory
*m
, DIR *d
) {
1568 FOREACH_DIRENT_ALL(de
, d
, goto fail
) {
1569 if (dirent_is_journal_file(de
))
1570 (void) add_file_by_name(j
, m
->path
, de
->d_name
);
1572 if (m
->is_root
&& dirent_is_journal_subdir(de
))
1573 (void) add_directory(j
, m
->path
, de
->d_name
);
1578 log_debug_errno(errno
, "Failed to enumerate directory %s, ignoring: %m", m
->path
);
1581 static void directory_watch(sd_journal
*j
, Directory
*m
, int fd
, uint32_t mask
) {
1588 /* Watch this directory if that's enabled and if it not being watched yet. */
1590 if (m
->wd
> 0) /* Already have a watch? */
1592 if (j
->inotify_fd
< 0) /* Not watching at all? */
1595 m
->wd
= inotify_add_watch_fd(j
->inotify_fd
, fd
, mask
);
1597 log_debug_errno(errno
, "Failed to watch journal directory '%s', ignoring: %m", m
->path
);
1601 r
= hashmap_put(j
->directories_by_wd
, INT_TO_PTR(m
->wd
), m
);
1603 log_debug_errno(r
, "Directory '%s' already being watched under a different path, ignoring: %m", m
->path
);
1605 log_debug_errno(r
, "Failed to add watch for journal directory '%s' to hashmap, ignoring: %m", m
->path
);
1606 (void) inotify_rm_watch(j
->inotify_fd
, m
->wd
);
1611 static int add_directory(
1614 const char *dirname
) {
1616 _cleanup_free_
char *path
= NULL
;
1617 _cleanup_closedir_
DIR *d
= NULL
;
1624 /* Adds a journal file directory to watch. If the directory is already tracked this updates the inotify watch
1625 * and reenumerates directory contents */
1627 path
= path_join(prefix
, dirname
);
1633 log_debug("Considering directory '%s'.", path
);
1635 /* We consider everything local that is in a directory for the local machine ID, or that is stored in /run */
1636 if ((j
->flags
& SD_JOURNAL_LOCAL_ONLY
) &&
1637 !((dirname
&& dirname_is_machine_id(dirname
) > 0) || path_has_prefix(j
, path
, "/run")))
1641 (!(FLAGS_SET(j
->flags
, SD_JOURNAL_ALL_NAMESPACES
) ||
1642 dirname_has_namespace(dirname
, j
->namespace) > 0 ||
1643 (FLAGS_SET(j
->flags
, SD_JOURNAL_INCLUDE_DEFAULT_NAMESPACE
) && dirname_has_namespace(dirname
, NULL
) > 0))))
1646 r
= directory_open(j
, path
, &d
);
1648 log_debug_errno(r
, "Failed to open directory '%s': %m", path
);
1652 m
= hashmap_get(j
->directories_by_path
, path
);
1654 m
= new(Directory
, 1);
1665 if (hashmap_put(j
->directories_by_path
, m
->path
, m
) < 0) {
1671 path
= NULL
; /* avoid freeing in cleanup */
1672 j
->current_invalidate_counter
++;
1674 log_debug("Directory %s added.", m
->path
);
1676 } else if (m
->is_root
)
1677 return 0; /* Don't 'downgrade' from root directory */
1679 m
->last_seen_generation
= j
->generation
;
1681 directory_watch(j
, m
, dirfd(d
),
1682 IN_CREATE
|IN_MOVED_TO
|IN_MODIFY
|IN_ATTRIB
|IN_DELETE
|
1683 IN_DELETE_SELF
|IN_MOVE_SELF
|IN_UNMOUNT
|IN_MOVED_FROM
|
1686 if (!j
->no_new_files
)
1687 directory_enumerate(j
, m
, d
);
1689 check_network(j
, dirfd(d
));
1694 k
= journal_put_error(j
, r
, path
?: prefix
);
1701 static int add_root_directory(sd_journal
*j
, const char *p
, bool missing_ok
) {
1703 _cleanup_closedir_
DIR *d
= NULL
;
1709 /* Adds a root directory to our set of directories to use. If the root directory is already in the set, we
1710 * update the inotify logic, and renumerate the directory entries. This call may hence be called to initially
1711 * populate the set, as well as to update it later. */
1714 /* If there's a path specified, use it. */
1716 log_debug("Considering root directory '%s'.", p
);
1718 if ((j
->flags
& SD_JOURNAL_RUNTIME_ONLY
) &&
1719 !path_has_prefix(j
, p
, "/run"))
1723 p
= strjoina(j
->prefix
, p
);
1725 r
= directory_open(j
, p
, &d
);
1726 if (r
== -ENOENT
&& missing_ok
)
1729 log_debug_errno(r
, "Failed to open root directory %s: %m", p
);
1733 _cleanup_close_
int dfd
= -1;
1735 /* If there's no path specified, then we use the top-level fd itself. We duplicate the fd here, since
1736 * opendir() will take possession of the fd, and close it, which we don't want. */
1738 p
= "."; /* store this as "." in the directories hashmap */
1740 dfd
= fcntl(j
->toplevel_fd
, F_DUPFD_CLOEXEC
, 3);
1746 d
= take_fdopendir(&dfd
);
1755 m
= hashmap_get(j
->directories_by_path
, p
);
1757 m
= new0(Directory
, 1);
1765 m
->path
= strdup(p
);
1772 if (hashmap_put(j
->directories_by_path
, m
->path
, m
) < 0) {
1779 j
->current_invalidate_counter
++;
1781 log_debug("Root directory %s added.", m
->path
);
1783 } else if (!m
->is_root
)
1786 directory_watch(j
, m
, dirfd(d
),
1787 IN_CREATE
|IN_MOVED_TO
|IN_MODIFY
|IN_ATTRIB
|IN_DELETE
|
1790 if (!j
->no_new_files
)
1791 directory_enumerate(j
, m
, d
);
1793 check_network(j
, dirfd(d
));
1798 k
= journal_put_error(j
, r
, p
);
1805 static void remove_directory(sd_journal
*j
, Directory
*d
) {
1809 hashmap_remove(j
->directories_by_wd
, INT_TO_PTR(d
->wd
));
1811 if (j
->inotify_fd
>= 0)
1812 (void) inotify_rm_watch(j
->inotify_fd
, d
->wd
);
1815 hashmap_remove(j
->directories_by_path
, d
->path
);
1818 log_debug("Root directory %s removed.", d
->path
);
1820 log_debug("Directory %s removed.", d
->path
);
1826 static int add_search_paths(sd_journal
*j
) {
1828 static const char search_paths
[] =
1829 "/run/log/journal\0"
1830 "/var/log/journal\0";
1835 /* We ignore most errors here, since the idea is to only open
1836 * what's actually accessible, and ignore the rest. */
1838 NULSTR_FOREACH(p
, search_paths
)
1839 (void) add_root_directory(j
, p
, true);
1841 if (!(j
->flags
& SD_JOURNAL_LOCAL_ONLY
))
1842 (void) add_root_directory(j
, "/var/log/journal/remote", true);
1847 static int add_current_paths(sd_journal
*j
) {
1851 assert(j
->no_new_files
);
1853 /* Simply adds all directories for files we have open as directories. We don't expect errors here, so we
1854 * treat them as fatal. */
1856 ORDERED_HASHMAP_FOREACH(f
, j
->files
) {
1857 _cleanup_free_
char *dir
= NULL
;
1860 dir
= dirname_malloc(f
->path
);
1864 r
= add_directory(j
, dir
, NULL
);
1872 static int allocate_inotify(sd_journal
*j
) {
1875 if (j
->inotify_fd
< 0) {
1876 j
->inotify_fd
= inotify_init1(IN_NONBLOCK
|IN_CLOEXEC
);
1877 if (j
->inotify_fd
< 0)
1881 return hashmap_ensure_allocated(&j
->directories_by_wd
, NULL
);
1884 static sd_journal
*journal_new(int flags
, const char *path
, const char *namespace) {
1885 _cleanup_(sd_journal_closep
) sd_journal
*j
= NULL
;
1887 j
= new0(sd_journal
, 1);
1891 j
->original_pid
= getpid_cached();
1892 j
->toplevel_fd
= -1;
1895 j
->data_threshold
= DEFAULT_DATA_THRESHOLD
;
1904 if (flags
& SD_JOURNAL_OS_ROOT
)
1911 j
->namespace = strdup(namespace);
1916 j
->files
= ordered_hashmap_new(&path_hash_ops
);
1920 j
->files_cache
= ordered_hashmap_iterated_cache_new(j
->files
);
1921 j
->directories_by_path
= hashmap_new(&path_hash_ops
);
1922 j
->mmap
= mmap_cache_new();
1923 if (!j
->files_cache
|| !j
->directories_by_path
|| !j
->mmap
)
1929 #define OPEN_ALLOWED_FLAGS \
1930 (SD_JOURNAL_LOCAL_ONLY | \
1931 SD_JOURNAL_RUNTIME_ONLY | \
1932 SD_JOURNAL_SYSTEM | \
1933 SD_JOURNAL_CURRENT_USER | \
1934 SD_JOURNAL_ALL_NAMESPACES | \
1935 SD_JOURNAL_INCLUDE_DEFAULT_NAMESPACE)
1937 _public_
int sd_journal_open_namespace(sd_journal
**ret
, const char *namespace, int flags
) {
1938 _cleanup_(sd_journal_closep
) sd_journal
*j
= NULL
;
1941 assert_return(ret
, -EINVAL
);
1942 assert_return((flags
& ~OPEN_ALLOWED_FLAGS
) == 0, -EINVAL
);
1944 j
= journal_new(flags
, NULL
, namespace);
1948 r
= add_search_paths(j
);
1956 _public_
int sd_journal_open(sd_journal
**ret
, int flags
) {
1957 return sd_journal_open_namespace(ret
, NULL
, flags
);
1960 #define OPEN_CONTAINER_ALLOWED_FLAGS \
1961 (SD_JOURNAL_LOCAL_ONLY | SD_JOURNAL_SYSTEM)
1963 _public_
int sd_journal_open_container(sd_journal
**ret
, const char *machine
, int flags
) {
1964 _cleanup_free_
char *root
= NULL
, *class = NULL
;
1965 _cleanup_(sd_journal_closep
) sd_journal
*j
= NULL
;
1969 /* This is deprecated, people should use machined's OpenMachineRootDirectory() call instead in
1970 * combination with sd_journal_open_directory_fd(). */
1972 assert_return(machine
, -EINVAL
);
1973 assert_return(ret
, -EINVAL
);
1974 assert_return((flags
& ~OPEN_CONTAINER_ALLOWED_FLAGS
) == 0, -EINVAL
);
1975 assert_return(hostname_is_valid(machine
, 0), -EINVAL
);
1977 p
= strjoina("/run/systemd/machines/", machine
);
1978 r
= parse_env_file(NULL
, p
,
1988 if (!streq_ptr(class, "container"))
1991 j
= journal_new(flags
, root
, NULL
);
1995 r
= add_search_paths(j
);
2003 #define OPEN_DIRECTORY_ALLOWED_FLAGS \
2004 (SD_JOURNAL_OS_ROOT | \
2005 SD_JOURNAL_SYSTEM | SD_JOURNAL_CURRENT_USER )
2007 _public_
int sd_journal_open_directory(sd_journal
**ret
, const char *path
, int flags
) {
2008 _cleanup_(sd_journal_closep
) sd_journal
*j
= NULL
;
2011 assert_return(ret
, -EINVAL
);
2012 assert_return(path
, -EINVAL
);
2013 assert_return((flags
& ~OPEN_DIRECTORY_ALLOWED_FLAGS
) == 0, -EINVAL
);
2015 j
= journal_new(flags
, path
, NULL
);
2019 if (flags
& SD_JOURNAL_OS_ROOT
)
2020 r
= add_search_paths(j
);
2022 r
= add_root_directory(j
, path
, false);
2030 _public_
int sd_journal_open_files(sd_journal
**ret
, const char **paths
, int flags
) {
2031 _cleanup_(sd_journal_closep
) sd_journal
*j
= NULL
;
2035 assert_return(ret
, -EINVAL
);
2036 assert_return(flags
== 0, -EINVAL
);
2038 j
= journal_new(flags
, NULL
, NULL
);
2042 STRV_FOREACH(path
, paths
) {
2043 r
= add_any_file(j
, -1, *path
);
2048 j
->no_new_files
= true;
2054 #define OPEN_DIRECTORY_FD_ALLOWED_FLAGS \
2055 (SD_JOURNAL_OS_ROOT | \
2056 SD_JOURNAL_SYSTEM | SD_JOURNAL_CURRENT_USER )
2058 _public_
int sd_journal_open_directory_fd(sd_journal
**ret
, int fd
, int flags
) {
2059 _cleanup_(sd_journal_closep
) sd_journal
*j
= NULL
;
2063 assert_return(ret
, -EINVAL
);
2064 assert_return(fd
>= 0, -EBADF
);
2065 assert_return((flags
& ~OPEN_DIRECTORY_FD_ALLOWED_FLAGS
) == 0, -EINVAL
);
2067 if (fstat(fd
, &st
) < 0)
2070 if (!S_ISDIR(st
.st_mode
))
2073 j
= journal_new(flags
, NULL
, NULL
);
2077 j
->toplevel_fd
= fd
;
2079 if (flags
& SD_JOURNAL_OS_ROOT
)
2080 r
= add_search_paths(j
);
2082 r
= add_root_directory(j
, NULL
, false);
2090 _public_
int sd_journal_open_files_fd(sd_journal
**ret
, int fds
[], unsigned n_fds
, int flags
) {
2092 _cleanup_(sd_journal_closep
) sd_journal
*j
= NULL
;
2096 assert_return(ret
, -EINVAL
);
2097 assert_return(n_fds
> 0, -EBADF
);
2098 assert_return(flags
== 0, -EINVAL
);
2100 j
= journal_new(flags
, NULL
, NULL
);
2104 for (i
= 0; i
< n_fds
; i
++) {
2112 if (fstat(fds
[i
], &st
) < 0) {
2117 r
= stat_verify_regular(&st
);
2121 r
= add_any_file(j
, fds
[i
], NULL
);
2126 j
->no_new_files
= true;
2127 j
->no_inotify
= true;
2133 /* If we fail, make sure we don't take possession of the files we managed to make use of successfully, and they
2135 ORDERED_HASHMAP_FOREACH(f
, j
->files
)
2136 f
->close_fd
= false;
2141 _public_
void sd_journal_close(sd_journal
*j
) {
2147 sd_journal_flush_matches(j
);
2149 ordered_hashmap_free_with_destructor(j
->files
, journal_file_close
);
2150 iterated_cache_free(j
->files_cache
);
2152 while ((d
= hashmap_first(j
->directories_by_path
)))
2153 remove_directory(j
, d
);
2155 while ((d
= hashmap_first(j
->directories_by_wd
)))
2156 remove_directory(j
, d
);
2158 hashmap_free(j
->directories_by_path
);
2159 hashmap_free(j
->directories_by_wd
);
2161 safe_close(j
->inotify_fd
);
2164 mmap_cache_stats_log_debug(j
->mmap
);
2165 mmap_cache_unref(j
->mmap
);
2168 hashmap_free_free(j
->errors
);
2173 free(j
->unique_field
);
2174 free(j
->fields_buffer
);
2178 _public_
int sd_journal_get_realtime_usec(sd_journal
*j
, uint64_t *ret
) {
2183 assert_return(j
, -EINVAL
);
2184 assert_return(!journal_pid_changed(j
), -ECHILD
);
2185 assert_return(ret
, -EINVAL
);
2187 f
= j
->current_file
;
2189 return -EADDRNOTAVAIL
;
2191 if (f
->current_offset
<= 0)
2192 return -EADDRNOTAVAIL
;
2194 r
= journal_file_move_to_object(f
, OBJECT_ENTRY
, f
->current_offset
, &o
);
2198 *ret
= le64toh(o
->entry
.realtime
);
2202 _public_
int sd_journal_get_monotonic_usec(sd_journal
*j
, uint64_t *ret
, sd_id128_t
*ret_boot_id
) {
2207 assert_return(j
, -EINVAL
);
2208 assert_return(!journal_pid_changed(j
), -ECHILD
);
2210 f
= j
->current_file
;
2212 return -EADDRNOTAVAIL
;
2214 if (f
->current_offset
<= 0)
2215 return -EADDRNOTAVAIL
;
2217 r
= journal_file_move_to_object(f
, OBJECT_ENTRY
, f
->current_offset
, &o
);
2222 *ret_boot_id
= o
->entry
.boot_id
;
2226 r
= sd_id128_get_boot(&id
);
2230 if (!sd_id128_equal(id
, o
->entry
.boot_id
))
2235 *ret
= le64toh(o
->entry
.monotonic
);
2240 static bool field_is_valid(const char *field
) {
2248 if (startswith(field
, "__"))
2251 for (p
= field
; *p
; p
++) {
2256 if (*p
>= 'A' && *p
<= 'Z')
2259 if (*p
>= '0' && *p
<= '9')
2268 _public_
int sd_journal_get_data(sd_journal
*j
, const char *field
, const void **data
, size_t *size
) {
2271 size_t field_length
;
2275 assert_return(j
, -EINVAL
);
2276 assert_return(!journal_pid_changed(j
), -ECHILD
);
2277 assert_return(field
, -EINVAL
);
2278 assert_return(data
, -EINVAL
);
2279 assert_return(size
, -EINVAL
);
2280 assert_return(field_is_valid(field
), -EINVAL
);
2282 f
= j
->current_file
;
2284 return -EADDRNOTAVAIL
;
2286 if (f
->current_offset
<= 0)
2287 return -EADDRNOTAVAIL
;
2289 r
= journal_file_move_to_object(f
, OBJECT_ENTRY
, f
->current_offset
, &o
);
2293 field_length
= strlen(field
);
2295 n
= journal_file_entry_n_items(o
);
2296 for (i
= 0; i
< n
; i
++) {
2303 p
= le64toh(o
->entry
.items
[i
].object_offset
);
2304 le_hash
= o
->entry
.items
[i
].hash
;
2305 r
= journal_file_move_to_object(f
, OBJECT_DATA
, p
, &d
);
2306 if (IN_SET(r
, -EADDRNOTAVAIL
, -EBADMSG
)) {
2307 log_debug_errno(r
, "Entry item %"PRIu64
" data object is bad, skipping over it: %m", i
);
2313 if (le_hash
!= d
->data
.hash
) {
2314 log_debug("Entry item %"PRIu64
" hash is bad, skipping over it.", i
);
2318 l
= le64toh(d
->object
.size
) - offsetof(Object
, data
.payload
);
2320 compression
= d
->object
.flags
& OBJECT_COMPRESSION_MASK
;
2322 #if HAVE_COMPRESSION
2323 r
= decompress_startswith(compression
,
2325 &f
->compress_buffer
,
2326 field
, field_length
, '=');
2328 log_debug_errno(r
, "Cannot decompress %s object of length %"PRIu64
" at offset "OFSfmt
": %m",
2329 object_compressed_to_string(compression
), l
, p
);
2334 r
= decompress_blob(compression
,
2336 &f
->compress_buffer
, &rsize
,
2341 *data
= f
->compress_buffer
;
2342 *size
= (size_t) rsize
;
2347 return -EPROTONOSUPPORT
;
2349 } else if (l
>= field_length
+1 &&
2350 memcmp(d
->data
.payload
, field
, field_length
) == 0 &&
2351 d
->data
.payload
[field_length
] == '=') {
2355 if ((uint64_t) t
!= l
)
2358 *data
= d
->data
.payload
;
2368 static int return_data(
2372 const void **ret_data
,
2382 l
= le64toh(READ_NOW(o
->object
.size
));
2383 if (l
< offsetof(Object
, data
.payload
))
2385 l
-= offsetof(Object
, data
.payload
);
2387 /* We can't read objects larger than 4G on a 32bit machine */
2389 if ((uint64_t) t
!= l
)
2392 compression
= o
->object
.flags
& OBJECT_COMPRESSION_MASK
;
2394 #if HAVE_COMPRESSION
2398 r
= decompress_blob(
2401 &f
->compress_buffer
, &rsize
,
2407 *ret_data
= f
->compress_buffer
;
2409 *ret_size
= (size_t) rsize
;
2411 return -EPROTONOSUPPORT
;
2415 *ret_data
= o
->data
.payload
;
2423 _public_
int sd_journal_enumerate_data(sd_journal
*j
, const void **data
, size_t *size
) {
2428 assert_return(j
, -EINVAL
);
2429 assert_return(!journal_pid_changed(j
), -ECHILD
);
2430 assert_return(data
, -EINVAL
);
2431 assert_return(size
, -EINVAL
);
2433 f
= j
->current_file
;
2435 return -EADDRNOTAVAIL
;
2437 if (f
->current_offset
<= 0)
2438 return -EADDRNOTAVAIL
;
2440 r
= journal_file_move_to_object(f
, OBJECT_ENTRY
, f
->current_offset
, &o
);
2444 for (uint64_t n
= journal_file_entry_n_items(o
); j
->current_field
< n
; j
->current_field
++) {
2448 p
= le64toh(o
->entry
.items
[j
->current_field
].object_offset
);
2449 le_hash
= o
->entry
.items
[j
->current_field
].hash
;
2450 r
= journal_file_move_to_object(f
, OBJECT_DATA
, p
, &o
);
2451 if (IN_SET(r
, -EADDRNOTAVAIL
, -EBADMSG
)) {
2452 log_debug_errno(r
, "Entry item %"PRIu64
" data object is bad, skipping over it: %m", j
->current_field
);
2458 if (le_hash
!= o
->data
.hash
) {
2459 log_debug("Entry item %"PRIu64
" hash is bad, skipping over it.", j
->current_field
);
2463 r
= return_data(j
, f
, o
, data
, size
);
2464 if (r
== -EBADMSG
) {
2465 log_debug("Entry item %"PRIu64
" data payload is bad, skipping over it.", j
->current_field
);
2479 _public_
int sd_journal_enumerate_available_data(sd_journal
*j
, const void **data
, size_t *size
) {
2483 r
= sd_journal_enumerate_data(j
, data
, size
);
2486 if (!JOURNAL_ERRNO_IS_UNAVAILABLE_FIELD(r
))
2488 j
->current_field
++; /* Try with the next field */
2492 _public_
void sd_journal_restart_data(sd_journal
*j
) {
2496 j
->current_field
= 0;
2499 static int reiterate_all_paths(sd_journal
*j
) {
2502 if (j
->no_new_files
)
2503 return add_current_paths(j
);
2505 if (j
->flags
& SD_JOURNAL_OS_ROOT
)
2506 return add_search_paths(j
);
2508 if (j
->toplevel_fd
>= 0)
2509 return add_root_directory(j
, NULL
, false);
2512 return add_root_directory(j
, j
->path
, true);
2514 return add_search_paths(j
);
2517 _public_
int sd_journal_get_fd(sd_journal
*j
) {
2520 assert_return(j
, -EINVAL
);
2521 assert_return(!journal_pid_changed(j
), -ECHILD
);
2524 return -EMEDIUMTYPE
;
2526 if (j
->inotify_fd
>= 0)
2527 return j
->inotify_fd
;
2529 r
= allocate_inotify(j
);
2533 log_debug("Reiterating files to get inotify watches established.");
2535 /* Iterate through all dirs again, to add them to the inotify */
2536 r
= reiterate_all_paths(j
);
2540 return j
->inotify_fd
;
2543 _public_
int sd_journal_get_events(sd_journal
*j
) {
2546 assert_return(j
, -EINVAL
);
2547 assert_return(!journal_pid_changed(j
), -ECHILD
);
2549 fd
= sd_journal_get_fd(j
);
2556 _public_
int sd_journal_get_timeout(sd_journal
*j
, uint64_t *timeout_usec
) {
2559 assert_return(j
, -EINVAL
);
2560 assert_return(!journal_pid_changed(j
), -ECHILD
);
2561 assert_return(timeout_usec
, -EINVAL
);
2563 fd
= sd_journal_get_fd(j
);
2567 if (!j
->on_network
) {
2568 *timeout_usec
= UINT64_MAX
;
2572 /* If we are on the network we need to regularly check for
2573 * changes manually */
2575 *timeout_usec
= j
->last_process_usec
+ JOURNAL_FILES_RECHECK_USEC
;
2579 static void process_q_overflow(sd_journal
*j
) {
2585 /* When the inotify queue overruns we need to enumerate and re-validate all journal files to bring our list
2586 * back in sync with what's on disk. For this we pick a new generation counter value. It'll be assigned to all
2587 * journal files we encounter. All journal files and all directories that don't carry it after reenumeration
2588 * are subject for unloading. */
2590 log_debug("Inotify queue overrun, reiterating everything.");
2593 (void) reiterate_all_paths(j
);
2595 ORDERED_HASHMAP_FOREACH(f
, j
->files
) {
2597 if (f
->last_seen_generation
== j
->generation
)
2600 log_debug("File '%s' hasn't been seen in this enumeration, removing.", f
->path
);
2601 remove_file_real(j
, f
);
2604 HASHMAP_FOREACH(m
, j
->directories_by_path
) {
2606 if (m
->last_seen_generation
== j
->generation
)
2609 if (m
->is_root
) /* Never GC root directories */
2612 log_debug("Directory '%s' hasn't been seen in this enumeration, removing.", f
->path
);
2613 remove_directory(j
, m
);
2616 log_debug("Reiteration complete.");
2619 static void process_inotify_event(sd_journal
*j
, const struct inotify_event
*e
) {
2625 if (e
->mask
& IN_Q_OVERFLOW
) {
2626 process_q_overflow(j
);
2630 /* Is this a subdirectory we watch? */
2631 d
= hashmap_get(j
->directories_by_wd
, INT_TO_PTR(e
->wd
));
2633 if (!(e
->mask
& IN_ISDIR
) && e
->len
> 0 &&
2634 (endswith(e
->name
, ".journal") ||
2635 endswith(e
->name
, ".journal~"))) {
2637 /* Event for a journal file */
2639 if (e
->mask
& (IN_CREATE
|IN_MOVED_TO
|IN_MODIFY
|IN_ATTRIB
))
2640 (void) add_file_by_name(j
, d
->path
, e
->name
);
2641 else if (e
->mask
& (IN_DELETE
|IN_MOVED_FROM
|IN_UNMOUNT
))
2642 remove_file_by_name(j
, d
->path
, e
->name
);
2644 } else if (!d
->is_root
&& e
->len
== 0) {
2646 /* Event for a subdirectory */
2648 if (e
->mask
& (IN_DELETE_SELF
|IN_MOVE_SELF
|IN_UNMOUNT
))
2649 remove_directory(j
, d
);
2651 } else if (d
->is_root
&& (e
->mask
& IN_ISDIR
) && e
->len
> 0 && id128_is_valid(e
->name
)) {
2653 /* Event for root directory */
2655 if (e
->mask
& (IN_CREATE
|IN_MOVED_TO
|IN_MODIFY
|IN_ATTRIB
))
2656 (void) add_directory(j
, d
->path
, e
->name
);
2662 if (e
->mask
& IN_IGNORED
)
2665 log_debug("Unexpected inotify event.");
2668 static int determine_change(sd_journal
*j
) {
2673 b
= j
->current_invalidate_counter
!= j
->last_invalidate_counter
;
2674 j
->last_invalidate_counter
= j
->current_invalidate_counter
;
2676 return b
? SD_JOURNAL_INVALIDATE
: SD_JOURNAL_APPEND
;
2679 _public_
int sd_journal_process(sd_journal
*j
) {
2680 bool got_something
= false;
2682 assert_return(j
, -EINVAL
);
2683 assert_return(!journal_pid_changed(j
), -ECHILD
);
2685 if (j
->inotify_fd
< 0) /* We have no inotify fd yet? Then there's noting to process. */
2688 j
->last_process_usec
= now(CLOCK_MONOTONIC
);
2689 j
->last_invalidate_counter
= j
->current_invalidate_counter
;
2692 union inotify_event_buffer buffer
;
2693 struct inotify_event
*e
;
2696 l
= read(j
->inotify_fd
, &buffer
, sizeof(buffer
));
2698 if (ERRNO_IS_TRANSIENT(errno
))
2699 return got_something
? determine_change(j
) : SD_JOURNAL_NOP
;
2704 got_something
= true;
2706 FOREACH_INOTIFY_EVENT(e
, buffer
, l
)
2707 process_inotify_event(j
, e
);
2711 _public_
int sd_journal_wait(sd_journal
*j
, uint64_t timeout_usec
) {
2715 assert_return(j
, -EINVAL
);
2716 assert_return(!journal_pid_changed(j
), -ECHILD
);
2718 if (j
->inotify_fd
< 0) {
2721 /* This is the first invocation, hence create the
2723 r
= sd_journal_get_fd(j
);
2727 /* Server might have done some vacuuming while we weren't watching.
2728 Get rid of the deleted files now so they don't stay around indefinitely. */
2729 ORDERED_HASHMAP_FOREACH(f
, j
->files
) {
2730 r
= journal_file_fstat(f
);
2732 remove_file_real(j
, f
);
2734 log_debug_errno(r
,"Failed to fstat() journal file '%s' : %m", f
->path
);
2739 /* The journal might have changed since the context
2740 * object was created and we weren't watching before,
2741 * hence don't wait for anything, and return
2743 return determine_change(j
);
2746 r
= sd_journal_get_timeout(j
, &t
);
2750 if (t
!= UINT64_MAX
) {
2751 t
= usec_sub_unsigned(t
, now(CLOCK_MONOTONIC
));
2753 if (timeout_usec
== UINT64_MAX
|| timeout_usec
> t
)
2758 r
= fd_wait_for_event(j
->inotify_fd
, POLLIN
, timeout_usec
);
2759 } while (r
== -EINTR
);
2764 return sd_journal_process(j
);
2767 _public_
int sd_journal_get_cutoff_realtime_usec(sd_journal
*j
, uint64_t *from
, uint64_t *to
) {
2770 uint64_t fmin
= 0, tmax
= 0;
2773 assert_return(j
, -EINVAL
);
2774 assert_return(!journal_pid_changed(j
), -ECHILD
);
2775 assert_return(from
|| to
, -EINVAL
);
2776 assert_return(from
!= to
, -EINVAL
);
2778 ORDERED_HASHMAP_FOREACH(f
, j
->files
) {
2781 r
= journal_file_get_cutoff_realtime_usec(f
, &fr
, &t
);
2794 fmin
= MIN(fr
, fmin
);
2795 tmax
= MAX(t
, tmax
);
2804 return first
? 0 : 1;
2807 _public_
int sd_journal_get_cutoff_monotonic_usec(
2813 uint64_t from
= UINT64_MAX
, to
= UINT64_MAX
;
2818 assert_return(j
, -EINVAL
);
2819 assert_return(!journal_pid_changed(j
), -ECHILD
);
2820 assert_return(ret_from
!= ret_to
, -EINVAL
);
2822 ORDERED_HASHMAP_FOREACH(f
, j
->files
) {
2825 r
= journal_file_get_cutoff_monotonic_usec(f
, boot_id
, &ff
, &tt
);
2834 from
= MIN(ff
, from
);
2851 void journal_print_header(sd_journal
*j
) {
2853 bool newline
= false;
2857 ORDERED_HASHMAP_FOREACH(f
, j
->files
) {
2863 journal_file_print_header(f
);
2867 _public_
int sd_journal_get_usage(sd_journal
*j
, uint64_t *ret
) {
2871 assert_return(j
, -EINVAL
);
2872 assert_return(!journal_pid_changed(j
), -ECHILD
);
2873 assert_return(ret
, -EINVAL
);
2875 ORDERED_HASHMAP_FOREACH(f
, j
->files
) {
2879 if (fstat(f
->fd
, &st
) < 0)
2882 b
= (uint64_t) st
.st_blocks
;
2883 if (b
> UINT64_MAX
/ 512)
2887 if (sum
> UINT64_MAX
- b
)
2896 _public_
int sd_journal_query_unique(sd_journal
*j
, const char *field
) {
2899 assert_return(j
, -EINVAL
);
2900 assert_return(!journal_pid_changed(j
), -ECHILD
);
2901 assert_return(!isempty(field
), -EINVAL
);
2902 assert_return(field_is_valid(field
), -EINVAL
);
2904 r
= free_and_strdup(&j
->unique_field
, field
);
2908 j
->unique_file
= NULL
;
2909 j
->unique_offset
= 0;
2910 j
->unique_file_lost
= false;
2915 _public_
int sd_journal_enumerate_unique(
2917 const void **ret_data
,
2922 assert_return(j
, -EINVAL
);
2923 assert_return(!journal_pid_changed(j
), -ECHILD
);
2924 assert_return(j
->unique_field
, -EINVAL
);
2926 k
= strlen(j
->unique_field
);
2928 if (!j
->unique_file
) {
2929 if (j
->unique_file_lost
)
2932 j
->unique_file
= ordered_hashmap_first(j
->files
);
2933 if (!j
->unique_file
)
2936 j
->unique_offset
= 0;
2947 /* Proceed to next data object in the field's linked list */
2948 if (j
->unique_offset
== 0) {
2949 r
= journal_file_find_field_object(j
->unique_file
, j
->unique_field
, k
, &o
, NULL
);
2953 j
->unique_offset
= r
> 0 ? le64toh(o
->field
.head_data_offset
) : 0;
2955 r
= journal_file_move_to_object(j
->unique_file
, OBJECT_DATA
, j
->unique_offset
, &o
);
2959 j
->unique_offset
= le64toh(o
->data
.next_field_offset
);
2962 /* We reached the end of the list? Then start again, with the next file */
2963 if (j
->unique_offset
== 0) {
2964 j
->unique_file
= ordered_hashmap_next(j
->files
, j
->unique_file
->path
);
2965 if (!j
->unique_file
)
2971 /* We do not use OBJECT_DATA context here, but OBJECT_UNUSED
2972 * instead, so that we can look at this data object at the same
2973 * time as one on another file */
2974 r
= journal_file_move_to_object(j
->unique_file
, OBJECT_UNUSED
, j
->unique_offset
, &o
);
2978 /* Let's do the type check by hand, since we used 0 context above. */
2979 if (o
->object
.type
!= OBJECT_DATA
)
2980 return log_debug_errno(SYNTHETIC_ERRNO(EBADMSG
),
2981 "%s:offset " OFSfmt
": object has type %d, expected %d",
2982 j
->unique_file
->path
,
2984 o
->object
.type
, OBJECT_DATA
);
2986 r
= return_data(j
, j
->unique_file
, o
, &odata
, &ol
);
2990 /* Check if we have at least the field name and "=". */
2992 return log_debug_errno(SYNTHETIC_ERRNO(EBADMSG
),
2993 "%s:offset " OFSfmt
": object has size %zu, expected at least %zu",
2994 j
->unique_file
->path
,
2995 j
->unique_offset
, ol
, k
+ 1);
2997 if (memcmp(odata
, j
->unique_field
, k
) != 0 || ((const char*) odata
)[k
] != '=')
2998 return log_debug_errno(SYNTHETIC_ERRNO(EBADMSG
),
2999 "%s:offset " OFSfmt
": object does not start with \"%s=\"",
3000 j
->unique_file
->path
,
3004 /* OK, now let's see if we already returned this data object by checking if it exists in the
3005 * earlier traversed files. */
3007 ORDERED_HASHMAP_FOREACH(of
, j
->files
) {
3008 if (of
== j
->unique_file
)
3011 /* Skip this file it didn't have any fields indexed */
3012 if (JOURNAL_HEADER_CONTAINS(of
->header
, n_fields
) && le64toh(of
->header
->n_fields
) <= 0)
3015 /* We can reuse the hash from our current file only on old-style journal files
3016 * without keyed hashes. On new-style files we have to calculate the hash anew, to
3017 * take the per-file hash seed into consideration. */
3018 if (!JOURNAL_HEADER_KEYED_HASH(j
->unique_file
->header
) && !JOURNAL_HEADER_KEYED_HASH(of
->header
))
3019 r
= journal_file_find_data_object_with_hash(of
, odata
, ol
, le64toh(o
->data
.hash
), NULL
, NULL
);
3021 r
= journal_file_find_data_object(of
, odata
, ol
, NULL
, NULL
);
3033 r
= return_data(j
, j
->unique_file
, o
, ret_data
, ret_size
);
3041 _public_
int sd_journal_enumerate_available_unique(sd_journal
*j
, const void **data
, size_t *size
) {
3045 r
= sd_journal_enumerate_unique(j
, data
, size
);
3048 if (!JOURNAL_ERRNO_IS_UNAVAILABLE_FIELD(r
))
3050 /* Try with the next field. sd_journal_enumerate_unique() modifies state, so on the next try
3051 * we will access the next field. */
3055 _public_
void sd_journal_restart_unique(sd_journal
*j
) {
3059 j
->unique_file
= NULL
;
3060 j
->unique_offset
= 0;
3061 j
->unique_file_lost
= false;
3064 _public_
int sd_journal_enumerate_fields(sd_journal
*j
, const char **field
) {
3067 assert_return(j
, -EINVAL
);
3068 assert_return(!journal_pid_changed(j
), -ECHILD
);
3069 assert_return(field
, -EINVAL
);
3071 if (!j
->fields_file
) {
3072 if (j
->fields_file_lost
)
3075 j
->fields_file
= ordered_hashmap_first(j
->files
);
3076 if (!j
->fields_file
)
3079 j
->fields_hash_table_index
= 0;
3080 j
->fields_offset
= 0;
3084 JournalFile
*f
, *of
;
3092 if (j
->fields_offset
== 0) {
3095 /* We are not yet positioned at any field. Let's pick the first one */
3096 r
= journal_file_map_field_hash_table(f
);
3100 m
= le64toh(f
->header
->field_hash_table_size
) / sizeof(HashItem
);
3102 if (j
->fields_hash_table_index
>= m
) {
3103 /* Reached the end of the hash table, go to the next file. */
3108 j
->fields_offset
= le64toh(f
->field_hash_table
[j
->fields_hash_table_index
].head_hash_offset
);
3110 if (j
->fields_offset
!= 0)
3113 /* Empty hash table bucket, go to next one */
3114 j
->fields_hash_table_index
++;
3118 /* Proceed with next file */
3119 j
->fields_file
= ordered_hashmap_next(j
->files
, f
->path
);
3120 if (!j
->fields_file
) {
3125 j
->fields_offset
= 0;
3126 j
->fields_hash_table_index
= 0;
3131 /* We are already positioned at a field. If so, let's figure out the next field from it */
3133 r
= journal_file_move_to_object(f
, OBJECT_FIELD
, j
->fields_offset
, &o
);
3137 j
->fields_offset
= le64toh(o
->field
.next_hash_offset
);
3138 if (j
->fields_offset
== 0) {
3139 /* Reached the end of the hash table chain */
3140 j
->fields_hash_table_index
++;
3145 /* We use OBJECT_UNUSED here, so that the iterator below doesn't remove our mmap window */
3146 r
= journal_file_move_to_object(f
, OBJECT_UNUSED
, j
->fields_offset
, &o
);
3150 /* Because we used OBJECT_UNUSED above, we need to do our type check manually */
3151 if (o
->object
.type
!= OBJECT_FIELD
)
3152 return log_debug_errno(SYNTHETIC_ERRNO(EBADMSG
),
3153 "%s:offset " OFSfmt
": object has type %i, expected %i",
3154 f
->path
, j
->fields_offset
,
3155 o
->object
.type
, OBJECT_FIELD
);
3157 sz
= le64toh(o
->object
.size
) - offsetof(Object
, field
.payload
);
3159 /* Let's see if we already returned this field name before. */
3161 ORDERED_HASHMAP_FOREACH(of
, j
->files
) {
3165 /* Skip this file it didn't have any fields indexed */
3166 if (JOURNAL_HEADER_CONTAINS(of
->header
, n_fields
) && le64toh(of
->header
->n_fields
) <= 0)
3169 if (!JOURNAL_HEADER_KEYED_HASH(f
->header
) && !JOURNAL_HEADER_KEYED_HASH(of
->header
))
3170 r
= journal_file_find_field_object_with_hash(of
, o
->field
.payload
, sz
,
3171 le64toh(o
->field
.hash
), NULL
, NULL
);
3173 r
= journal_file_find_field_object(of
, o
->field
.payload
, sz
, NULL
, NULL
);
3185 /* Check if this is really a valid string containing no NUL byte */
3186 if (memchr(o
->field
.payload
, 0, sz
))
3189 if (j
->data_threshold
> 0 && sz
> j
->data_threshold
)
3190 sz
= j
->data_threshold
;
3192 if (!GREEDY_REALLOC(j
->fields_buffer
, sz
+ 1))
3195 memcpy(j
->fields_buffer
, o
->field
.payload
, sz
);
3196 j
->fields_buffer
[sz
] = 0;
3198 if (!field_is_valid(j
->fields_buffer
))
3201 *field
= j
->fields_buffer
;
3206 _public_
void sd_journal_restart_fields(sd_journal
*j
) {
3210 j
->fields_file
= NULL
;
3211 j
->fields_hash_table_index
= 0;
3212 j
->fields_offset
= 0;
3213 j
->fields_file_lost
= false;
3216 _public_
int sd_journal_reliable_fd(sd_journal
*j
) {
3217 assert_return(j
, -EINVAL
);
3218 assert_return(!journal_pid_changed(j
), -ECHILD
);
3220 return !j
->on_network
;
3223 static char *lookup_field(const char *field
, void *userdata
) {
3224 sd_journal
*j
= userdata
;
3232 r
= sd_journal_get_data(j
, field
, &data
, &size
);
3234 size
> REPLACE_VAR_MAX
)
3235 return strdup(field
);
3237 d
= strlen(field
) + 1;
3239 return strndup((const char*) data
+ d
, size
- d
);
3242 _public_
int sd_journal_get_catalog(sd_journal
*j
, char **ret
) {
3246 _cleanup_free_
char *text
= NULL
, *cid
= NULL
;
3250 assert_return(j
, -EINVAL
);
3251 assert_return(!journal_pid_changed(j
), -ECHILD
);
3252 assert_return(ret
, -EINVAL
);
3254 r
= sd_journal_get_data(j
, "MESSAGE_ID", &data
, &size
);
3258 cid
= strndup((const char*) data
+ 11, size
- 11);
3262 r
= sd_id128_from_string(cid
, &id
);
3266 r
= catalog_get(CATALOG_DATABASE
, id
, &text
);
3270 t
= replace_var(text
, lookup_field
, j
);
3278 _public_
int sd_journal_get_catalog_for_message_id(sd_id128_t id
, char **ret
) {
3279 assert_return(ret
, -EINVAL
);
3281 return catalog_get(CATALOG_DATABASE
, id
, ret
);
3284 _public_
int sd_journal_set_data_threshold(sd_journal
*j
, size_t sz
) {
3285 assert_return(j
, -EINVAL
);
3286 assert_return(!journal_pid_changed(j
), -ECHILD
);
3288 j
->data_threshold
= sz
;
3292 _public_
int sd_journal_get_data_threshold(sd_journal
*j
, size_t *sz
) {
3293 assert_return(j
, -EINVAL
);
3294 assert_return(!journal_pid_changed(j
), -ECHILD
);
3295 assert_return(sz
, -EINVAL
);
3297 *sz
= j
->data_threshold
;
3301 _public_
int sd_journal_has_runtime_files(sd_journal
*j
) {
3302 assert_return(j
, -EINVAL
);
3304 return j
->has_runtime_files
;
3307 _public_
int sd_journal_has_persistent_files(sd_journal
*j
) {
3308 assert_return(j
, -EINVAL
);
3310 return j
->has_persistent_files
;