log_info("got IO on %c", PTR_TO_INT(userdata));
if (userdata == INT_TO_PTR('a')) {
- assert_se(sd_event_source_set_enabled(s, SD_EVENT_OFF) >= 0);
- assert_se(!got_a);
+ ASSERT_OK(sd_event_source_set_enabled(s, SD_EVENT_OFF));
+ ASSERT_FALSE(got_a);
got_a = true;
} else if (userdata == INT_TO_PTR('b')) {
- assert_se(!got_b);
+ ASSERT_FALSE(got_b);
got_b = true;
} else if (userdata == INT_TO_PTR('d')) {
got_d++;
if (got_d < 2)
- assert_se(sd_event_source_set_enabled(s, SD_EVENT_ONESHOT) >= 0);
+ ASSERT_OK(sd_event_source_set_enabled(s, SD_EVENT_ONESHOT));
else
- assert_se(sd_event_source_set_enabled(s, SD_EVENT_OFF) >= 0);
+ ASSERT_OK(sd_event_source_set_enabled(s, SD_EVENT_OFF));
} else
assert_not_reached();
static int child_handler(sd_event_source *s, const siginfo_t *si, void *userdata) {
- assert_se(s);
- assert_se(si);
+ ASSERT_NOT_NULL(s);
+ ASSERT_NOT_NULL(si);
- assert_se(si->si_uid == getuid());
- assert_se(si->si_signo == SIGCHLD);
- assert_se(si->si_code == CLD_EXITED);
- assert_se(si->si_status == 78);
+ ASSERT_EQ(si->si_uid, getuid());
+ ASSERT_EQ(si->si_signo, SIGCHLD);
+ ASSERT_EQ(si->si_code, CLD_EXITED);
+ ASSERT_EQ(si->si_status, 78);
log_info("got child on %c", PTR_TO_INT(userdata));
- assert_se(userdata == INT_TO_PTR('f'));
+ ASSERT_PTR_EQ(userdata, INT_TO_PTR('f'));
- assert_se(sd_event_exit(sd_event_source_get_event(s), 0) >= 0);
+ ASSERT_OK(sd_event_exit(sd_event_source_get_event(s), 0));
sd_event_source_unref(s);
return 1;
pid_t pid;
siginfo_t plain_si;
- assert_se(s);
- assert_se(si);
+ ASSERT_NOT_NULL(s);
+ ASSERT_NOT_NULL(si);
log_info("got signal on %c", PTR_TO_INT(userdata));
- assert_se(userdata == INT_TO_PTR('e'));
+ ASSERT_PTR_EQ(userdata, INT_TO_PTR('e'));
- assert_se(sigprocmask_many(SIG_BLOCK, NULL, SIGCHLD, SIGUSR2) >= 0);
+ ASSERT_OK(sigprocmask_many(SIG_BLOCK, NULL, SIGCHLD, SIGUSR2));
- pid = fork();
- assert_se(pid >= 0);
+ ASSERT_OK_ERRNO(pid = fork());
if (pid == 0) {
sigset_t ss;
- assert_se(sigemptyset(&ss) >= 0);
- assert_se(sigaddset(&ss, SIGUSR2) >= 0);
+ ASSERT_OK_ERRNO(sigemptyset(&ss));
+ ASSERT_OK_ERRNO(sigaddset(&ss, SIGUSR2));
zero(plain_si);
- assert_se(sigwaitinfo(&ss, &plain_si) >= 0);
+ ASSERT_OK_ERRNO(sigwaitinfo(&ss, &plain_si));
- assert_se(plain_si.si_signo == SIGUSR2);
- assert_se(plain_si.si_value.sival_int == 4711);
+ ASSERT_EQ(plain_si.si_signo, SIGUSR2);
+ ASSERT_EQ(plain_si.si_value.sival_int, 4711);
_exit(78);
}
- assert_se(sd_event_add_child(sd_event_source_get_event(s), &p, pid, WEXITED, child_handler, INT_TO_PTR('f')) >= 0);
- assert_se(sd_event_source_set_enabled(p, SD_EVENT_ONESHOT) >= 0);
- assert_se(sd_event_source_set_child_process_own(p, true) >= 0);
+ ASSERT_OK(sd_event_add_child(sd_event_source_get_event(s), &p, pid, WEXITED, child_handler, INT_TO_PTR('f')));
+ ASSERT_OK(sd_event_source_set_enabled(p, SD_EVENT_ONESHOT));
+ ASSERT_OK(sd_event_source_set_child_process_own(p, true));
/* We can't use structured initialization here, since the structure contains various unions and these
* fields lie in overlapping (carefully aligned) unions that LLVM is allergic to allow assignments
plain_si.si_uid = getuid();
plain_si.si_value.sival_int = 4711;
- assert_se(sd_event_source_send_child_signal(p, SIGUSR2, &plain_si, 0) >= 0);
+ ASSERT_OK(sd_event_source_send_child_signal(p, SIGUSR2, &plain_si, 0));
sd_event_source_unref(s);
static int defer_handler(sd_event_source *s, void *userdata) {
sd_event_source *p = NULL;
- assert_se(s);
+ ASSERT_NOT_NULL(s);
log_info("got defer on %c", PTR_TO_INT(userdata));
- assert_se(userdata == INT_TO_PTR('d'));
+ ASSERT_PTR_EQ(userdata, INT_TO_PTR('d'));
- assert_se(sigprocmask_many(SIG_BLOCK, NULL, SIGUSR1) >= 0);
+ ASSERT_OK(sigprocmask_many(SIG_BLOCK, NULL, SIGUSR1));
- assert_se(sd_event_add_signal(sd_event_source_get_event(s), &p, SIGUSR1, signal_handler, INT_TO_PTR('e')) >= 0);
- assert_se(sd_event_source_set_enabled(p, SD_EVENT_ONESHOT) >= 0);
+ ASSERT_OK(sd_event_add_signal(sd_event_source_get_event(s), &p, SIGUSR1, signal_handler, INT_TO_PTR('e')));
+ ASSERT_OK(sd_event_source_set_enabled(p, SD_EVENT_ONESHOT));
raise(SIGUSR1);
sd_event_source_unref(s);
if (do_quit) {
sd_event_source *p;
- assert_se(sd_event_add_defer(sd_event_source_get_event(s), &p, defer_handler, INT_TO_PTR('d')) >= 0);
- assert_se(sd_event_source_set_enabled(p, SD_EVENT_ONESHOT) >= 0);
+ ASSERT_OK(sd_event_add_defer(sd_event_source_get_event(s), &p, defer_handler, INT_TO_PTR('d')));
+ ASSERT_OK(sd_event_source_set_enabled(p, SD_EVENT_ONESHOT));
} else {
- assert_se(!got_c);
+ ASSERT_FALSE(got_c);
got_c = true;
}
} else
uint64_t event_now;
int64_t priority;
- assert_se(pipe(a) >= 0);
- assert_se(pipe(b) >= 0);
- assert_se(pipe(d) >= 0);
- assert_se(pipe(k) >= 0);
+ ASSERT_OK_ERRNO(pipe(a));
+ ASSERT_OK_ERRNO(pipe(b));
+ ASSERT_OK_ERRNO(pipe(d));
+ ASSERT_OK_ERRNO(pipe(k));
- assert_se(sd_event_default(&e) >= 0);
- assert_se(sd_event_now(e, CLOCK_MONOTONIC, &event_now) > 0);
+ ASSERT_OK(sd_event_default(&e));
+ ASSERT_OK_POSITIVE(sd_event_now(e, CLOCK_MONOTONIC, &event_now));
- assert_se(sd_event_set_watchdog(e, true) >= 0);
+ ASSERT_OK(sd_event_set_watchdog(e, true));
/* Test whether we cleanly can destroy an io event source from its own handler */
got_unref = false;
- assert_se(sd_event_add_io(e, &t, k[0], EPOLLIN, unref_handler, NULL) >= 0);
- assert_se(write(k[1], &ch, 1) == 1);
- assert_se(sd_event_run(e, UINT64_MAX) >= 1);
- assert_se(got_unref);
+ ASSERT_OK(sd_event_add_io(e, &t, k[0], EPOLLIN, unref_handler, NULL));
+ ASSERT_OK_EQ_ERRNO(write(k[1], &ch, 1), 1);
+ ASSERT_OK_POSITIVE(sd_event_run(e, UINT64_MAX));
+ ASSERT_TRUE(got_unref);
got_a = false, got_b = false, got_c = false, got_d = 0;
/* Add a oneshot handler, trigger it, reenable it, and trigger it again. */
- assert_se(sd_event_add_io(e, &w, d[0], EPOLLIN, io_handler, INT_TO_PTR('d')) >= 0);
- assert_se(sd_event_source_set_enabled(w, SD_EVENT_ONESHOT) >= 0);
- assert_se(write(d[1], &ch, 1) >= 0);
- assert_se(sd_event_run(e, UINT64_MAX) >= 1);
- assert_se(got_d == 1);
- assert_se(write(d[1], &ch, 1) >= 0);
- assert_se(sd_event_run(e, UINT64_MAX) >= 1);
- assert_se(got_d == 2);
-
- assert_se(sd_event_add_io(e, &x, a[0], EPOLLIN, io_handler, INT_TO_PTR('a')) >= 0);
- assert_se(sd_event_add_io(e, &y, b[0], EPOLLIN, io_handler, INT_TO_PTR('b')) >= 0);
+ ASSERT_OK(sd_event_add_io(e, &w, d[0], EPOLLIN, io_handler, INT_TO_PTR('d')));
+ ASSERT_OK(sd_event_source_set_enabled(w, SD_EVENT_ONESHOT));
+ ASSERT_OK_ERRNO(write(d[1], &ch, 1));
+ ASSERT_OK_POSITIVE(sd_event_run(e, UINT64_MAX));
+ ASSERT_EQ(got_d, 1U);
+ ASSERT_OK_ERRNO(write(d[1], &ch, 1));
+ ASSERT_OK_POSITIVE(sd_event_run(e, UINT64_MAX));
+ ASSERT_EQ(got_d, 2U);
+
+ ASSERT_OK(sd_event_add_io(e, &x, a[0], EPOLLIN, io_handler, INT_TO_PTR('a')));
+ ASSERT_OK(sd_event_add_io(e, &y, b[0], EPOLLIN, io_handler, INT_TO_PTR('b')));
do_quit = false;
- assert_se(sd_event_add_time(e, &z, CLOCK_MONOTONIC, 0, 0, time_handler, INT_TO_PTR('c')) >= 0);
- assert_se(sd_event_add_exit(e, &q, exit_handler, INT_TO_PTR('g')) >= 0);
-
- assert_se(sd_event_source_set_priority(x, 99) >= 0);
- assert_se(sd_event_source_get_priority(x, &priority) >= 0);
- assert_se(priority == 99);
- assert_se(sd_event_source_set_enabled(y, SD_EVENT_ONESHOT) >= 0);
- assert_se(sd_event_source_set_prepare(x, prepare_handler) >= 0);
- assert_se(sd_event_source_set_priority(z, 50) >= 0);
- assert_se(sd_event_source_set_enabled(z, SD_EVENT_ONESHOT) >= 0);
- assert_se(sd_event_source_set_prepare(z, prepare_handler) >= 0);
+ ASSERT_OK(sd_event_add_time(e, &z, CLOCK_MONOTONIC, 0, 0, time_handler, INT_TO_PTR('c')));
+ ASSERT_OK(sd_event_add_exit(e, &q, exit_handler, INT_TO_PTR('g')));
+
+ ASSERT_OK(sd_event_source_set_priority(x, 99));
+ ASSERT_OK(sd_event_source_get_priority(x, &priority));
+ ASSERT_EQ(priority, 99);
+ ASSERT_OK(sd_event_source_set_enabled(y, SD_EVENT_ONESHOT));
+ ASSERT_OK(sd_event_source_set_prepare(x, prepare_handler));
+ ASSERT_OK(sd_event_source_set_priority(z, 50));
+ ASSERT_OK(sd_event_source_set_enabled(z, SD_EVENT_ONESHOT));
+ ASSERT_OK(sd_event_source_set_prepare(z, prepare_handler));
/* Test for floating event sources */
- assert_se(sigprocmask_many(SIG_BLOCK, NULL, SIGRTMIN+1) >= 0);
- assert_se(sd_event_add_signal(e, NULL, SIGRTMIN+1, NULL, NULL) >= 0);
+ ASSERT_OK(sigprocmask_many(SIG_BLOCK, NULL, SIGRTMIN+1));
+ ASSERT_OK(sd_event_add_signal(e, NULL, SIGRTMIN+1, NULL, NULL));
- assert_se(write(a[1], &ch, 1) >= 0);
- assert_se(write(b[1], &ch, 1) >= 0);
+ ASSERT_OK_ERRNO(write(a[1], &ch, 1));
+ ASSERT_OK_ERRNO(write(b[1], &ch, 1));
- assert_se(!got_a && !got_b && !got_c);
+ ASSERT_FALSE(got_a);
+ ASSERT_FALSE(got_b);
+ ASSERT_FALSE(got_c);
- assert_se(sd_event_run(e, UINT64_MAX) >= 1);
+ ASSERT_OK_POSITIVE(sd_event_run(e, UINT64_MAX));
- assert_se(!got_a && got_b && !got_c);
+ ASSERT_FALSE(got_a);
+ ASSERT_TRUE(got_b);
+ ASSERT_FALSE(got_c);
- assert_se(sd_event_run(e, UINT64_MAX) >= 1);
+ ASSERT_OK_POSITIVE(sd_event_run(e, UINT64_MAX));
- assert_se(!got_a && got_b && got_c);
+ ASSERT_FALSE(got_a);
+ ASSERT_TRUE(got_b);
+ ASSERT_TRUE(got_c);
- assert_se(sd_event_run(e, UINT64_MAX) >= 1);
+ ASSERT_OK_POSITIVE(sd_event_run(e, UINT64_MAX));
- assert_se(got_a && got_b && got_c);
+ ASSERT_TRUE(got_a);
+ ASSERT_TRUE(got_b);
+ ASSERT_TRUE(got_c);
sd_event_source_unref(x);
sd_event_source_unref(y);
do_quit = true;
- assert_se(sd_event_add_post(e, NULL, post_handler, NULL) >= 0);
- assert_se(sd_event_now(e, CLOCK_MONOTONIC, &event_now) == 0);
- assert_se(sd_event_source_set_time(z, event_now + 200 * USEC_PER_MSEC) >= 0);
- assert_se(sd_event_source_set_enabled(z, SD_EVENT_ONESHOT) >= 0);
+ ASSERT_OK(sd_event_add_post(e, NULL, post_handler, NULL));
+ ASSERT_OK_ZERO(sd_event_now(e, CLOCK_MONOTONIC, &event_now));
+ ASSERT_OK(sd_event_source_set_time(z, event_now + 200 * USEC_PER_MSEC));
+ ASSERT_OK(sd_event_source_set_enabled(z, SD_EVENT_ONESHOT));
- assert_se(sd_event_loop(e) >= 0);
- assert_se(got_post);
- assert_se(got_exit);
+ ASSERT_OK(sd_event_loop(e));
+ ASSERT_TRUE(got_post);
+ ASSERT_TRUE(got_exit);
sd_event_source_unref(z);
sd_event_source_unref(q);
_cleanup_(sd_event_unrefp) sd_event *e = NULL;
uint64_t event_now;
- assert_se(sd_event_new(&e) >= 0);
- assert_se(sd_event_now(e, CLOCK_MONOTONIC, &event_now) > 0);
- assert_se(sd_event_now(e, CLOCK_REALTIME, &event_now) > 0);
- assert_se(sd_event_now(e, CLOCK_REALTIME_ALARM, &event_now) > 0);
- assert_se(sd_event_now(e, CLOCK_BOOTTIME, &event_now) > 0);
- assert_se(sd_event_now(e, CLOCK_BOOTTIME_ALARM, &event_now) > 0);
- assert_se(sd_event_now(e, -1, &event_now) == -EOPNOTSUPP);
- assert_se(sd_event_now(e, 900 /* arbitrary big number */, &event_now) == -EOPNOTSUPP);
+ ASSERT_OK(sd_event_new(&e));
+ ASSERT_OK_POSITIVE(sd_event_now(e, CLOCK_MONOTONIC, &event_now));
+ ASSERT_OK_POSITIVE(sd_event_now(e, CLOCK_REALTIME, &event_now));
+ ASSERT_OK_POSITIVE(sd_event_now(e, CLOCK_REALTIME_ALARM, &event_now));
+ ASSERT_OK_POSITIVE(sd_event_now(e, CLOCK_BOOTTIME, &event_now));
+ ASSERT_OK_POSITIVE(sd_event_now(e, CLOCK_BOOTTIME_ALARM, &event_now));
+ ASSERT_ERROR(sd_event_now(e, -1, &event_now), EOPNOTSUPP);
+ ASSERT_ERROR(sd_event_now(e, 900 /* arbitrary big number */, &event_now), EOPNOTSUPP);
- assert_se(sd_event_run(e, 0) == 0);
+ ASSERT_OK_ZERO(sd_event_run(e, 0));
- assert_se(sd_event_now(e, CLOCK_MONOTONIC, &event_now) == 0);
- assert_se(sd_event_now(e, CLOCK_REALTIME, &event_now) == 0);
- assert_se(sd_event_now(e, CLOCK_REALTIME_ALARM, &event_now) == 0);
- assert_se(sd_event_now(e, CLOCK_BOOTTIME, &event_now) == 0);
- assert_se(sd_event_now(e, CLOCK_BOOTTIME_ALARM, &event_now) == 0);
- assert_se(sd_event_now(e, -1, &event_now) == -EOPNOTSUPP);
- assert_se(sd_event_now(e, 900 /* arbitrary big number */, &event_now) == -EOPNOTSUPP);
+ ASSERT_OK_ZERO(sd_event_now(e, CLOCK_MONOTONIC, &event_now));
+ ASSERT_OK_ZERO(sd_event_now(e, CLOCK_REALTIME, &event_now));
+ ASSERT_OK_ZERO(sd_event_now(e, CLOCK_REALTIME_ALARM, &event_now));
+ ASSERT_OK_ZERO(sd_event_now(e, CLOCK_BOOTTIME, &event_now));
+ ASSERT_OK_ZERO(sd_event_now(e, CLOCK_BOOTTIME_ALARM, &event_now));
+ ASSERT_ERROR(sd_event_now(e, -1, &event_now), EOPNOTSUPP);
+ ASSERT_ERROR(sd_event_now(e, 900 /* arbitrary big number */, &event_now), EOPNOTSUPP);
}
static int last_rtqueue_sigval = 0;
sd_event_source *u = NULL, *v = NULL, *s = NULL;
sd_event *e = NULL;
- assert_se(sd_event_default(&e) >= 0);
+ ASSERT_OK(sd_event_default(&e));
- assert_se(sigprocmask_many(SIG_BLOCK, NULL, SIGRTMIN+2, SIGRTMIN+3, SIGUSR2) >= 0);
- assert_se(sd_event_add_signal(e, &u, SIGRTMIN+2, rtqueue_handler, NULL) >= 0);
- assert_se(sd_event_add_signal(e, &v, SIGRTMIN+3, rtqueue_handler, NULL) >= 0);
- assert_se(sd_event_add_signal(e, &s, SIGUSR2, rtqueue_handler, NULL) >= 0);
+ ASSERT_OK(sigprocmask_many(SIG_BLOCK, NULL, SIGRTMIN+2, SIGRTMIN+3, SIGUSR2));
+ ASSERT_OK(sd_event_add_signal(e, &u, SIGRTMIN+2, rtqueue_handler, NULL));
+ ASSERT_OK(sd_event_add_signal(e, &v, SIGRTMIN+3, rtqueue_handler, NULL));
+ ASSERT_OK(sd_event_add_signal(e, &s, SIGUSR2, rtqueue_handler, NULL));
- assert_se(sd_event_source_set_priority(v, -10) >= 0);
+ ASSERT_OK(sd_event_source_set_priority(v, -10));
- assert_se(sigqueue(getpid_cached(), SIGRTMIN+2, (union sigval) { .sival_int = 1 }) >= 0);
- assert_se(sigqueue(getpid_cached(), SIGRTMIN+3, (union sigval) { .sival_int = 2 }) >= 0);
- assert_se(sigqueue(getpid_cached(), SIGUSR2, (union sigval) { .sival_int = 3 }) >= 0);
- assert_se(sigqueue(getpid_cached(), SIGRTMIN+3, (union sigval) { .sival_int = 4 }) >= 0);
- assert_se(sigqueue(getpid_cached(), SIGUSR2, (union sigval) { .sival_int = 5 }) >= 0);
+ ASSERT_OK_ERRNO(sigqueue(getpid_cached(), SIGRTMIN+2, (union sigval) { .sival_int = 1 }));
+ ASSERT_OK_ERRNO(sigqueue(getpid_cached(), SIGRTMIN+3, (union sigval) { .sival_int = 2 }));
+ ASSERT_OK_ERRNO(sigqueue(getpid_cached(), SIGUSR2, (union sigval) { .sival_int = 3 }));
+ ASSERT_OK_ERRNO(sigqueue(getpid_cached(), SIGRTMIN+3, (union sigval) { .sival_int = 4 }));
+ ASSERT_OK_ERRNO(sigqueue(getpid_cached(), SIGUSR2, (union sigval) { .sival_int = 5 }));
- assert_se(n_rtqueue == 0);
- assert_se(last_rtqueue_sigval == 0);
+ ASSERT_EQ(n_rtqueue, 0);
+ ASSERT_EQ(last_rtqueue_sigval, 0);
- assert_se(sd_event_run(e, UINT64_MAX) >= 1);
- assert_se(n_rtqueue == 1);
- assert_se(last_rtqueue_sigval == 2); /* first SIGRTMIN+3 */
+ ASSERT_OK_EQ(sd_event_run(e, UINT64_MAX), 1);
+ ASSERT_EQ(n_rtqueue, 1);
+ ASSERT_EQ(last_rtqueue_sigval, 2); /* first SIGRTMIN+3 */
- assert_se(sd_event_run(e, UINT64_MAX) >= 1);
- assert_se(n_rtqueue == 2);
- assert_se(last_rtqueue_sigval == 4); /* second SIGRTMIN+3 */
+ ASSERT_OK_EQ(sd_event_run(e, UINT64_MAX), 1);
+ ASSERT_EQ(n_rtqueue, 2);
+ ASSERT_EQ(last_rtqueue_sigval, 4); /* second SIGRTMIN+3 */
- assert_se(sd_event_run(e, UINT64_MAX) >= 1);
- assert_se(n_rtqueue == 3);
- assert_se(last_rtqueue_sigval == 3); /* first SIGUSR2 */
+ ASSERT_OK_EQ(sd_event_run(e, UINT64_MAX), 1);
+ ASSERT_EQ(n_rtqueue, 3);
+ ASSERT_EQ(last_rtqueue_sigval, 3); /* first SIGUSR2 */
- assert_se(sd_event_run(e, UINT64_MAX) >= 1);
- assert_se(n_rtqueue == 4);
- assert_se(last_rtqueue_sigval == 1); /* SIGRTMIN+2 */
+ ASSERT_OK_EQ(sd_event_run(e, UINT64_MAX), 1);
+ ASSERT_EQ(n_rtqueue, 4);
+ ASSERT_EQ(last_rtqueue_sigval, 1); /* SIGRTMIN+2 */
- assert_se(sd_event_run(e, 0) == 0); /* the other SIGUSR2 is dropped, because the first one was still queued */
- assert_se(n_rtqueue == 4);
- assert_se(last_rtqueue_sigval == 1);
+ ASSERT_OK_ZERO(sd_event_run(e, 0)); /* the other SIGUSR2 is dropped, because the first one was still queued */
+ ASSERT_EQ(n_rtqueue, 4);
+ ASSERT_EQ(last_rtqueue_sigval, 1);
sd_event_source_unref(u);
sd_event_source_unref(v);
static void maybe_exit(sd_event_source *s, struct inotify_context *c) {
unsigned n;
- assert_se(s);
- assert_se(c);
+ ASSERT_NOT_NULL(s);
+ ASSERT_NOT_NULL(c);
if (!c->delete_self_handler_called)
return;
const char *path, *description;
unsigned bit, n;
- assert_se(sd_event_source_get_inotify_path(s, &path) >= 0);
+ ASSERT_OK(sd_event_source_get_inotify_path(s, &path));
- assert_se(sd_event_source_get_description(s, &description) >= 0);
- assert_se(safe_atou(description, &n) >= 0);
+ ASSERT_OK(sd_event_source_get_description(s, &description));
+ ASSERT_OK(safe_atou(description, &n));
- assert_se(n <= 3);
+ ASSERT_LE(n, 3U);
bit = 1U << n;
if (ev->mask & IN_Q_OVERFLOW) {
log_info("inotify-handler for %s <%s>: overflow", path, description);
c->create_overflow |= bit;
} else if (ev->mask & IN_CREATE) {
- assert_se(path_equal_or_inode_same(path, c->path, 0));
+ ASSERT_TRUE(path_equal_or_inode_same(path, c->path, 0));
if (streq(ev->name, "sub"))
log_debug("inotify-handler for %s <%s>: create on %s", path, description, ev->name);
else {
unsigned i;
- assert_se(safe_atou(ev->name, &i) >= 0);
- assert_se(i < c->n_create_events);
+ ASSERT_OK(safe_atou(ev->name, &i));
+ ASSERT_LT(i, c->n_create_events);
c->create_called[i] |= bit;
}
} else if (ev->mask & IN_DELETE) {
log_info("inotify-handler for %s <%s>: delete of %s", path, description, ev->name);
- assert_se(streq(ev->name, "sub"));
+ ASSERT_STREQ(ev->name, "sub");
} else
assert_not_reached();
struct inotify_context *c = ASSERT_PTR(userdata);
const char *path;
- assert_se(sd_event_source_get_inotify_path(s, &path) >= 0);
+ ASSERT_OK(sd_event_source_get_inotify_path(s, &path));
if (ev->mask & IN_Q_OVERFLOW) {
log_info("delete-self-handler for %s: overflow", path);
log_info("/* %s(%u) */", __func__, n_create_events);
- assert_se(sd_event_default(&e) >= 0);
+ ASSERT_OK(sd_event_default(&e));
- assert_se(mkdtemp_malloc("/tmp/test-inotify-XXXXXX", &p) >= 0);
+ ASSERT_OK(mkdtemp_malloc("/tmp/test-inotify-XXXXXX", &p));
context.path = p;
- assert_se(sd_event_add_inotify(e, &a, p, IN_CREATE|IN_ONLYDIR, inotify_handler, &context) >= 0);
- assert_se(sd_event_add_inotify(e, &b, p, IN_CREATE|IN_DELETE|IN_DONT_FOLLOW, inotify_handler, &context) >= 0);
- assert_se(sd_event_source_set_priority(b, SD_EVENT_PRIORITY_IDLE) >= 0);
- assert_se(sd_event_source_set_priority(b, SD_EVENT_PRIORITY_NORMAL) >= 0);
- assert_se(sd_event_add_inotify(e, &c, p, IN_CREATE|IN_DELETE|IN_EXCL_UNLINK, inotify_handler, &context) >= 0);
- assert_se(sd_event_source_set_priority(c, SD_EVENT_PRIORITY_IDLE) >= 0);
+ ASSERT_OK(sd_event_add_inotify(e, &a, p, IN_CREATE|IN_ONLYDIR, inotify_handler, &context));
+ ASSERT_OK(sd_event_add_inotify(e, &b, p, IN_CREATE|IN_DELETE|IN_DONT_FOLLOW, inotify_handler, &context));
+ ASSERT_OK(sd_event_source_set_priority(b, SD_EVENT_PRIORITY_IDLE));
+ ASSERT_OK(sd_event_source_set_priority(b, SD_EVENT_PRIORITY_NORMAL));
+ ASSERT_OK(sd_event_add_inotify(e, &c, p, IN_CREATE|IN_DELETE|IN_EXCL_UNLINK, inotify_handler, &context));
+ ASSERT_OK(sd_event_source_set_priority(c, SD_EVENT_PRIORITY_IDLE));
- assert_se(sd_event_source_set_description(a, "0") >= 0);
- assert_se(sd_event_source_set_description(b, "1") >= 0);
- assert_se(sd_event_source_set_description(c, "2") >= 0);
+ ASSERT_OK(sd_event_source_set_description(a, "0"));
+ ASSERT_OK(sd_event_source_set_description(b, "1"));
+ ASSERT_OK(sd_event_source_set_description(c, "2"));
- assert_se(sd_event_source_get_inotify_path(a, &pp) >= 0);
- assert_se(path_equal_or_inode_same(pp, p, 0));
- assert_se(sd_event_source_get_inotify_path(b, &pp) >= 0);
- assert_se(path_equal_or_inode_same(pp, p, 0));
- assert_se(sd_event_source_get_inotify_path(b, &pp) >= 0);
- assert_se(path_equal_or_inode_same(pp, p, 0));
+ ASSERT_OK(sd_event_source_get_inotify_path(a, &pp));
+ ASSERT_TRUE(path_equal_or_inode_same(pp, p, 0));
+ ASSERT_OK(sd_event_source_get_inotify_path(b, &pp));
+ ASSERT_TRUE(path_equal_or_inode_same(pp, p, 0));
+ ASSERT_OK(sd_event_source_get_inotify_path(b, &pp));
+ ASSERT_TRUE(path_equal_or_inode_same(pp, p, 0));
q = strjoina(p, "/sub");
- assert_se(touch(q) >= 0);
- assert_se(sd_event_add_inotify(e, &d, q, IN_DELETE_SELF, delete_self_handler, &context) >= 0);
+ ASSERT_OK(touch(q));
+ ASSERT_OK(sd_event_add_inotify(e, &d, q, IN_DELETE_SELF, delete_self_handler, &context));
for (i = 0; i < n_create_events; i++) {
char buf[DECIMAL_STR_MAX(unsigned)+1];
_cleanup_free_ char *z = NULL;
xsprintf(buf, "%u", i);
- assert_se(z = path_join(p, buf));
+ ASSERT_NOT_NULL(z = path_join(p, buf));
- assert_se(touch(z) >= 0);
+ ASSERT_OK(touch(z));
}
- assert_se(unlink(q) >= 0);
+ ASSERT_OK_ERRNO(unlink(q));
- assert_se(sd_event_loop(e) >= 0);
+ ASSERT_OK(sd_event_loop(e));
sd_event_source_unref(a);
sd_event_source_unref(b);
}
static int pidfd_handler(sd_event_source *s, const siginfo_t *si, void *userdata) {
- assert_se(s);
- assert_se(si);
+ ASSERT_NOT_NULL(s);
+ ASSERT_NOT_NULL(si);
- assert_se(si->si_uid == getuid());
- assert_se(si->si_signo == SIGCHLD);
- assert_se(si->si_code == CLD_EXITED);
- assert_se(si->si_status == 66);
+ ASSERT_EQ(si->si_uid, getuid());
+ ASSERT_EQ(si->si_signo, SIGCHLD);
+ ASSERT_EQ(si->si_code, CLD_EXITED);
+ ASSERT_EQ(si->si_status, 66);
log_info("got pidfd on %c", PTR_TO_INT(userdata));
- assert_se(userdata == INT_TO_PTR('p'));
+ ASSERT_PTR_EQ(userdata, INT_TO_PTR('p'));
- assert_se(sd_event_exit(sd_event_source_get_event(s), 0) >= 0);
+ ASSERT_OK(sd_event_exit(sd_event_source_get_event(s), 0));
sd_event_source_unref(s);
return 0;
int pidfd;
pid_t pid, pid2;
- assert_se(sigprocmask_many(SIG_BLOCK, NULL, SIGCHLD) >= 0);
+ ASSERT_OK(sigprocmask_many(SIG_BLOCK, NULL, SIGCHLD));
- pid = fork();
+ ASSERT_OK_ERRNO(pid = fork());
if (pid == 0)
/* child */
_exit(66);
- assert_se(pid > 1);
-
ASSERT_OK(pidfd = pidfd_open(pid, 0));
- pid2 = fork();
+ ASSERT_OK_ERRNO(pid2 = fork());
if (pid2 == 0)
freeze();
- assert_se(pid > 2);
-
- assert_se(sd_event_default(&e) >= 0);
- assert_se(sd_event_add_child_pidfd(e, &s, pidfd, WEXITED, pidfd_handler, INT_TO_PTR('p')) >= 0);
- assert_se(sd_event_source_set_child_pidfd_own(s, true) >= 0);
+ ASSERT_OK(sd_event_default(&e));
+ ASSERT_OK(sd_event_add_child_pidfd(e, &s, pidfd, WEXITED, pidfd_handler, INT_TO_PTR('p')));
+ ASSERT_OK(sd_event_source_set_child_pidfd_own(s, true));
/* This one should never trigger, since our second child lives forever */
- assert_se(sd_event_add_child(e, &t, pid2, WEXITED, pidfd_handler, INT_TO_PTR('q')) >= 0);
- assert_se(sd_event_source_set_child_process_own(t, true) >= 0);
+ ASSERT_OK(sd_event_add_child(e, &t, pid2, WEXITED, pidfd_handler, INT_TO_PTR('q')));
+ ASSERT_OK(sd_event_source_set_child_process_own(t, true));
- assert_se(sd_event_loop(e) >= 0);
+ ASSERT_OK(sd_event_loop(e));
/* Child should still be alive */
- assert_se(kill(pid2, 0) >= 0);
+ ASSERT_OK_ERRNO(kill(pid2, 0));
t = sd_event_source_unref(t);
/* Child should now be dead, since we dropped the ref */
- assert_se(kill(pid2, 0) < 0 && errno == ESRCH);
+ ASSERT_ERROR_ERRNO(kill(pid2, 0), ESRCH);
sd_event_unref(e);
}
uint64_t interval;
unsigned count, burst;
- assert_se(sd_event_default(&e) >= 0);
- assert_se(pipe2(p, O_CLOEXEC|O_NONBLOCK) >= 0);
+ ASSERT_OK(sd_event_default(&e));
+ ASSERT_OK_ERRNO(pipe2(p, O_CLOEXEC|O_NONBLOCK));
- assert_se(sd_event_add_io(e, &s, p[0], EPOLLIN, ratelimit_io_handler, &count) >= 0);
- assert_se(sd_event_source_set_description(s, "test-ratelimit-io") >= 0);
- assert_se(sd_event_source_set_ratelimit(s, 1 * USEC_PER_SEC, 5) >= 0);
- assert_se(sd_event_source_get_ratelimit(s, &interval, &burst) >= 0);
- assert_se(interval == 1 * USEC_PER_SEC && burst == 5);
+ ASSERT_OK(sd_event_add_io(e, &s, p[0], EPOLLIN, ratelimit_io_handler, &count));
+ ASSERT_OK(sd_event_source_set_description(s, "test-ratelimit-io"));
+ ASSERT_OK(sd_event_source_set_ratelimit(s, 1 * USEC_PER_SEC, 5));
+ ASSERT_OK(sd_event_source_get_ratelimit(s, &interval, &burst));
+ ASSERT_EQ(interval, 1 * USEC_PER_SEC);
+ ASSERT_EQ(burst, 5U);
- assert_se(write(p[1], "1", 1) == 1);
+ ASSERT_OK_EQ_ERRNO(write(p[1], "1", 1), 1);
count = 0;
for (unsigned i = 0; i < 10; i++) {
log_debug("slow loop iteration %u", i);
- assert_se(sd_event_run(e, UINT64_MAX) >= 0);
- assert_se(usleep_safe(250 * USEC_PER_MSEC) >= 0);
+ ASSERT_OK(sd_event_run(e, UINT64_MAX));
+ ASSERT_OK(usleep_safe(250 * USEC_PER_MSEC));
}
- assert_se(sd_event_source_is_ratelimited(s) == 0);
- assert_se(count == 10);
+ ASSERT_OK_ZERO(sd_event_source_is_ratelimited(s));
+ ASSERT_EQ(count, 10U);
log_info("ratelimit_io_handler: called %u times, event source not ratelimited", count);
- assert_se(sd_event_source_set_ratelimit(s, 0, 0) >= 0);
- assert_se(sd_event_source_set_ratelimit(s, 1 * USEC_PER_SEC, 5) >= 0);
+ ASSERT_OK(sd_event_source_set_ratelimit(s, 0, 0));
+ ASSERT_OK(sd_event_source_set_ratelimit(s, 1 * USEC_PER_SEC, 5));
count = 0;
for (unsigned i = 0; i < 10; i++) {
log_debug("fast event loop iteration %u", i);
- assert_se(sd_event_run(e, UINT64_MAX) >= 0);
- assert_se(usleep_safe(10) >= 0);
+ ASSERT_OK(sd_event_run(e, UINT64_MAX));
+ ASSERT_OK(usleep_safe(10));
}
log_info("ratelimit_io_handler: called %u times, event source got ratelimited", count);
- assert_se(count < 10);
+ ASSERT_LT(count, 10U);
s = sd_event_source_unref(s);
safe_close_pair(p);
count = 0;
- assert_se(sd_event_add_time_relative(e, &s, CLOCK_MONOTONIC, 1000, 1, ratelimit_time_handler, &count) >= 0);
- assert_se(sd_event_source_set_ratelimit(s, 1 * USEC_PER_SEC, 10) == 0);
+ ASSERT_OK(sd_event_add_time_relative(e, &s, CLOCK_MONOTONIC, 1000, 1, ratelimit_time_handler, &count));
+ ASSERT_OK_ZERO(sd_event_source_set_ratelimit(s, 1 * USEC_PER_SEC, 10));
do {
- assert_se(sd_event_run(e, UINT64_MAX) >= 0);
+ ASSERT_OK(sd_event_run(e, UINT64_MAX));
} while (!sd_event_source_is_ratelimited(s));
log_info("ratelimit_time_handler: called %u times, event source got ratelimited", count);
- assert_se(count == 10);
+ ASSERT_EQ(count, 10U);
/* In order to get rid of active rate limit client needs to disable it explicitly */
- assert_se(sd_event_source_set_ratelimit(s, 0, 0) >= 0);
- assert_se(!sd_event_source_is_ratelimited(s));
+ ASSERT_OK(sd_event_source_set_ratelimit(s, 0, 0));
+ ASSERT_OK_ZERO(sd_event_source_is_ratelimited(s));
- assert_se(sd_event_source_set_ratelimit(s, 1 * USEC_PER_SEC, 10) >= 0);
+ ASSERT_OK(sd_event_source_set_ratelimit(s, 1 * USEC_PER_SEC, 10));
/* Set callback that will be invoked when we leave rate limited state. */
- assert_se(sd_event_source_set_ratelimit_expire_callback(s, ratelimit_expired) >= 0);
+ ASSERT_OK(sd_event_source_set_ratelimit_expire_callback(s, ratelimit_expired));
do {
- assert_se(sd_event_run(e, UINT64_MAX) >= 0);
+ ASSERT_OK(sd_event_run(e, UINT64_MAX));
} while (!sd_event_source_is_ratelimited(s));
log_info("ratelimit_time_handler: called 10 more times, event source got ratelimited");
- assert_se(count == 20);
+ ASSERT_EQ(count, 20U);
/* Dispatch the event loop once more and check that ratelimit expiration callback got called */
- assert_se(sd_event_run(e, UINT64_MAX) >= 0);
- assert_se(expired == 0);
+ ASSERT_OK(sd_event_run(e, UINT64_MAX));
+ ASSERT_EQ(expired, 0);
}
TEST(simple_timeout) {
some_time = random_u64_range(2 * USEC_PER_SEC);
- assert_se(sd_event_default(&e) >= 0);
+ ASSERT_OK(sd_event_default(&e));
- assert_se(sd_event_prepare(e) == 0);
+ ASSERT_OK_ZERO(sd_event_prepare(e));
f = now(CLOCK_MONOTONIC);
- assert_se(sd_event_wait(e, some_time) >= 0);
+ ASSERT_OK(sd_event_wait(e, some_time));
t = now(CLOCK_MONOTONIC);
/* The event loop may sleep longer than the specified time (timer accuracy, scheduling latencies, …),
* but never shorter. Let's check that. */
- assert_se(t >= usec_add(f, some_time));
+ ASSERT_GE(t, usec_add(f, some_time));
}
static int inotify_self_destroy_handler(sd_event_source *s, const struct inotify_event *ev, void *userdata) {
sd_event_source **p = userdata;
- assert_se(ev);
- assert_se(p);
- assert_se(*p == s);
+ ASSERT_NOT_NULL(ev);
+ ASSERT_NOT_NULL(p);
+ ASSERT_PTR_EQ(*p, s);
- assert_se(FLAGS_SET(ev->mask, IN_ATTRIB));
+ ASSERT_TRUE(FLAGS_SET(ev->mask, IN_ATTRIB));
- assert_se(sd_event_exit(sd_event_source_get_event(s), 0) >= 0);
+ ASSERT_OK(sd_event_exit(sd_event_source_get_event(s), 0));
*p = sd_event_source_unref(*p); /* here's what we actually intend to test: we destroy the event
* source from inside the event source handler */
/* Tests that destroying an inotify event source from its own handler is safe */
- assert_se(sd_event_default(&e) >= 0);
+ ASSERT_OK(sd_event_default(&e));
- fd = mkostemp_safe(path);
- assert_se(fd >= 0);
- assert_se(sd_event_add_inotify_fd(e, &s, fd, IN_ATTRIB, inotify_self_destroy_handler, &s) >= 0);
+ ASSERT_OK(fd = mkostemp_safe(path));
+ ASSERT_OK(sd_event_add_inotify_fd(e, &s, fd, IN_ATTRIB, inotify_self_destroy_handler, &s));
fd = safe_close(fd);
- assert_se(unlink(path) >= 0); /* This will trigger IN_ATTRIB because link count goes to zero */
- assert_se(sd_event_loop(e) >= 0);
+ ASSERT_OK_ERRNO(unlink(path)); /* This will trigger IN_ATTRIB because link count goes to zero */
+ ASSERT_OK(sd_event_loop(e));
}
struct inotify_process_buffered_data_context {
struct inotify_process_buffered_data_context *c = ASSERT_PTR(userdata);
const char *description;
- assert_se(sd_event_source_get_description(s, &description) >= 0);
+ ASSERT_OK(sd_event_source_get_description(s, &description));
- assert_se(c->i < 2);
- assert_se(streq(c->path[c->i], description));
+ ASSERT_LT(c->i, 2U);
+ ASSERT_STREQ(c->path[c->i], description);
c->i++;
return 1;
/* For issue #23826 */
- assert_se(sd_event_default(&e) >= 0);
+ ASSERT_OK(sd_event_default(&e));
- assert_se(mkdtemp_malloc("/tmp/test-inotify-XXXXXX", &p) >= 0);
- assert_se(mkdtemp_malloc("/tmp/test-inotify-XXXXXX", &q) >= 0);
+ ASSERT_OK(mkdtemp_malloc("/tmp/test-inotify-XXXXXX", &p));
+ ASSERT_OK(mkdtemp_malloc("/tmp/test-inotify-XXXXXX", &q));
struct inotify_process_buffered_data_context context = {
.path = { p, q },
};
- assert_se(sd_event_add_inotify(e, &a, p, IN_CREATE, inotify_process_buffered_data_handler, &context) >= 0);
- assert_se(sd_event_add_inotify(e, &b, q, IN_CREATE, inotify_process_buffered_data_handler, &context) >= 0);
+ ASSERT_OK(sd_event_add_inotify(e, &a, p, IN_CREATE, inotify_process_buffered_data_handler, &context));
+ ASSERT_OK(sd_event_add_inotify(e, &b, q, IN_CREATE, inotify_process_buffered_data_handler, &context));
- assert_se(z = path_join(p, "aaa"));
- assert_se(touch(z) >= 0);
+ ASSERT_NOT_NULL(z = path_join(p, "aaa"));
+ ASSERT_OK(touch(z));
z = mfree(z);
- assert_se(z = path_join(q, "bbb"));
- assert_se(touch(z) >= 0);
+ ASSERT_NOT_NULL(z = path_join(q, "bbb"));
+ ASSERT_OK(touch(z));
z = mfree(z);
- assert_se(sd_event_run(e, 10 * USEC_PER_SEC) > 0);
- assert_se(sd_event_prepare(e) > 0); /* issue #23826: this was 0. */
- assert_se(sd_event_dispatch(e) > 0);
- assert_se(sd_event_prepare(e) == 0);
- assert_se(sd_event_wait(e, 0) == 0);
+ ASSERT_OK_POSITIVE(sd_event_run(e, 10 * USEC_PER_SEC));
+ ASSERT_OK_POSITIVE(sd_event_prepare(e)); /* issue #23826: this was 0. */
+ ASSERT_OK_POSITIVE(sd_event_dispatch(e));
+ ASSERT_OK_ZERO(sd_event_prepare(e));
+ ASSERT_OK_ZERO(sd_event_wait(e, 0));
}
static int inotify_handler_issue_38265(sd_event_source *s, const struct inotify_event *event, void *userdata) {
_cleanup_(sd_event_unrefp) sd_event *e = NULL;
int r;
- assert_se(sd_event_default(&e) >= 0);
- assert_se(sd_event_prepare(e) == 0);
+ ASSERT_OK(sd_event_default(&e));
+ ASSERT_OK_ZERO(sd_event_prepare(e));
/* Check that after a fork the cleanup functions return NULL */
r = safe_fork("(bus-fork-test)", FORK_WAIT|FORK_LOG, NULL);
if (r == 0) {
- assert_se(e);
- assert_se(sd_event_ref(e) == NULL);
- assert_se(sd_event_unref(e) == NULL);
+ ASSERT_NOT_NULL(e);
+ ASSERT_NULL(sd_event_ref(e));
+ ASSERT_NULL(sd_event_unref(e));
_exit(EXIT_SUCCESS);
}
- assert_se(r >= 0);
+ ASSERT_OK(r);
}
TEST(sd_event_source_set_io_fd) {
_cleanup_(sd_event_unrefp) sd_event *e = NULL;
_cleanup_close_pair_ int pfd_a[2] = EBADF_PAIR, pfd_b[2] = EBADF_PAIR;
- assert_se(sd_event_default(&e) >= 0);
+ ASSERT_OK(sd_event_default(&e));
- assert_se(pipe2(pfd_a, O_CLOEXEC) >= 0);
- assert_se(pipe2(pfd_b, O_CLOEXEC) >= 0);
+ ASSERT_OK_ERRNO(pipe2(pfd_a, O_CLOEXEC));
+ ASSERT_OK_ERRNO(pipe2(pfd_b, O_CLOEXEC));
- assert_se(sd_event_add_io(e, &s, pfd_a[0], EPOLLIN, NULL, INT_TO_PTR(-ENOANO)) >= 0);
- assert_se(sd_event_source_set_io_fd_own(s, true) >= 0);
+ ASSERT_OK(sd_event_add_io(e, &s, pfd_a[0], EPOLLIN, NULL, INT_TO_PTR(-ENOANO)));
+ ASSERT_OK(sd_event_source_set_io_fd_own(s, true));
TAKE_FD(pfd_a[0]);
- assert_se(sd_event_source_set_io_fd(s, pfd_b[0]) >= 0);
+ ASSERT_OK(sd_event_source_set_io_fd(s, pfd_b[0]));
TAKE_FD(pfd_b[0]);
}
static int hup_callback(sd_event_source *s, int fd, uint32_t revents, void *userdata) {
unsigned *c = userdata;
- assert_se(revents == EPOLLHUP);
+ ASSERT_EQ(revents, (uint32_t) EPOLLHUP);
(*c)++;
return 0;
unsigned c = 0;
int r;
- assert_se(sd_event_default(&e) >= 0);
+ ASSERT_OK(sd_event_default(&e));
/* Create an event source that will continuously fire by creating a pipe whose write side is closed,
* and which hence will only see EOF and constant EPOLLHUP */
- assert_se(pipe2(pfd, O_CLOEXEC) >= 0);
- assert_se(sd_event_add_io(e, &s, pfd[0], EPOLLIN, hup_callback, &c) >= 0);
- assert_se(sd_event_source_set_io_fd_own(s, true) >= 0);
- assert_se(sd_event_source_set_ratelimit(s, 5*USEC_PER_MINUTE, 5) >= 0);
+ ASSERT_OK_ERRNO(pipe2(pfd, O_CLOEXEC));
+ ASSERT_OK(sd_event_add_io(e, &s, pfd[0], EPOLLIN, hup_callback, &c));
+ ASSERT_OK(sd_event_source_set_io_fd_own(s, true));
+ ASSERT_OK(sd_event_source_set_ratelimit(s, 5*USEC_PER_MINUTE, 5));
pfd[0] = -EBADF;
pfd[1] = safe_close(pfd[1]); /* Trigger continuous EOF */
for (;;) {
- r = sd_event_prepare(e);
- assert_se(r >= 0);
+ ASSERT_OK(r = sd_event_prepare(e));
- if (r == 0) {
- r = sd_event_wait(e, UINT64_MAX);
- assert_se(r > 0);
- }
+ if (r == 0)
+ ASSERT_OK_POSITIVE(sd_event_wait(e, UINT64_MAX));
- r = sd_event_dispatch(e);
- assert_se(r > 0);
+ ASSERT_OK_POSITIVE(sd_event_dispatch(e));
- r = sd_event_source_is_ratelimited(s);
- assert_se(r >= 0);
+ ASSERT_OK(r = sd_event_source_is_ratelimited(s));
if (c < 5)
/* First four dispatches should just work */
- assert_se(!r);
+ ASSERT_FALSE(r);
else if (c == 5) {
/* The fifth dispatch should still work, but we now expect the ratelimit to be hit subsequently */
if (!expect_ratelimit) {
- assert_se(!r);
- assert_se(sd_event_source_leave_ratelimit(s) == 0); /* this should be a NOP, and return 0 hence */
+ ASSERT_FALSE(r);
+ ASSERT_OK_ZERO(sd_event_source_leave_ratelimit(s)); /* this should be a NOP, and return 0 hence */
expect_ratelimit = true;
} else {
/* We expected the ratelimit, let's leave it manually, and verify it */
- assert_se(r);
- assert_se(sd_event_source_leave_ratelimit(s) > 0); /* we are ratelimited, hence should return > 0 */
- assert_se(sd_event_source_is_ratelimited(s) == 0);
+ ASSERT_TRUE(r);
+ ASSERT_OK_POSITIVE(sd_event_source_leave_ratelimit(s)); /* we are ratelimited, hence should return > 0 */
+ ASSERT_OK_ZERO(sd_event_source_is_ratelimited(s));
manually_left_ratelimit = true;
}
}
/* Verify we definitely hit the ratelimit and left it manually again */
- assert_se(manually_left_ratelimit);
+ ASSERT_TRUE(manually_left_ratelimit);
}
DEFINE_TEST_MAIN(LOG_DEBUG);