1 /*-*- Mode: C; c-basic-offset: 8; indent-tabs-mode: nil -*-*/
4 This file is part of systemd.
6 Copyright 2010 Lennart Poettering
8 systemd is free software; you can redistribute it and/or modify it
9 under the terms of the GNU Lesser General Public License as published by
10 the Free Software Foundation; either version 2.1 of the License, or
11 (at your option) any later version.
13 systemd is distributed in the hope that it will be useful, but
14 WITHOUT ANY WARRANTY; without even the implied warranty of
15 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
16 Lesser General Public License for more details.
18 You should have received a copy of the GNU Lesser General Public License
19 along with systemd; If not, see <http://www.gnu.org/licenses/>.
29 #include "sd-messages.h"
34 #include "path-util.h"
35 #include "load-fragment.h"
36 #include "load-dropin.h"
38 #include "unit-name.h"
39 #include "dbus-unit.h"
41 #include "cgroup-util.h"
44 #include "fileio-label.h"
45 #include "bus-common-errors.h"
49 #include "formats-util.h"
50 #include "process-util.h"
52 const UnitVTable
* const unit_vtable
[_UNIT_TYPE_MAX
] = {
53 [UNIT_SERVICE
] = &service_vtable
,
54 [UNIT_SOCKET
] = &socket_vtable
,
55 [UNIT_BUSNAME
] = &busname_vtable
,
56 [UNIT_TARGET
] = &target_vtable
,
57 [UNIT_SNAPSHOT
] = &snapshot_vtable
,
58 [UNIT_DEVICE
] = &device_vtable
,
59 [UNIT_MOUNT
] = &mount_vtable
,
60 [UNIT_AUTOMOUNT
] = &automount_vtable
,
61 [UNIT_SWAP
] = &swap_vtable
,
62 [UNIT_TIMER
] = &timer_vtable
,
63 [UNIT_PATH
] = &path_vtable
,
64 [UNIT_SLICE
] = &slice_vtable
,
65 [UNIT_SCOPE
] = &scope_vtable
68 static int maybe_warn_about_dependency(const char *id
, const char *other
, UnitDependency dependency
);
70 Unit
*unit_new(Manager
*m
, size_t size
) {
74 assert(size
>= sizeof(Unit
));
80 u
->names
= set_new(&string_hash_ops
);
87 u
->type
= _UNIT_TYPE_INVALID
;
88 u
->deserialized_job
= _JOB_TYPE_INVALID
;
89 u
->default_dependencies
= true;
90 u
->unit_file_state
= _UNIT_FILE_STATE_INVALID
;
91 u
->unit_file_preset
= -1;
92 u
->on_failure_job_mode
= JOB_REPLACE
;
97 bool unit_has_name(Unit
*u
, const char *name
) {
101 return !!set_get(u
->names
, (char*) name
);
104 static void unit_init(Unit
*u
) {
111 assert(u
->type
>= 0);
113 cc
= unit_get_cgroup_context(u
);
115 cgroup_context_init(cc
);
117 /* Copy in the manager defaults into the cgroup
118 * context, _before_ the rest of the settings have
119 * been initialized */
121 cc
->cpu_accounting
= u
->manager
->default_cpu_accounting
;
122 cc
->blockio_accounting
= u
->manager
->default_blockio_accounting
;
123 cc
->memory_accounting
= u
->manager
->default_memory_accounting
;
126 ec
= unit_get_exec_context(u
);
128 exec_context_init(ec
);
130 kc
= unit_get_kill_context(u
);
132 kill_context_init(kc
);
134 if (UNIT_VTABLE(u
)->init
)
135 UNIT_VTABLE(u
)->init(u
);
138 int unit_add_name(Unit
*u
, const char *text
) {
139 _cleanup_free_
char *s
= NULL
, *i
= NULL
;
146 if (unit_name_is_template(text
)) {
151 s
= unit_name_replace_instance(text
, u
->instance
);
157 if (!unit_name_is_valid(s
, TEMPLATE_INVALID
))
160 assert_se((t
= unit_name_to_type(s
)) >= 0);
162 if (u
->type
!= _UNIT_TYPE_INVALID
&& t
!= u
->type
)
165 r
= unit_name_to_instance(s
, &i
);
169 if (i
&& unit_vtable
[t
]->no_instances
)
172 /* Ensure that this unit is either instanced or not instanced,
174 if (u
->type
!= _UNIT_TYPE_INVALID
&& !u
->instance
!= !i
)
177 if (unit_vtable
[t
]->no_alias
&&
178 !set_isempty(u
->names
) &&
179 !set_get(u
->names
, s
))
182 if (hashmap_size(u
->manager
->units
) >= MANAGER_MAX_NAMES
)
185 r
= set_put(u
->names
, s
);
189 r
= hashmap_put(u
->manager
->units
, s
, u
);
191 set_remove(u
->names
, s
);
195 if (u
->type
== _UNIT_TYPE_INVALID
) {
200 LIST_PREPEND(units_by_type
, u
->manager
->units_by_type
[t
], u
);
209 unit_add_to_dbus_queue(u
);
213 int unit_choose_id(Unit
*u
, const char *name
) {
214 _cleanup_free_
char *t
= NULL
;
221 if (unit_name_is_template(name
)) {
226 t
= unit_name_replace_instance(name
, u
->instance
);
233 /* Selects one of the names of this unit as the id */
234 s
= set_get(u
->names
, (char*) name
);
238 r
= unit_name_to_instance(s
, &i
);
247 unit_add_to_dbus_queue(u
);
252 int unit_set_description(Unit
*u
, const char *description
) {
257 if (isempty(description
))
260 s
= strdup(description
);
265 free(u
->description
);
268 unit_add_to_dbus_queue(u
);
272 bool unit_check_gc(Unit
*u
) {
273 UnitActiveState state
;
282 state
= unit_active_state(u
);
284 /* If the unit is inactive and failed and no job is queued for
285 * it, then release its runtime resources */
286 if (UNIT_IS_INACTIVE_OR_FAILED(state
) &&
287 UNIT_VTABLE(u
)->release_resources
)
288 UNIT_VTABLE(u
)->release_resources(u
);
290 /* But we keep the unit object around for longer when it is
291 * referenced or configured to not be gc'ed */
292 if (state
!= UNIT_INACTIVE
)
295 if (UNIT_VTABLE(u
)->no_gc
)
304 if (UNIT_VTABLE(u
)->check_gc
)
305 if (UNIT_VTABLE(u
)->check_gc(u
))
311 void unit_add_to_load_queue(Unit
*u
) {
313 assert(u
->type
!= _UNIT_TYPE_INVALID
);
315 if (u
->load_state
!= UNIT_STUB
|| u
->in_load_queue
)
318 LIST_PREPEND(load_queue
, u
->manager
->load_queue
, u
);
319 u
->in_load_queue
= true;
322 void unit_add_to_cleanup_queue(Unit
*u
) {
325 if (u
->in_cleanup_queue
)
328 LIST_PREPEND(cleanup_queue
, u
->manager
->cleanup_queue
, u
);
329 u
->in_cleanup_queue
= true;
332 void unit_add_to_gc_queue(Unit
*u
) {
335 if (u
->in_gc_queue
|| u
->in_cleanup_queue
)
338 if (unit_check_gc(u
))
341 LIST_PREPEND(gc_queue
, u
->manager
->gc_queue
, u
);
342 u
->in_gc_queue
= true;
344 u
->manager
->n_in_gc_queue
++;
347 void unit_add_to_dbus_queue(Unit
*u
) {
349 assert(u
->type
!= _UNIT_TYPE_INVALID
);
351 if (u
->load_state
== UNIT_STUB
|| u
->in_dbus_queue
)
354 /* Shortcut things if nobody cares */
355 if (sd_bus_track_count(u
->manager
->subscribed
) <= 0 &&
356 set_isempty(u
->manager
->private_buses
)) {
357 u
->sent_dbus_new_signal
= true;
361 LIST_PREPEND(dbus_queue
, u
->manager
->dbus_unit_queue
, u
);
362 u
->in_dbus_queue
= true;
365 static void bidi_set_free(Unit
*u
, Set
*s
) {
371 /* Frees the set and makes sure we are dropped from the
372 * inverse pointers */
374 SET_FOREACH(other
, s
, i
) {
377 for (d
= 0; d
< _UNIT_DEPENDENCY_MAX
; d
++)
378 set_remove(other
->dependencies
[d
], u
);
380 unit_add_to_gc_queue(other
);
386 static void unit_remove_transient(Unit
*u
) {
394 if (u
->fragment_path
)
395 unlink(u
->fragment_path
);
397 STRV_FOREACH(i
, u
->dropin_paths
) {
398 _cleanup_free_
char *p
= NULL
;
403 r
= path_get_parent(*i
, &p
);
409 static void unit_free_requires_mounts_for(Unit
*u
) {
412 STRV_FOREACH(j
, u
->requires_mounts_for
) {
413 char s
[strlen(*j
) + 1];
415 PATH_FOREACH_PREFIX_MORE(s
, *j
) {
419 x
= hashmap_get2(u
->manager
->units_requiring_mounts_for
, s
, (void**) &y
);
425 if (set_isempty(x
)) {
426 hashmap_remove(u
->manager
->units_requiring_mounts_for
, y
);
433 strv_free(u
->requires_mounts_for
);
434 u
->requires_mounts_for
= NULL
;
437 static void unit_done(Unit
*u
) {
446 if (UNIT_VTABLE(u
)->done
)
447 UNIT_VTABLE(u
)->done(u
);
449 ec
= unit_get_exec_context(u
);
451 exec_context_done(ec
);
453 cc
= unit_get_cgroup_context(u
);
455 cgroup_context_done(cc
);
458 void unit_free(Unit
*u
) {
465 if (u
->manager
->n_reloading
<= 0)
466 unit_remove_transient(u
);
468 bus_unit_send_removed_signal(u
);
472 unit_free_requires_mounts_for(u
);
474 SET_FOREACH(t
, u
->names
, i
)
475 hashmap_remove_value(u
->manager
->units
, t
, u
);
489 for (d
= 0; d
< _UNIT_DEPENDENCY_MAX
; d
++)
490 bidi_set_free(u
, u
->dependencies
[d
]);
492 if (u
->type
!= _UNIT_TYPE_INVALID
)
493 LIST_REMOVE(units_by_type
, u
->manager
->units_by_type
[u
->type
], u
);
495 if (u
->in_load_queue
)
496 LIST_REMOVE(load_queue
, u
->manager
->load_queue
, u
);
498 if (u
->in_dbus_queue
)
499 LIST_REMOVE(dbus_queue
, u
->manager
->dbus_unit_queue
, u
);
501 if (u
->in_cleanup_queue
)
502 LIST_REMOVE(cleanup_queue
, u
->manager
->cleanup_queue
, u
);
504 if (u
->in_gc_queue
) {
505 LIST_REMOVE(gc_queue
, u
->manager
->gc_queue
, u
);
506 u
->manager
->n_in_gc_queue
--;
509 if (u
->in_cgroup_queue
)
510 LIST_REMOVE(cgroup_queue
, u
->manager
->cgroup_queue
, u
);
512 if (u
->cgroup_path
) {
513 hashmap_remove(u
->manager
->cgroup_unit
, u
->cgroup_path
);
514 free(u
->cgroup_path
);
517 manager_update_failed_units(u
->manager
, u
, false);
518 set_remove(u
->manager
->startup_units
, u
);
520 free(u
->description
);
521 strv_free(u
->documentation
);
522 free(u
->fragment_path
);
523 free(u
->source_path
);
524 strv_free(u
->dropin_paths
);
527 free(u
->job_timeout_reboot_arg
);
529 set_free_free(u
->names
);
531 unit_unwatch_all_pids(u
);
533 condition_free_list(u
->conditions
);
534 condition_free_list(u
->asserts
);
536 unit_ref_unset(&u
->slice
);
539 unit_ref_unset(u
->refs
);
544 UnitActiveState
unit_active_state(Unit
*u
) {
547 if (u
->load_state
== UNIT_MERGED
)
548 return unit_active_state(unit_follow_merge(u
));
550 /* After a reload it might happen that a unit is not correctly
551 * loaded but still has a process around. That's why we won't
552 * shortcut failed loading to UNIT_INACTIVE_FAILED. */
554 return UNIT_VTABLE(u
)->active_state(u
);
557 const char* unit_sub_state_to_string(Unit
*u
) {
560 return UNIT_VTABLE(u
)->sub_state_to_string(u
);
563 static int complete_move(Set
**s
, Set
**other
) {
573 r
= set_move(*s
, *other
);
584 static int merge_names(Unit
*u
, Unit
*other
) {
592 r
= complete_move(&u
->names
, &other
->names
);
596 set_free_free(other
->names
);
600 SET_FOREACH(t
, u
->names
, i
)
601 assert_se(hashmap_replace(u
->manager
->units
, t
, u
) == 0);
606 static int reserve_dependencies(Unit
*u
, Unit
*other
, UnitDependency d
) {
611 assert(d
< _UNIT_DEPENDENCY_MAX
);
614 * If u does not have this dependency set allocated, there is no need
615 * to reserve anything. In that case other's set will be transferred
616 * as a whole to u by complete_move().
618 if (!u
->dependencies
[d
])
621 /* merge_dependencies() will skip a u-on-u dependency */
622 n_reserve
= set_size(other
->dependencies
[d
]) - !!set_get(other
->dependencies
[d
], u
);
624 return set_reserve(u
->dependencies
[d
], n_reserve
);
627 static void merge_dependencies(Unit
*u
, Unit
*other
, const char *other_id
, UnitDependency d
) {
634 assert(d
< _UNIT_DEPENDENCY_MAX
);
636 /* Fix backwards pointers */
637 SET_FOREACH(back
, other
->dependencies
[d
], i
) {
640 for (k
= 0; k
< _UNIT_DEPENDENCY_MAX
; k
++) {
641 /* Do not add dependencies between u and itself */
643 if (set_remove(back
->dependencies
[k
], other
))
644 maybe_warn_about_dependency(u
->id
, other_id
, k
);
646 r
= set_remove_and_put(back
->dependencies
[k
], other
, u
);
648 set_remove(back
->dependencies
[k
], other
);
650 assert(r
>= 0 || r
== -ENOENT
);
655 /* Also do not move dependencies on u to itself */
656 back
= set_remove(other
->dependencies
[d
], u
);
658 maybe_warn_about_dependency(u
->id
, other_id
, d
);
660 /* The move cannot fail. The caller must have performed a reservation. */
661 assert_se(complete_move(&u
->dependencies
[d
], &other
->dependencies
[d
]) == 0);
663 set_free(other
->dependencies
[d
]);
664 other
->dependencies
[d
] = NULL
;
667 int unit_merge(Unit
*u
, Unit
*other
) {
669 const char *other_id
= NULL
;
674 assert(u
->manager
== other
->manager
);
675 assert(u
->type
!= _UNIT_TYPE_INVALID
);
677 other
= unit_follow_merge(other
);
682 if (u
->type
!= other
->type
)
685 if (!u
->instance
!= !other
->instance
)
688 if (other
->load_state
!= UNIT_STUB
&&
689 other
->load_state
!= UNIT_NOT_FOUND
)
698 if (!UNIT_IS_INACTIVE_OR_FAILED(unit_active_state(other
)))
702 other_id
= strdupa(other
->id
);
704 /* Make reservations to ensure merge_dependencies() won't fail */
705 for (d
= 0; d
< _UNIT_DEPENDENCY_MAX
; d
++) {
706 r
= reserve_dependencies(u
, other
, d
);
708 * We don't rollback reservations if we fail. We don't have
709 * a way to undo reservations. A reservation is not a leak.
716 r
= merge_names(u
, other
);
720 /* Redirect all references */
722 unit_ref_set(other
->refs
, u
);
724 /* Merge dependencies */
725 for (d
= 0; d
< _UNIT_DEPENDENCY_MAX
; d
++)
726 merge_dependencies(u
, other
, other_id
, d
);
728 other
->load_state
= UNIT_MERGED
;
729 other
->merged_into
= u
;
731 /* If there is still some data attached to the other node, we
732 * don't need it anymore, and can free it. */
733 if (other
->load_state
!= UNIT_STUB
)
734 if (UNIT_VTABLE(other
)->done
)
735 UNIT_VTABLE(other
)->done(other
);
737 unit_add_to_dbus_queue(u
);
738 unit_add_to_cleanup_queue(other
);
743 int unit_merge_by_name(Unit
*u
, const char *name
) {
746 _cleanup_free_
char *s
= NULL
;
751 if (unit_name_is_template(name
)) {
755 s
= unit_name_replace_instance(name
, u
->instance
);
762 other
= manager_get_unit(u
->manager
, name
);
764 r
= unit_add_name(u
, name
);
766 r
= unit_merge(u
, other
);
771 Unit
* unit_follow_merge(Unit
*u
) {
774 while (u
->load_state
== UNIT_MERGED
)
775 assert_se(u
= u
->merged_into
);
780 int unit_add_exec_dependencies(Unit
*u
, ExecContext
*c
) {
786 if (c
->working_directory
) {
787 r
= unit_require_mounts_for(u
, c
->working_directory
);
792 if (c
->root_directory
) {
793 r
= unit_require_mounts_for(u
, c
->root_directory
);
798 if (u
->manager
->running_as
!= SYSTEMD_SYSTEM
)
801 if (c
->private_tmp
) {
802 r
= unit_require_mounts_for(u
, "/tmp");
806 r
= unit_require_mounts_for(u
, "/var/tmp");
811 if (c
->std_output
!= EXEC_OUTPUT_KMSG
&&
812 c
->std_output
!= EXEC_OUTPUT_SYSLOG
&&
813 c
->std_output
!= EXEC_OUTPUT_JOURNAL
&&
814 c
->std_output
!= EXEC_OUTPUT_KMSG_AND_CONSOLE
&&
815 c
->std_output
!= EXEC_OUTPUT_SYSLOG_AND_CONSOLE
&&
816 c
->std_output
!= EXEC_OUTPUT_JOURNAL_AND_CONSOLE
&&
817 c
->std_error
!= EXEC_OUTPUT_KMSG
&&
818 c
->std_error
!= EXEC_OUTPUT_SYSLOG
&&
819 c
->std_error
!= EXEC_OUTPUT_JOURNAL
&&
820 c
->std_error
!= EXEC_OUTPUT_KMSG_AND_CONSOLE
&&
821 c
->std_error
!= EXEC_OUTPUT_JOURNAL_AND_CONSOLE
&&
822 c
->std_error
!= EXEC_OUTPUT_SYSLOG_AND_CONSOLE
)
825 /* If syslog or kernel logging is requested, make sure our own
826 * logging daemon is run first. */
828 r
= unit_add_dependency_by_name(u
, UNIT_AFTER
, SPECIAL_JOURNALD_SOCKET
, NULL
, true);
835 const char *unit_description(Unit
*u
) {
839 return u
->description
;
844 void unit_dump(Unit
*u
, FILE *f
, const char *prefix
) {
850 timestamp1
[FORMAT_TIMESTAMP_MAX
],
851 timestamp2
[FORMAT_TIMESTAMP_MAX
],
852 timestamp3
[FORMAT_TIMESTAMP_MAX
],
853 timestamp4
[FORMAT_TIMESTAMP_MAX
],
854 timespan
[FORMAT_TIMESPAN_MAX
];
856 _cleanup_set_free_ Set
*following_set
= NULL
;
860 assert(u
->type
>= 0);
862 prefix
= strempty(prefix
);
863 prefix2
= strjoina(prefix
, "\t");
867 "%s\tDescription: %s\n"
869 "%s\tUnit Load State: %s\n"
870 "%s\tUnit Active State: %s\n"
871 "%s\tInactive Exit Timestamp: %s\n"
872 "%s\tActive Enter Timestamp: %s\n"
873 "%s\tActive Exit Timestamp: %s\n"
874 "%s\tInactive Enter Timestamp: %s\n"
875 "%s\tGC Check Good: %s\n"
876 "%s\tNeed Daemon Reload: %s\n"
877 "%s\tTransient: %s\n"
880 "%s\tCGroup realized: %s\n"
881 "%s\tCGroup mask: 0x%x\n"
882 "%s\tCGroup members mask: 0x%x\n",
884 prefix
, unit_description(u
),
885 prefix
, strna(u
->instance
),
886 prefix
, unit_load_state_to_string(u
->load_state
),
887 prefix
, unit_active_state_to_string(unit_active_state(u
)),
888 prefix
, strna(format_timestamp(timestamp1
, sizeof(timestamp1
), u
->inactive_exit_timestamp
.realtime
)),
889 prefix
, strna(format_timestamp(timestamp2
, sizeof(timestamp2
), u
->active_enter_timestamp
.realtime
)),
890 prefix
, strna(format_timestamp(timestamp3
, sizeof(timestamp3
), u
->active_exit_timestamp
.realtime
)),
891 prefix
, strna(format_timestamp(timestamp4
, sizeof(timestamp4
), u
->inactive_enter_timestamp
.realtime
)),
892 prefix
, yes_no(unit_check_gc(u
)),
893 prefix
, yes_no(unit_need_daemon_reload(u
)),
894 prefix
, yes_no(u
->transient
),
895 prefix
, strna(unit_slice_name(u
)),
896 prefix
, strna(u
->cgroup_path
),
897 prefix
, yes_no(u
->cgroup_realized
),
898 prefix
, u
->cgroup_realized_mask
,
899 prefix
, u
->cgroup_members_mask
);
901 SET_FOREACH(t
, u
->names
, i
)
902 fprintf(f
, "%s\tName: %s\n", prefix
, t
);
904 STRV_FOREACH(j
, u
->documentation
)
905 fprintf(f
, "%s\tDocumentation: %s\n", prefix
, *j
);
907 following
= unit_following(u
);
909 fprintf(f
, "%s\tFollowing: %s\n", prefix
, following
->id
);
911 r
= unit_following_set(u
, &following_set
);
915 SET_FOREACH(other
, following_set
, i
)
916 fprintf(f
, "%s\tFollowing Set Member: %s\n", prefix
, other
->id
);
919 if (u
->fragment_path
)
920 fprintf(f
, "%s\tFragment Path: %s\n", prefix
, u
->fragment_path
);
923 fprintf(f
, "%s\tSource Path: %s\n", prefix
, u
->source_path
);
925 STRV_FOREACH(j
, u
->dropin_paths
)
926 fprintf(f
, "%s\tDropIn Path: %s\n", prefix
, *j
);
928 if (u
->job_timeout
> 0)
929 fprintf(f
, "%s\tJob Timeout: %s\n", prefix
, format_timespan(timespan
, sizeof(timespan
), u
->job_timeout
, 0));
931 if (u
->job_timeout_action
!= FAILURE_ACTION_NONE
)
932 fprintf(f
, "%s\tJob Timeout Action: %s\n", prefix
, failure_action_to_string(u
->job_timeout_action
));
934 if (u
->job_timeout_reboot_arg
)
935 fprintf(f
, "%s\tJob Timeout Reboot Argument: %s\n", prefix
, u
->job_timeout_reboot_arg
);
937 condition_dump_list(u
->conditions
, f
, prefix
, condition_type_to_string
);
938 condition_dump_list(u
->asserts
, f
, prefix
, assert_type_to_string
);
940 if (dual_timestamp_is_set(&u
->condition_timestamp
))
942 "%s\tCondition Timestamp: %s\n"
943 "%s\tCondition Result: %s\n",
944 prefix
, strna(format_timestamp(timestamp1
, sizeof(timestamp1
), u
->condition_timestamp
.realtime
)),
945 prefix
, yes_no(u
->condition_result
));
947 if (dual_timestamp_is_set(&u
->assert_timestamp
))
949 "%s\tAssert Timestamp: %s\n"
950 "%s\tAssert Result: %s\n",
951 prefix
, strna(format_timestamp(timestamp1
, sizeof(timestamp1
), u
->assert_timestamp
.realtime
)),
952 prefix
, yes_no(u
->assert_result
));
954 for (d
= 0; d
< _UNIT_DEPENDENCY_MAX
; d
++) {
957 SET_FOREACH(other
, u
->dependencies
[d
], i
)
958 fprintf(f
, "%s\t%s: %s\n", prefix
, unit_dependency_to_string(d
), other
->id
);
961 if (!strv_isempty(u
->requires_mounts_for
)) {
963 "%s\tRequiresMountsFor:", prefix
);
965 STRV_FOREACH(j
, u
->requires_mounts_for
)
966 fprintf(f
, " %s", *j
);
971 if (u
->load_state
== UNIT_LOADED
) {
974 "%s\tStopWhenUnneeded: %s\n"
975 "%s\tRefuseManualStart: %s\n"
976 "%s\tRefuseManualStop: %s\n"
977 "%s\tDefaultDependencies: %s\n"
978 "%s\tOnFailureJobMode: %s\n"
979 "%s\tIgnoreOnIsolate: %s\n"
980 "%s\tIgnoreOnSnapshot: %s\n",
981 prefix
, yes_no(u
->stop_when_unneeded
),
982 prefix
, yes_no(u
->refuse_manual_start
),
983 prefix
, yes_no(u
->refuse_manual_stop
),
984 prefix
, yes_no(u
->default_dependencies
),
985 prefix
, job_mode_to_string(u
->on_failure_job_mode
),
986 prefix
, yes_no(u
->ignore_on_isolate
),
987 prefix
, yes_no(u
->ignore_on_snapshot
));
989 if (UNIT_VTABLE(u
)->dump
)
990 UNIT_VTABLE(u
)->dump(u
, f
, prefix2
);
992 } else if (u
->load_state
== UNIT_MERGED
)
994 "%s\tMerged into: %s\n",
995 prefix
, u
->merged_into
->id
);
996 else if (u
->load_state
== UNIT_ERROR
)
997 fprintf(f
, "%s\tLoad Error Code: %s\n", prefix
, strerror(-u
->load_error
));
1001 job_dump(u
->job
, f
, prefix2
);
1004 job_dump(u
->nop_job
, f
, prefix2
);
1008 /* Common implementation for multiple backends */
1009 int unit_load_fragment_and_dropin(Unit
*u
) {
1014 /* Load a .{service,socket,...} file */
1015 r
= unit_load_fragment(u
);
1019 if (u
->load_state
== UNIT_STUB
)
1022 /* Load drop-in directory data */
1023 r
= unit_load_dropin(unit_follow_merge(u
));
1030 /* Common implementation for multiple backends */
1031 int unit_load_fragment_and_dropin_optional(Unit
*u
) {
1036 /* Same as unit_load_fragment_and_dropin(), but whether
1037 * something can be loaded or not doesn't matter. */
1039 /* Load a .service file */
1040 r
= unit_load_fragment(u
);
1044 if (u
->load_state
== UNIT_STUB
)
1045 u
->load_state
= UNIT_LOADED
;
1047 /* Load drop-in directory data */
1048 r
= unit_load_dropin(unit_follow_merge(u
));
1055 int unit_add_default_target_dependency(Unit
*u
, Unit
*target
) {
1059 if (target
->type
!= UNIT_TARGET
)
1062 /* Only add the dependency if both units are loaded, so that
1063 * that loop check below is reliable */
1064 if (u
->load_state
!= UNIT_LOADED
||
1065 target
->load_state
!= UNIT_LOADED
)
1068 /* If either side wants no automatic dependencies, then let's
1070 if (!u
->default_dependencies
||
1071 !target
->default_dependencies
)
1074 /* Don't create loops */
1075 if (set_get(target
->dependencies
[UNIT_BEFORE
], u
))
1078 return unit_add_dependency(target
, UNIT_AFTER
, u
, true);
1081 static int unit_add_target_dependencies(Unit
*u
) {
1083 static const UnitDependency deps
[] = {
1085 UNIT_REQUIRED_BY_OVERRIDABLE
,
1097 for (k
= 0; k
< ELEMENTSOF(deps
); k
++)
1098 SET_FOREACH(target
, u
->dependencies
[deps
[k
]], i
) {
1099 r
= unit_add_default_target_dependency(u
, target
);
1107 static int unit_add_slice_dependencies(Unit
*u
) {
1110 if (!unit_get_cgroup_context(u
))
1113 if (UNIT_ISSET(u
->slice
))
1114 return unit_add_two_dependencies(u
, UNIT_AFTER
, UNIT_WANTS
, UNIT_DEREF(u
->slice
), true);
1116 if (streq(u
->id
, SPECIAL_ROOT_SLICE
))
1119 return unit_add_two_dependencies_by_name(u
, UNIT_AFTER
, UNIT_WANTS
, SPECIAL_ROOT_SLICE
, NULL
, true);
1122 static int unit_add_mount_dependencies(Unit
*u
) {
1128 STRV_FOREACH(i
, u
->requires_mounts_for
) {
1129 char prefix
[strlen(*i
) + 1];
1131 PATH_FOREACH_PREFIX_MORE(prefix
, *i
) {
1134 r
= manager_get_unit_by_path(u
->manager
, prefix
, ".mount", &m
);
1142 if (m
->load_state
!= UNIT_LOADED
)
1145 r
= unit_add_dependency(u
, UNIT_AFTER
, m
, true);
1149 if (m
->fragment_path
) {
1150 r
= unit_add_dependency(u
, UNIT_REQUIRES
, m
, true);
1160 static int unit_add_startup_units(Unit
*u
) {
1163 c
= unit_get_cgroup_context(u
);
1167 if (c
->startup_cpu_shares
== (unsigned long) -1 &&
1168 c
->startup_blockio_weight
== (unsigned long) -1)
1171 return set_put(u
->manager
->startup_units
, u
);
1174 int unit_load(Unit
*u
) {
1179 if (u
->in_load_queue
) {
1180 LIST_REMOVE(load_queue
, u
->manager
->load_queue
, u
);
1181 u
->in_load_queue
= false;
1184 if (u
->type
== _UNIT_TYPE_INVALID
)
1187 if (u
->load_state
!= UNIT_STUB
)
1190 if (UNIT_VTABLE(u
)->load
) {
1191 r
= UNIT_VTABLE(u
)->load(u
);
1196 if (u
->load_state
== UNIT_STUB
) {
1201 if (u
->load_state
== UNIT_LOADED
) {
1203 r
= unit_add_target_dependencies(u
);
1207 r
= unit_add_slice_dependencies(u
);
1211 r
= unit_add_mount_dependencies(u
);
1215 r
= unit_add_startup_units(u
);
1219 if (u
->on_failure_job_mode
== JOB_ISOLATE
&& set_size(u
->dependencies
[UNIT_ON_FAILURE
]) > 1) {
1220 log_unit_error(u
->id
, "More than one OnFailure= dependencies specified for %s but OnFailureJobMode=isolate set. Refusing.", u
->id
);
1225 unit_update_cgroup_members_masks(u
);
1228 assert((u
->load_state
!= UNIT_MERGED
) == !u
->merged_into
);
1230 unit_add_to_dbus_queue(unit_follow_merge(u
));
1231 unit_add_to_gc_queue(u
);
1236 u
->load_state
= u
->load_state
== UNIT_STUB
? UNIT_NOT_FOUND
: UNIT_ERROR
;
1238 unit_add_to_dbus_queue(u
);
1239 unit_add_to_gc_queue(u
);
1241 log_unit_debug(u
->id
, "Failed to load configuration for %s: %s",
1242 u
->id
, strerror(-r
));
1247 static bool unit_condition_test_list(Unit
*u
, Condition
*first
, const char *(*to_string
)(ConditionType t
)) {
1254 /* If the condition list is empty, then it is true */
1258 /* Otherwise, if all of the non-trigger conditions apply and
1259 * if any of the trigger conditions apply (unless there are
1260 * none) we return true */
1261 LIST_FOREACH(conditions
, c
, first
) {
1264 r
= condition_test(c
);
1266 log_unit_warning(u
->id
,
1267 "Couldn't determine result for %s=%s%s%s for %s, assuming failed: %s",
1269 c
->trigger
? "|" : "",
1270 c
->negate
? "!" : "",
1275 log_unit_debug(u
->id
,
1276 "%s=%s%s%s %s for %s.",
1278 c
->trigger
? "|" : "",
1279 c
->negate
? "!" : "",
1281 condition_result_to_string(c
->result
),
1284 if (!c
->trigger
&& r
<= 0)
1287 if (c
->trigger
&& triggered
<= 0)
1291 return triggered
!= 0;
1294 static bool unit_condition_test(Unit
*u
) {
1297 dual_timestamp_get(&u
->condition_timestamp
);
1298 u
->condition_result
= unit_condition_test_list(u
, u
->conditions
, condition_type_to_string
);
1300 return u
->condition_result
;
1303 static bool unit_assert_test(Unit
*u
) {
1306 dual_timestamp_get(&u
->assert_timestamp
);
1307 u
->assert_result
= unit_condition_test_list(u
, u
->asserts
, assert_type_to_string
);
1309 return u
->assert_result
;
1312 _pure_
static const char* unit_get_status_message_format(Unit
*u
, JobType t
) {
1313 const UnitStatusMessageFormats
*format_table
;
1317 assert(t
< _JOB_TYPE_MAX
);
1319 if (t
!= JOB_START
&& t
!= JOB_STOP
)
1322 format_table
= &UNIT_VTABLE(u
)->status_message_formats
;
1326 return format_table
->starting_stopping
[t
== JOB_STOP
];
1329 _pure_
static const char *unit_get_status_message_format_try_harder(Unit
*u
, JobType t
) {
1334 assert(t
< _JOB_TYPE_MAX
);
1336 format
= unit_get_status_message_format(u
, t
);
1340 /* Return generic strings */
1342 return "Starting %s.";
1343 else if (t
== JOB_STOP
)
1344 return "Stopping %s.";
1345 else if (t
== JOB_RELOAD
)
1346 return "Reloading %s.";
1351 static void unit_status_print_starting_stopping(Unit
*u
, JobType t
) {
1356 /* We only print status messages for selected units on
1357 * selected operations. */
1359 format
= unit_get_status_message_format(u
, t
);
1363 DISABLE_WARNING_FORMAT_NONLITERAL
;
1364 unit_status_printf(u
, "", format
);
1368 static void unit_status_log_starting_stopping_reloading(Unit
*u
, JobType t
) {
1375 if (t
!= JOB_START
&& t
!= JOB_STOP
&& t
!= JOB_RELOAD
)
1378 if (log_on_console())
1381 /* We log status messages for all units and all operations. */
1383 format
= unit_get_status_message_format_try_harder(u
, t
);
1387 DISABLE_WARNING_FORMAT_NONLITERAL
;
1388 snprintf(buf
, sizeof(buf
), format
, unit_description(u
));
1391 mid
= t
== JOB_START
? SD_MESSAGE_UNIT_STARTING
:
1392 t
== JOB_STOP
? SD_MESSAGE_UNIT_STOPPING
:
1393 SD_MESSAGE_UNIT_RELOADING
;
1395 log_unit_struct(u
->id
,
1397 LOG_MESSAGE_ID(mid
),
1398 LOG_MESSAGE("%s", buf
),
1403 * -EBADR: This unit type does not support starting.
1404 * -EALREADY: Unit is already started.
1405 * -EAGAIN: An operation is already in progress. Retry later.
1406 * -ECANCELED: Too many requests for now.
1407 * -EPROTO: Assert failed
1409 int unit_start(Unit
*u
) {
1410 UnitActiveState state
;
1416 if (u
->load_state
!= UNIT_LOADED
)
1419 /* If this is already started, then this will succeed. Note
1420 * that this will even succeed if this unit is not startable
1421 * by the user. This is relied on to detect when we need to
1422 * wait for units and when waiting is finished. */
1423 state
= unit_active_state(u
);
1424 if (UNIT_IS_ACTIVE_OR_RELOADING(state
))
1427 /* If the conditions failed, don't do anything at all. If we
1428 * already are activating this call might still be useful to
1429 * speed up activation in case there is some hold-off time,
1430 * but we don't want to recheck the condition in that case. */
1431 if (state
!= UNIT_ACTIVATING
&&
1432 !unit_condition_test(u
)) {
1433 log_unit_debug(u
->id
, "Starting of %s requested but condition failed. Not starting unit.", u
->id
);
1437 /* If the asserts failed, fail the entire job */
1438 if (state
!= UNIT_ACTIVATING
&&
1439 !unit_assert_test(u
)) {
1440 log_unit_debug(u
->id
, "Starting of %s requested but asserts failed.", u
->id
);
1444 /* Forward to the main object, if we aren't it. */
1445 following
= unit_following(u
);
1447 log_unit_debug(u
->id
, "Redirecting start request from %s to %s.", u
->id
, following
->id
);
1448 return unit_start(following
);
1451 if (UNIT_VTABLE(u
)->supported
&& !UNIT_VTABLE(u
)->supported(u
->manager
))
1454 /* If it is stopped, but we cannot start it, then fail */
1455 if (!UNIT_VTABLE(u
)->start
)
1458 /* We don't suppress calls to ->start() here when we are
1459 * already starting, to allow this request to be used as a
1460 * "hurry up" call, for example when the unit is in some "auto
1461 * restart" state where it waits for a holdoff timer to elapse
1462 * before it will start again. */
1464 unit_add_to_dbus_queue(u
);
1466 r
= UNIT_VTABLE(u
)->start(u
);
1470 /* Log if the start function actually did something */
1471 unit_status_log_starting_stopping_reloading(u
, JOB_START
);
1472 unit_status_print_starting_stopping(u
, JOB_START
);
1476 bool unit_can_start(Unit
*u
) {
1479 return !!UNIT_VTABLE(u
)->start
;
1482 bool unit_can_isolate(Unit
*u
) {
1485 return unit_can_start(u
) &&
1490 * -EBADR: This unit type does not support stopping.
1491 * -EALREADY: Unit is already stopped.
1492 * -EAGAIN: An operation is already in progress. Retry later.
1494 int unit_stop(Unit
*u
) {
1495 UnitActiveState state
;
1501 state
= unit_active_state(u
);
1502 if (UNIT_IS_INACTIVE_OR_FAILED(state
))
1505 following
= unit_following(u
);
1507 log_unit_debug(u
->id
, "Redirecting stop request from %s to %s.", u
->id
, following
->id
);
1508 return unit_stop(following
);
1511 if (!UNIT_VTABLE(u
)->stop
)
1514 unit_add_to_dbus_queue(u
);
1516 r
= UNIT_VTABLE(u
)->stop(u
);
1520 unit_status_log_starting_stopping_reloading(u
, JOB_STOP
);
1521 unit_status_print_starting_stopping(u
, JOB_STOP
);
1526 * -EBADR: This unit type does not support reloading.
1527 * -ENOEXEC: Unit is not started.
1528 * -EAGAIN: An operation is already in progress. Retry later.
1530 int unit_reload(Unit
*u
) {
1531 UnitActiveState state
;
1537 if (u
->load_state
!= UNIT_LOADED
)
1540 if (!unit_can_reload(u
))
1543 state
= unit_active_state(u
);
1544 if (state
== UNIT_RELOADING
)
1547 if (state
!= UNIT_ACTIVE
) {
1548 log_unit_warning(u
->id
, "Unit %s cannot be reloaded because it is inactive.", u
->id
);
1552 following
= unit_following(u
);
1554 log_unit_debug(u
->id
, "Redirecting reload request from %s to %s.", u
->id
, following
->id
);
1555 return unit_reload(following
);
1558 unit_add_to_dbus_queue(u
);
1560 r
= UNIT_VTABLE(u
)->reload(u
);
1564 unit_status_log_starting_stopping_reloading(u
, JOB_RELOAD
);
1568 bool unit_can_reload(Unit
*u
) {
1571 if (!UNIT_VTABLE(u
)->reload
)
1574 if (!UNIT_VTABLE(u
)->can_reload
)
1577 return UNIT_VTABLE(u
)->can_reload(u
);
1580 static void unit_check_unneeded(Unit
*u
) {
1586 /* If this service shall be shut down when unneeded then do
1589 if (!u
->stop_when_unneeded
)
1592 if (!UNIT_IS_ACTIVE_OR_ACTIVATING(unit_active_state(u
)))
1595 SET_FOREACH(other
, u
->dependencies
[UNIT_REQUIRED_BY
], i
)
1596 if (unit_active_or_pending(other
))
1599 SET_FOREACH(other
, u
->dependencies
[UNIT_REQUIRED_BY_OVERRIDABLE
], i
)
1600 if (unit_active_or_pending(other
))
1603 SET_FOREACH(other
, u
->dependencies
[UNIT_WANTED_BY
], i
)
1604 if (unit_active_or_pending(other
))
1607 SET_FOREACH(other
, u
->dependencies
[UNIT_BOUND_BY
], i
)
1608 if (unit_active_or_pending(other
))
1611 log_unit_info(u
->id
, "Unit %s is not needed anymore. Stopping.", u
->id
);
1613 /* Ok, nobody needs us anymore. Sniff. Then let's commit suicide */
1614 manager_add_job(u
->manager
, JOB_STOP
, u
, JOB_FAIL
, true, NULL
, NULL
);
1617 static void unit_check_binds_to(Unit
*u
) {
1627 if (unit_active_state(u
) != UNIT_ACTIVE
)
1630 SET_FOREACH(other
, u
->dependencies
[UNIT_BINDS_TO
], i
) {
1634 if (!UNIT_IS_INACTIVE_OR_FAILED(unit_active_state(other
)))
1645 log_unit_info(u
->id
, "Unit %s is bound to inactive unit %s. Stopping, too.", u
->id
, other
->id
);
1647 /* A unit we need to run is gone. Sniff. Let's stop this. */
1648 manager_add_job(u
->manager
, JOB_STOP
, u
, JOB_FAIL
, true, NULL
, NULL
);
1651 static void retroactively_start_dependencies(Unit
*u
) {
1656 assert(UNIT_IS_ACTIVE_OR_ACTIVATING(unit_active_state(u
)));
1658 SET_FOREACH(other
, u
->dependencies
[UNIT_REQUIRES
], i
)
1659 if (!set_get(u
->dependencies
[UNIT_AFTER
], other
) &&
1660 !UNIT_IS_ACTIVE_OR_ACTIVATING(unit_active_state(other
)))
1661 manager_add_job(u
->manager
, JOB_START
, other
, JOB_REPLACE
, true, NULL
, NULL
);
1663 SET_FOREACH(other
, u
->dependencies
[UNIT_BINDS_TO
], i
)
1664 if (!set_get(u
->dependencies
[UNIT_AFTER
], other
) &&
1665 !UNIT_IS_ACTIVE_OR_ACTIVATING(unit_active_state(other
)))
1666 manager_add_job(u
->manager
, JOB_START
, other
, JOB_REPLACE
, true, NULL
, NULL
);
1668 SET_FOREACH(other
, u
->dependencies
[UNIT_REQUIRES_OVERRIDABLE
], i
)
1669 if (!set_get(u
->dependencies
[UNIT_AFTER
], other
) &&
1670 !UNIT_IS_ACTIVE_OR_ACTIVATING(unit_active_state(other
)))
1671 manager_add_job(u
->manager
, JOB_START
, other
, JOB_FAIL
, false, NULL
, NULL
);
1673 SET_FOREACH(other
, u
->dependencies
[UNIT_WANTS
], i
)
1674 if (!set_get(u
->dependencies
[UNIT_AFTER
], other
) &&
1675 !UNIT_IS_ACTIVE_OR_ACTIVATING(unit_active_state(other
)))
1676 manager_add_job(u
->manager
, JOB_START
, other
, JOB_FAIL
, false, NULL
, NULL
);
1678 SET_FOREACH(other
, u
->dependencies
[UNIT_CONFLICTS
], i
)
1679 if (!UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(other
)))
1680 manager_add_job(u
->manager
, JOB_STOP
, other
, JOB_REPLACE
, true, NULL
, NULL
);
1682 SET_FOREACH(other
, u
->dependencies
[UNIT_CONFLICTED_BY
], i
)
1683 if (!UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(other
)))
1684 manager_add_job(u
->manager
, JOB_STOP
, other
, JOB_REPLACE
, true, NULL
, NULL
);
1687 static void retroactively_stop_dependencies(Unit
*u
) {
1692 assert(UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(u
)));
1694 /* Pull down units which are bound to us recursively if enabled */
1695 SET_FOREACH(other
, u
->dependencies
[UNIT_BOUND_BY
], i
)
1696 if (!UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(other
)))
1697 manager_add_job(u
->manager
, JOB_STOP
, other
, JOB_REPLACE
, true, NULL
, NULL
);
1700 static void check_unneeded_dependencies(Unit
*u
) {
1705 assert(UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(u
)));
1707 /* Garbage collect services that might not be needed anymore, if enabled */
1708 SET_FOREACH(other
, u
->dependencies
[UNIT_REQUIRES
], i
)
1709 if (!UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(other
)))
1710 unit_check_unneeded(other
);
1711 SET_FOREACH(other
, u
->dependencies
[UNIT_REQUIRES_OVERRIDABLE
], i
)
1712 if (!UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(other
)))
1713 unit_check_unneeded(other
);
1714 SET_FOREACH(other
, u
->dependencies
[UNIT_WANTS
], i
)
1715 if (!UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(other
)))
1716 unit_check_unneeded(other
);
1717 SET_FOREACH(other
, u
->dependencies
[UNIT_REQUISITE
], i
)
1718 if (!UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(other
)))
1719 unit_check_unneeded(other
);
1720 SET_FOREACH(other
, u
->dependencies
[UNIT_REQUISITE_OVERRIDABLE
], i
)
1721 if (!UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(other
)))
1722 unit_check_unneeded(other
);
1723 SET_FOREACH(other
, u
->dependencies
[UNIT_BINDS_TO
], i
)
1724 if (!UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(other
)))
1725 unit_check_unneeded(other
);
1728 void unit_start_on_failure(Unit
*u
) {
1734 if (set_size(u
->dependencies
[UNIT_ON_FAILURE
]) <= 0)
1737 log_unit_info(u
->id
, "Triggering OnFailure= dependencies of %s.", u
->id
);
1739 SET_FOREACH(other
, u
->dependencies
[UNIT_ON_FAILURE
], i
) {
1742 r
= manager_add_job(u
->manager
, JOB_START
, other
, u
->on_failure_job_mode
, true, NULL
, NULL
);
1744 log_unit_error_errno(u
->id
, r
, "Failed to enqueue OnFailure= job: %m");
1748 void unit_trigger_notify(Unit
*u
) {
1754 SET_FOREACH(other
, u
->dependencies
[UNIT_TRIGGERED_BY
], i
)
1755 if (UNIT_VTABLE(other
)->trigger_notify
)
1756 UNIT_VTABLE(other
)->trigger_notify(other
, u
);
1759 void unit_notify(Unit
*u
, UnitActiveState os
, UnitActiveState ns
, bool reload_success
) {
1764 assert(os
< _UNIT_ACTIVE_STATE_MAX
);
1765 assert(ns
< _UNIT_ACTIVE_STATE_MAX
);
1767 /* Note that this is called for all low-level state changes,
1768 * even if they might map to the same high-level
1769 * UnitActiveState! That means that ns == os is an expected
1770 * behavior here. For example: if a mount point is remounted
1771 * this function will be called too! */
1775 /* Update timestamps for state changes */
1776 if (m
->n_reloading
<= 0) {
1779 dual_timestamp_get(&ts
);
1781 if (UNIT_IS_INACTIVE_OR_FAILED(os
) && !UNIT_IS_INACTIVE_OR_FAILED(ns
))
1782 u
->inactive_exit_timestamp
= ts
;
1783 else if (!UNIT_IS_INACTIVE_OR_FAILED(os
) && UNIT_IS_INACTIVE_OR_FAILED(ns
))
1784 u
->inactive_enter_timestamp
= ts
;
1786 if (!UNIT_IS_ACTIVE_OR_RELOADING(os
) && UNIT_IS_ACTIVE_OR_RELOADING(ns
))
1787 u
->active_enter_timestamp
= ts
;
1788 else if (UNIT_IS_ACTIVE_OR_RELOADING(os
) && !UNIT_IS_ACTIVE_OR_RELOADING(ns
))
1789 u
->active_exit_timestamp
= ts
;
1792 /* Keep track of failed units */
1793 manager_update_failed_units(u
->manager
, u
, ns
== UNIT_FAILED
);
1795 /* Make sure the cgroup is always removed when we become inactive */
1796 if (UNIT_IS_INACTIVE_OR_FAILED(ns
))
1797 unit_destroy_cgroup_if_empty(u
);
1799 /* Note that this doesn't apply to RemainAfterExit services exiting
1800 * successfully, since there's no change of state in that case. Which is
1801 * why it is handled in service_set_state() */
1802 if (UNIT_IS_INACTIVE_OR_FAILED(os
) != UNIT_IS_INACTIVE_OR_FAILED(ns
)) {
1805 ec
= unit_get_exec_context(u
);
1806 if (ec
&& exec_context_may_touch_console(ec
)) {
1807 if (UNIT_IS_INACTIVE_OR_FAILED(ns
)) {
1810 if (m
->n_on_console
== 0)
1811 /* unset no_console_output flag, since the console is free */
1812 m
->no_console_output
= false;
1821 if (u
->job
->state
== JOB_WAITING
)
1823 /* So we reached a different state for this
1824 * job. Let's see if we can run it now if it
1825 * failed previously due to EAGAIN. */
1826 job_add_to_run_queue(u
->job
);
1828 /* Let's check whether this state change constitutes a
1829 * finished job, or maybe contradicts a running job and
1830 * hence needs to invalidate jobs. */
1832 switch (u
->job
->type
) {
1835 case JOB_VERIFY_ACTIVE
:
1837 if (UNIT_IS_ACTIVE_OR_RELOADING(ns
))
1838 job_finish_and_invalidate(u
->job
, JOB_DONE
, true);
1839 else if (u
->job
->state
== JOB_RUNNING
&& ns
!= UNIT_ACTIVATING
) {
1842 if (UNIT_IS_INACTIVE_OR_FAILED(ns
))
1843 job_finish_and_invalidate(u
->job
, ns
== UNIT_FAILED
? JOB_FAILED
: JOB_DONE
, true);
1849 case JOB_RELOAD_OR_START
:
1851 if (u
->job
->state
== JOB_RUNNING
) {
1852 if (ns
== UNIT_ACTIVE
)
1853 job_finish_and_invalidate(u
->job
, reload_success
? JOB_DONE
: JOB_FAILED
, true);
1854 else if (ns
!= UNIT_ACTIVATING
&& ns
!= UNIT_RELOADING
) {
1857 if (UNIT_IS_INACTIVE_OR_FAILED(ns
))
1858 job_finish_and_invalidate(u
->job
, ns
== UNIT_FAILED
? JOB_FAILED
: JOB_DONE
, true);
1866 case JOB_TRY_RESTART
:
1868 if (UNIT_IS_INACTIVE_OR_FAILED(ns
))
1869 job_finish_and_invalidate(u
->job
, JOB_DONE
, true);
1870 else if (u
->job
->state
== JOB_RUNNING
&& ns
!= UNIT_DEACTIVATING
) {
1872 job_finish_and_invalidate(u
->job
, JOB_FAILED
, true);
1878 assert_not_reached("Job type unknown");
1884 if (m
->n_reloading
<= 0) {
1886 /* If this state change happened without being
1887 * requested by a job, then let's retroactively start
1888 * or stop dependencies. We skip that step when
1889 * deserializing, since we don't want to create any
1890 * additional jobs just because something is already
1894 if (UNIT_IS_INACTIVE_OR_FAILED(os
) && UNIT_IS_ACTIVE_OR_ACTIVATING(ns
))
1895 retroactively_start_dependencies(u
);
1896 else if (UNIT_IS_ACTIVE_OR_ACTIVATING(os
) && UNIT_IS_INACTIVE_OR_DEACTIVATING(ns
))
1897 retroactively_stop_dependencies(u
);
1900 /* stop unneeded units regardless if going down was expected or not */
1901 if (UNIT_IS_INACTIVE_OR_DEACTIVATING(ns
))
1902 check_unneeded_dependencies(u
);
1904 if (ns
!= os
&& ns
== UNIT_FAILED
) {
1905 log_unit_notice(u
->id
, "Unit %s entered failed state.", u
->id
);
1906 unit_start_on_failure(u
);
1910 /* Some names are special */
1911 if (UNIT_IS_ACTIVE_OR_RELOADING(ns
)) {
1913 if (unit_has_name(u
, SPECIAL_DBUS_SERVICE
))
1914 /* The bus might have just become available,
1915 * hence try to connect to it, if we aren't
1919 if (u
->type
== UNIT_SERVICE
&&
1920 !UNIT_IS_ACTIVE_OR_RELOADING(os
) &&
1921 m
->n_reloading
<= 0) {
1922 /* Write audit record if we have just finished starting up */
1923 manager_send_unit_audit(m
, u
, AUDIT_SERVICE_START
, true);
1927 if (!UNIT_IS_ACTIVE_OR_RELOADING(os
))
1928 manager_send_unit_plymouth(m
, u
);
1932 /* We don't care about D-Bus here, since we'll get an
1933 * asynchronous notification for it anyway. */
1935 if (u
->type
== UNIT_SERVICE
&&
1936 UNIT_IS_INACTIVE_OR_FAILED(ns
) &&
1937 !UNIT_IS_INACTIVE_OR_FAILED(os
) &&
1938 m
->n_reloading
<= 0) {
1940 /* Hmm, if there was no start record written
1941 * write it now, so that we always have a nice
1944 manager_send_unit_audit(m
, u
, AUDIT_SERVICE_START
, ns
== UNIT_INACTIVE
);
1946 if (ns
== UNIT_INACTIVE
)
1947 manager_send_unit_audit(m
, u
, AUDIT_SERVICE_STOP
, true);
1949 /* Write audit record if we have just finished shutting down */
1950 manager_send_unit_audit(m
, u
, AUDIT_SERVICE_STOP
, ns
== UNIT_INACTIVE
);
1952 u
->in_audit
= false;
1956 manager_recheck_journal(m
);
1957 unit_trigger_notify(u
);
1959 if (u
->manager
->n_reloading
<= 0) {
1960 /* Maybe we finished startup and are now ready for
1961 * being stopped because unneeded? */
1962 unit_check_unneeded(u
);
1964 /* Maybe we finished startup, but something we needed
1965 * has vanished? Let's die then. (This happens when
1966 * something BindsTo= to a Type=oneshot unit, as these
1967 * units go directly from starting to inactive,
1968 * without ever entering started.) */
1969 unit_check_binds_to(u
);
1972 unit_add_to_dbus_queue(u
);
1973 unit_add_to_gc_queue(u
);
1976 int unit_watch_pid(Unit
*u
, pid_t pid
) {
1982 /* Watch a specific PID. We only support one or two units
1983 * watching each PID for now, not more. */
1985 r
= set_ensure_allocated(&u
->pids
, NULL
);
1989 r
= hashmap_ensure_allocated(&u
->manager
->watch_pids1
, NULL
);
1993 r
= hashmap_put(u
->manager
->watch_pids1
, LONG_TO_PTR(pid
), u
);
1995 r
= hashmap_ensure_allocated(&u
->manager
->watch_pids2
, NULL
);
1999 r
= hashmap_put(u
->manager
->watch_pids2
, LONG_TO_PTR(pid
), u
);
2002 q
= set_put(u
->pids
, LONG_TO_PTR(pid
));
2009 void unit_unwatch_pid(Unit
*u
, pid_t pid
) {
2013 hashmap_remove_value(u
->manager
->watch_pids1
, LONG_TO_PTR(pid
), u
);
2014 hashmap_remove_value(u
->manager
->watch_pids2
, LONG_TO_PTR(pid
), u
);
2015 set_remove(u
->pids
, LONG_TO_PTR(pid
));
2018 void unit_unwatch_all_pids(Unit
*u
) {
2021 while (!set_isempty(u
->pids
))
2022 unit_unwatch_pid(u
, PTR_TO_LONG(set_first(u
->pids
)));
2028 static int unit_watch_pids_in_path(Unit
*u
, const char *path
) {
2029 _cleanup_closedir_
DIR *d
= NULL
;
2030 _cleanup_fclose_
FILE *f
= NULL
;
2036 /* Adds all PIDs from a specific cgroup path to the set of PIDs we watch. */
2038 r
= cg_enumerate_processes(SYSTEMD_CGROUP_CONTROLLER
, path
, &f
);
2042 while ((r
= cg_read_pid(f
, &pid
)) > 0) {
2043 r
= unit_watch_pid(u
, pid
);
2044 if (r
< 0 && ret
>= 0)
2047 if (r
< 0 && ret
>= 0)
2050 } else if (ret
>= 0)
2053 r
= cg_enumerate_subgroups(SYSTEMD_CGROUP_CONTROLLER
, path
, &d
);
2057 while ((r
= cg_read_subgroup(d
, &fn
)) > 0) {
2058 _cleanup_free_
char *p
= NULL
;
2060 p
= strjoin(path
, "/", fn
, NULL
);
2066 r
= unit_watch_pids_in_path(u
, p
);
2067 if (r
< 0 && ret
>= 0)
2070 if (r
< 0 && ret
>= 0)
2073 } else if (ret
>= 0)
2079 int unit_watch_all_pids(Unit
*u
) {
2082 /* Adds all PIDs from our cgroup to the set of PIDs we watch */
2084 if (!u
->cgroup_path
)
2087 return unit_watch_pids_in_path(u
, u
->cgroup_path
);
2090 void unit_tidy_watch_pids(Unit
*u
, pid_t except1
, pid_t except2
) {
2096 /* Cleans dead PIDs from our list */
2098 SET_FOREACH(e
, u
->pids
, i
) {
2099 pid_t pid
= PTR_TO_LONG(e
);
2101 if (pid
== except1
|| pid
== except2
)
2104 if (!pid_is_unwaited(pid
))
2105 unit_unwatch_pid(u
, pid
);
2109 bool unit_job_is_applicable(Unit
*u
, JobType j
) {
2111 assert(j
>= 0 && j
< _JOB_TYPE_MAX
);
2115 case JOB_VERIFY_ACTIVE
:
2122 case JOB_TRY_RESTART
:
2123 return unit_can_start(u
);
2126 return unit_can_reload(u
);
2128 case JOB_RELOAD_OR_START
:
2129 return unit_can_reload(u
) && unit_can_start(u
);
2132 assert_not_reached("Invalid job type");
2136 static int maybe_warn_about_dependency(const char *id
, const char *other
, UnitDependency dependency
) {
2139 switch (dependency
) {
2141 case UNIT_REQUIRES_OVERRIDABLE
:
2143 case UNIT_REQUISITE
:
2144 case UNIT_REQUISITE_OVERRIDABLE
:
2147 case UNIT_REQUIRED_BY
:
2148 case UNIT_REQUIRED_BY_OVERRIDABLE
:
2149 case UNIT_WANTED_BY
:
2151 case UNIT_CONSISTS_OF
:
2152 case UNIT_REFERENCES
:
2153 case UNIT_REFERENCED_BY
:
2154 case UNIT_PROPAGATES_RELOAD_TO
:
2155 case UNIT_RELOAD_PROPAGATED_FROM
:
2156 case UNIT_JOINS_NAMESPACE_OF
:
2159 case UNIT_CONFLICTS
:
2160 case UNIT_CONFLICTED_BY
:
2163 case UNIT_ON_FAILURE
:
2165 case UNIT_TRIGGERED_BY
:
2166 if (streq_ptr(id
, other
))
2167 log_unit_warning(id
, "Dependency %s=%s dropped from unit %s",
2168 unit_dependency_to_string(dependency
), id
, other
);
2170 log_unit_warning(id
, "Dependency %s=%s dropped from unit %s merged into %s",
2171 unit_dependency_to_string(dependency
), id
,
2175 case _UNIT_DEPENDENCY_MAX
:
2176 case _UNIT_DEPENDENCY_INVALID
:
2180 assert_not_reached("Invalid dependency type");
2183 int unit_add_dependency(Unit
*u
, UnitDependency d
, Unit
*other
, bool add_reference
) {
2185 static const UnitDependency inverse_table
[_UNIT_DEPENDENCY_MAX
] = {
2186 [UNIT_REQUIRES
] = UNIT_REQUIRED_BY
,
2187 [UNIT_REQUIRES_OVERRIDABLE
] = UNIT_REQUIRED_BY_OVERRIDABLE
,
2188 [UNIT_WANTS
] = UNIT_WANTED_BY
,
2189 [UNIT_REQUISITE
] = UNIT_REQUIRED_BY
,
2190 [UNIT_REQUISITE_OVERRIDABLE
] = UNIT_REQUIRED_BY_OVERRIDABLE
,
2191 [UNIT_BINDS_TO
] = UNIT_BOUND_BY
,
2192 [UNIT_PART_OF
] = UNIT_CONSISTS_OF
,
2193 [UNIT_REQUIRED_BY
] = _UNIT_DEPENDENCY_INVALID
,
2194 [UNIT_REQUIRED_BY_OVERRIDABLE
] = _UNIT_DEPENDENCY_INVALID
,
2195 [UNIT_WANTED_BY
] = _UNIT_DEPENDENCY_INVALID
,
2196 [UNIT_BOUND_BY
] = UNIT_BINDS_TO
,
2197 [UNIT_CONSISTS_OF
] = UNIT_PART_OF
,
2198 [UNIT_CONFLICTS
] = UNIT_CONFLICTED_BY
,
2199 [UNIT_CONFLICTED_BY
] = UNIT_CONFLICTS
,
2200 [UNIT_BEFORE
] = UNIT_AFTER
,
2201 [UNIT_AFTER
] = UNIT_BEFORE
,
2202 [UNIT_ON_FAILURE
] = _UNIT_DEPENDENCY_INVALID
,
2203 [UNIT_REFERENCES
] = UNIT_REFERENCED_BY
,
2204 [UNIT_REFERENCED_BY
] = UNIT_REFERENCES
,
2205 [UNIT_TRIGGERS
] = UNIT_TRIGGERED_BY
,
2206 [UNIT_TRIGGERED_BY
] = UNIT_TRIGGERS
,
2207 [UNIT_PROPAGATES_RELOAD_TO
] = UNIT_RELOAD_PROPAGATED_FROM
,
2208 [UNIT_RELOAD_PROPAGATED_FROM
] = UNIT_PROPAGATES_RELOAD_TO
,
2209 [UNIT_JOINS_NAMESPACE_OF
] = UNIT_JOINS_NAMESPACE_OF
,
2211 int r
, q
= 0, v
= 0, w
= 0;
2212 Unit
*orig_u
= u
, *orig_other
= other
;
2215 assert(d
>= 0 && d
< _UNIT_DEPENDENCY_MAX
);
2218 u
= unit_follow_merge(u
);
2219 other
= unit_follow_merge(other
);
2221 /* We won't allow dependencies on ourselves. We will not
2222 * consider them an error however. */
2224 maybe_warn_about_dependency(orig_u
->id
, orig_other
->id
, d
);
2228 r
= set_ensure_allocated(&u
->dependencies
[d
], NULL
);
2232 if (inverse_table
[d
] != _UNIT_DEPENDENCY_INVALID
) {
2233 r
= set_ensure_allocated(&other
->dependencies
[inverse_table
[d
]], NULL
);
2238 if (add_reference
) {
2239 r
= set_ensure_allocated(&u
->dependencies
[UNIT_REFERENCES
], NULL
);
2243 r
= set_ensure_allocated(&other
->dependencies
[UNIT_REFERENCED_BY
], NULL
);
2248 q
= set_put(u
->dependencies
[d
], other
);
2252 if (inverse_table
[d
] != _UNIT_DEPENDENCY_INVALID
&& inverse_table
[d
] != d
) {
2253 v
= set_put(other
->dependencies
[inverse_table
[d
]], u
);
2260 if (add_reference
) {
2261 w
= set_put(u
->dependencies
[UNIT_REFERENCES
], other
);
2267 r
= set_put(other
->dependencies
[UNIT_REFERENCED_BY
], u
);
2272 unit_add_to_dbus_queue(u
);
2277 set_remove(u
->dependencies
[d
], other
);
2280 set_remove(other
->dependencies
[inverse_table
[d
]], u
);
2283 set_remove(u
->dependencies
[UNIT_REFERENCES
], other
);
2288 int unit_add_two_dependencies(Unit
*u
, UnitDependency d
, UnitDependency e
, Unit
*other
, bool add_reference
) {
2293 r
= unit_add_dependency(u
, d
, other
, add_reference
);
2297 r
= unit_add_dependency(u
, e
, other
, add_reference
);
2304 static const char *resolve_template(Unit
*u
, const char *name
, const char*path
, char **p
) {
2308 assert(name
|| path
);
2312 name
= basename(path
);
2314 if (!unit_name_is_template(name
)) {
2320 s
= unit_name_replace_instance(name
, u
->instance
);
2322 _cleanup_free_
char *i
= NULL
;
2324 i
= unit_name_to_prefix(u
->id
);
2328 s
= unit_name_replace_instance(name
, i
);
2338 int unit_add_dependency_by_name(Unit
*u
, UnitDependency d
, const char *name
, const char *path
, bool add_reference
) {
2341 _cleanup_free_
char *s
= NULL
;
2344 assert(name
|| path
);
2346 name
= resolve_template(u
, name
, path
, &s
);
2350 r
= manager_load_unit(u
->manager
, name
, path
, NULL
, &other
);
2354 return unit_add_dependency(u
, d
, other
, add_reference
);
2357 int unit_add_two_dependencies_by_name(Unit
*u
, UnitDependency d
, UnitDependency e
, const char *name
, const char *path
, bool add_reference
) {
2358 _cleanup_free_
char *s
= NULL
;
2363 assert(name
|| path
);
2365 name
= resolve_template(u
, name
, path
, &s
);
2369 r
= manager_load_unit(u
->manager
, name
, path
, NULL
, &other
);
2373 return unit_add_two_dependencies(u
, d
, e
, other
, add_reference
);
2376 int unit_add_dependency_by_name_inverse(Unit
*u
, UnitDependency d
, const char *name
, const char *path
, bool add_reference
) {
2379 _cleanup_free_
char *s
= NULL
;
2382 assert(name
|| path
);
2384 name
= resolve_template(u
, name
, path
, &s
);
2388 r
= manager_load_unit(u
->manager
, name
, path
, NULL
, &other
);
2392 return unit_add_dependency(other
, d
, u
, add_reference
);
2395 int unit_add_two_dependencies_by_name_inverse(Unit
*u
, UnitDependency d
, UnitDependency e
, const char *name
, const char *path
, bool add_reference
) {
2398 _cleanup_free_
char *s
= NULL
;
2401 assert(name
|| path
);
2403 name
= resolve_template(u
, name
, path
, &s
);
2407 r
= manager_load_unit(u
->manager
, name
, path
, NULL
, &other
);
2411 r
= unit_add_two_dependencies(other
, d
, e
, u
, add_reference
);
2418 int set_unit_path(const char *p
) {
2419 /* This is mostly for debug purposes */
2420 if (setenv("SYSTEMD_UNIT_PATH", p
, 0) < 0)
2426 char *unit_dbus_path(Unit
*u
) {
2432 return unit_dbus_path_from_name(u
->id
);
2435 char *unit_default_cgroup_path(Unit
*u
) {
2436 _cleanup_free_
char *escaped
= NULL
, *slice
= NULL
;
2441 if (unit_has_name(u
, SPECIAL_ROOT_SLICE
))
2442 return strdup(u
->manager
->cgroup_root
);
2444 if (UNIT_ISSET(u
->slice
) && !unit_has_name(UNIT_DEREF(u
->slice
), SPECIAL_ROOT_SLICE
)) {
2445 r
= cg_slice_to_path(UNIT_DEREF(u
->slice
)->id
, &slice
);
2450 escaped
= cg_escape(u
->id
);
2455 return strjoin(u
->manager
->cgroup_root
, "/", slice
, "/", escaped
, NULL
);
2457 return strjoin(u
->manager
->cgroup_root
, "/", escaped
, NULL
);
2460 int unit_add_default_slice(Unit
*u
, CGroupContext
*c
) {
2461 _cleanup_free_
char *b
= NULL
;
2462 const char *slice_name
;
2469 if (UNIT_ISSET(u
->slice
))
2473 _cleanup_free_
char *prefix
= NULL
, *escaped
= NULL
;
2475 /* Implicitly place all instantiated units in their
2476 * own per-template slice */
2478 prefix
= unit_name_to_prefix(u
->id
);
2482 /* The prefix is already escaped, but it might include
2483 * "-" which has a special meaning for slice units,
2484 * hence escape it here extra. */
2485 escaped
= strreplace(prefix
, "-", "\\x2d");
2489 if (u
->manager
->running_as
== SYSTEMD_SYSTEM
)
2490 b
= strjoin("system-", escaped
, ".slice", NULL
);
2492 b
= strappend(escaped
, ".slice");
2499 u
->manager
->running_as
== SYSTEMD_SYSTEM
2500 ? SPECIAL_SYSTEM_SLICE
2501 : SPECIAL_ROOT_SLICE
;
2503 r
= manager_load_unit(u
->manager
, slice_name
, NULL
, NULL
, &slice
);
2507 unit_ref_set(&u
->slice
, slice
);
2511 const char *unit_slice_name(Unit
*u
) {
2514 if (!UNIT_ISSET(u
->slice
))
2517 return UNIT_DEREF(u
->slice
)->id
;
2520 int unit_load_related_unit(Unit
*u
, const char *type
, Unit
**_found
) {
2521 _cleanup_free_
char *t
= NULL
;
2528 t
= unit_name_change_suffix(u
->id
, type
);
2532 assert(!unit_has_name(u
, t
));
2534 r
= manager_load_unit(u
->manager
, t
, NULL
, NULL
, _found
);
2535 assert(r
< 0 || *_found
!= u
);
2539 int unit_watch_bus_name(Unit
*u
, const char *name
) {
2543 /* Watch a specific name on the bus. We only support one unit
2544 * watching each name for now. */
2546 return hashmap_put(u
->manager
->watch_bus
, name
, u
);
2549 void unit_unwatch_bus_name(Unit
*u
, const char *name
) {
2553 hashmap_remove_value(u
->manager
->watch_bus
, name
, u
);
2556 bool unit_can_serialize(Unit
*u
) {
2559 return UNIT_VTABLE(u
)->serialize
&& UNIT_VTABLE(u
)->deserialize_item
;
2562 int unit_serialize(Unit
*u
, FILE *f
, FDSet
*fds
, bool serialize_jobs
) {
2569 if (unit_can_serialize(u
)) {
2572 r
= UNIT_VTABLE(u
)->serialize(u
, f
, fds
);
2576 rt
= unit_get_exec_runtime(u
);
2578 r
= exec_runtime_serialize(rt
, u
, f
, fds
);
2584 dual_timestamp_serialize(f
, "inactive-exit-timestamp", &u
->inactive_exit_timestamp
);
2585 dual_timestamp_serialize(f
, "active-enter-timestamp", &u
->active_enter_timestamp
);
2586 dual_timestamp_serialize(f
, "active-exit-timestamp", &u
->active_exit_timestamp
);
2587 dual_timestamp_serialize(f
, "inactive-enter-timestamp", &u
->inactive_enter_timestamp
);
2588 dual_timestamp_serialize(f
, "condition-timestamp", &u
->condition_timestamp
);
2589 dual_timestamp_serialize(f
, "assert-timestamp", &u
->assert_timestamp
);
2591 if (dual_timestamp_is_set(&u
->condition_timestamp
))
2592 unit_serialize_item(u
, f
, "condition-result", yes_no(u
->condition_result
));
2594 if (dual_timestamp_is_set(&u
->assert_timestamp
))
2595 unit_serialize_item(u
, f
, "assert-result", yes_no(u
->assert_result
));
2597 unit_serialize_item(u
, f
, "transient", yes_no(u
->transient
));
2598 unit_serialize_item_format(u
, f
, "cpuacct-usage-base", "%" PRIu64
, u
->cpuacct_usage_base
);
2601 unit_serialize_item(u
, f
, "cgroup", u
->cgroup_path
);
2603 if (serialize_jobs
) {
2605 fprintf(f
, "job\n");
2606 job_serialize(u
->job
, f
, fds
);
2610 fprintf(f
, "job\n");
2611 job_serialize(u
->nop_job
, f
, fds
);
2620 void unit_serialize_item_format(Unit
*u
, FILE *f
, const char *key
, const char *format
, ...) {
2631 va_start(ap
, format
);
2632 vfprintf(f
, format
, ap
);
2638 void unit_serialize_item(Unit
*u
, FILE *f
, const char *key
, const char *value
) {
2644 fprintf(f
, "%s=%s\n", key
, value
);
2647 static int unit_set_cgroup_path(Unit
*u
, const char *path
) {
2648 _cleanup_free_
char *p
= NULL
;
2660 if (streq_ptr(u
->cgroup_path
, p
))
2664 r
= hashmap_put(u
->manager
->cgroup_unit
, p
, u
);
2669 if (u
->cgroup_path
) {
2670 log_unit_debug(u
->id
, "%s: Changing cgroup path from %s to %s.", u
->id
, u
->cgroup_path
, strna(p
));
2671 hashmap_remove(u
->manager
->cgroup_unit
, u
->cgroup_path
);
2672 free(u
->cgroup_path
);
2681 int unit_deserialize(Unit
*u
, FILE *f
, FDSet
*fds
) {
2682 ExecRuntime
**rt
= NULL
;
2690 offset
= UNIT_VTABLE(u
)->exec_runtime_offset
;
2692 rt
= (ExecRuntime
**) ((uint8_t*) u
+ offset
);
2695 char line
[LINE_MAX
], *l
, *v
;
2698 if (!fgets(line
, sizeof(line
), f
)) {
2711 k
= strcspn(l
, "=");
2719 if (streq(l
, "job")) {
2721 /* new-style serialized job */
2728 r
= job_deserialize(j
, f
, fds
);
2734 r
= hashmap_put(u
->manager
->jobs
, UINT32_TO_PTR(j
->id
), j
);
2740 r
= job_install_deserialized(j
);
2742 hashmap_remove(u
->manager
->jobs
, UINT32_TO_PTR(j
->id
));
2750 type
= job_type_from_string(v
);
2752 log_debug("Failed to parse job type value %s", v
);
2754 u
->deserialized_job
= type
;
2757 } else if (streq(l
, "inactive-exit-timestamp")) {
2758 dual_timestamp_deserialize(v
, &u
->inactive_exit_timestamp
);
2760 } else if (streq(l
, "active-enter-timestamp")) {
2761 dual_timestamp_deserialize(v
, &u
->active_enter_timestamp
);
2763 } else if (streq(l
, "active-exit-timestamp")) {
2764 dual_timestamp_deserialize(v
, &u
->active_exit_timestamp
);
2766 } else if (streq(l
, "inactive-enter-timestamp")) {
2767 dual_timestamp_deserialize(v
, &u
->inactive_enter_timestamp
);
2769 } else if (streq(l
, "condition-timestamp")) {
2770 dual_timestamp_deserialize(v
, &u
->condition_timestamp
);
2772 } else if (streq(l
, "assert-timestamp")) {
2773 dual_timestamp_deserialize(v
, &u
->assert_timestamp
);
2775 } else if (streq(l
, "condition-result")) {
2777 r
= parse_boolean(v
);
2779 log_debug("Failed to parse condition result value %s, ignoring.", v
);
2781 u
->condition_result
= r
;
2785 } else if (streq(l
, "assert-result")) {
2787 r
= parse_boolean(v
);
2789 log_debug("Failed to parse assert result value %s, ignoring.", v
);
2791 u
->assert_result
= r
;
2795 } else if (streq(l
, "transient")) {
2797 r
= parse_boolean(v
);
2799 log_debug("Failed to parse transient bool %s, ignoring.", v
);
2805 } else if (streq(l
, "cpuacct-usage-base")) {
2807 r
= safe_atou64(v
, &u
->cpuacct_usage_base
);
2809 log_debug("Failed to parse CPU usage %s, ignoring.", v
);
2813 } else if (streq(l
, "cgroup")) {
2815 r
= unit_set_cgroup_path(u
, v
);
2817 log_debug_errno(r
, "Failed to set cgroup path %s, ignoring: %m", v
);
2822 if (unit_can_serialize(u
)) {
2824 r
= exec_runtime_deserialize_item(rt
, u
, l
, v
, fds
);
2826 log_unit_warning(u
->id
, "Failed to deserialize runtime parameter '%s', ignoring.", l
);
2830 /* Returns positive if key was handled by the call */
2835 r
= UNIT_VTABLE(u
)->deserialize_item(u
, l
, v
, fds
);
2837 log_unit_warning(u
->id
, "Failed to deserialize unit parameter '%s', ignoring.", l
);
2842 int unit_add_node_link(Unit
*u
, const char *what
, bool wants
) {
2844 _cleanup_free_
char *e
= NULL
;
2852 /* Adds in links to the device node that this unit is based on */
2854 if (!is_device_path(what
))
2857 e
= unit_name_from_path(what
, ".device");
2861 r
= manager_load_unit(u
->manager
, e
, NULL
, NULL
, &device
);
2865 r
= unit_add_two_dependencies(u
, UNIT_AFTER
, u
->manager
->running_as
== SYSTEMD_SYSTEM
? UNIT_BINDS_TO
: UNIT_WANTS
, device
, true);
2870 r
= unit_add_dependency(device
, UNIT_WANTS
, u
, false);
2878 static int unit_add_deserialized_job_coldplug(Unit
*u
) {
2881 r
= manager_add_job(u
->manager
, u
->deserialized_job
, u
, JOB_IGNORE_REQUIREMENTS
, false, NULL
, NULL
);
2885 u
->deserialized_job
= _JOB_TYPE_INVALID
;
2890 int unit_coldplug(Unit
*u
, Hashmap
*deferred_work
) {
2895 if (UNIT_VTABLE(u
)->coldplug
)
2896 if ((r
= UNIT_VTABLE(u
)->coldplug(u
, deferred_work
)) < 0)
2900 r
= job_coldplug(u
->job
);
2903 } else if (u
->deserialized_job
>= 0)
2905 hashmap_put(deferred_work
, u
, &unit_add_deserialized_job_coldplug
);
2910 void unit_status_printf(Unit
*u
, const char *status
, const char *unit_status_msg_format
) {
2911 DISABLE_WARNING_FORMAT_NONLITERAL
;
2912 manager_status_printf(u
->manager
, STATUS_TYPE_NORMAL
,
2913 status
, unit_status_msg_format
, unit_description(u
));
2917 bool unit_need_daemon_reload(Unit
*u
) {
2918 _cleanup_strv_free_
char **t
= NULL
;
2921 unsigned loaded_cnt
, current_cnt
;
2925 if (u
->fragment_path
) {
2927 if (stat(u
->fragment_path
, &st
) < 0)
2928 /* What, cannot access this anymore? */
2931 if (u
->fragment_mtime
> 0 &&
2932 timespec_load(&st
.st_mtim
) != u
->fragment_mtime
)
2936 if (u
->source_path
) {
2938 if (stat(u
->source_path
, &st
) < 0)
2941 if (u
->source_mtime
> 0 &&
2942 timespec_load(&st
.st_mtim
) != u
->source_mtime
)
2946 (void) unit_find_dropin_paths(u
, &t
);
2947 loaded_cnt
= strv_length(t
);
2948 current_cnt
= strv_length(u
->dropin_paths
);
2950 if (loaded_cnt
== current_cnt
) {
2951 if (loaded_cnt
== 0)
2954 if (strv_overlap(u
->dropin_paths
, t
)) {
2955 STRV_FOREACH(path
, u
->dropin_paths
) {
2957 if (stat(*path
, &st
) < 0)
2960 if (u
->dropin_mtime
> 0 &&
2961 timespec_load(&st
.st_mtim
) > u
->dropin_mtime
)
2972 void unit_reset_failed(Unit
*u
) {
2975 if (UNIT_VTABLE(u
)->reset_failed
)
2976 UNIT_VTABLE(u
)->reset_failed(u
);
2979 Unit
*unit_following(Unit
*u
) {
2982 if (UNIT_VTABLE(u
)->following
)
2983 return UNIT_VTABLE(u
)->following(u
);
2988 bool unit_stop_pending(Unit
*u
) {
2991 /* This call does check the current state of the unit. It's
2992 * hence useful to be called from state change calls of the
2993 * unit itself, where the state isn't updated yet. This is
2994 * different from unit_inactive_or_pending() which checks both
2995 * the current state and for a queued job. */
2997 return u
->job
&& u
->job
->type
== JOB_STOP
;
3000 bool unit_inactive_or_pending(Unit
*u
) {
3003 /* Returns true if the unit is inactive or going down */
3005 if (UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(u
)))
3008 if (unit_stop_pending(u
))
3014 bool unit_active_or_pending(Unit
*u
) {
3017 /* Returns true if the unit is active or going up */
3019 if (UNIT_IS_ACTIVE_OR_ACTIVATING(unit_active_state(u
)))
3023 (u
->job
->type
== JOB_START
||
3024 u
->job
->type
== JOB_RELOAD_OR_START
||
3025 u
->job
->type
== JOB_RESTART
))
3031 int unit_kill(Unit
*u
, KillWho w
, int signo
, sd_bus_error
*error
) {
3033 assert(w
>= 0 && w
< _KILL_WHO_MAX
);
3035 assert(signo
< _NSIG
);
3037 if (!UNIT_VTABLE(u
)->kill
)
3040 return UNIT_VTABLE(u
)->kill(u
, w
, signo
, error
);
3043 static Set
*unit_pid_set(pid_t main_pid
, pid_t control_pid
) {
3047 pid_set
= set_new(NULL
);
3051 /* Exclude the main/control pids from being killed via the cgroup */
3053 r
= set_put(pid_set
, LONG_TO_PTR(main_pid
));
3058 if (control_pid
> 0) {
3059 r
= set_put(pid_set
, LONG_TO_PTR(control_pid
));
3071 int unit_kill_common(
3077 sd_bus_error
*error
) {
3081 if (who
== KILL_MAIN
&& main_pid
<= 0) {
3083 return sd_bus_error_setf(error
, BUS_ERROR_NO_SUCH_PROCESS
, "%s units have no main processes", unit_type_to_string(u
->type
));
3085 return sd_bus_error_set_const(error
, BUS_ERROR_NO_SUCH_PROCESS
, "No main process to kill");
3088 if (who
== KILL_CONTROL
&& control_pid
<= 0) {
3089 if (control_pid
< 0)
3090 return sd_bus_error_setf(error
, BUS_ERROR_NO_SUCH_PROCESS
, "%s units have no control processes", unit_type_to_string(u
->type
));
3092 return sd_bus_error_set_const(error
, BUS_ERROR_NO_SUCH_PROCESS
, "No control process to kill");
3095 if (who
== KILL_CONTROL
|| who
== KILL_ALL
)
3096 if (control_pid
> 0)
3097 if (kill(control_pid
, signo
) < 0)
3100 if (who
== KILL_MAIN
|| who
== KILL_ALL
)
3102 if (kill(main_pid
, signo
) < 0)
3105 if (who
== KILL_ALL
&& u
->cgroup_path
) {
3106 _cleanup_set_free_ Set
*pid_set
= NULL
;
3109 /* Exclude the main/control pids from being killed via the cgroup */
3110 pid_set
= unit_pid_set(main_pid
, control_pid
);
3114 q
= cg_kill_recursive(SYSTEMD_CGROUP_CONTROLLER
, u
->cgroup_path
, signo
, false, true, false, pid_set
);
3115 if (q
< 0 && q
!= -EAGAIN
&& q
!= -ESRCH
&& q
!= -ENOENT
)
3122 int unit_following_set(Unit
*u
, Set
**s
) {
3126 if (UNIT_VTABLE(u
)->following_set
)
3127 return UNIT_VTABLE(u
)->following_set(u
, s
);
3133 UnitFileState
unit_get_unit_file_state(Unit
*u
) {
3136 if (u
->unit_file_state
< 0 && u
->fragment_path
)
3137 u
->unit_file_state
= unit_file_get_state(
3138 u
->manager
->running_as
== SYSTEMD_SYSTEM
? UNIT_FILE_SYSTEM
: UNIT_FILE_USER
,
3139 NULL
, basename(u
->fragment_path
));
3141 return u
->unit_file_state
;
3144 int unit_get_unit_file_preset(Unit
*u
) {
3147 if (u
->unit_file_preset
< 0 && u
->fragment_path
)
3148 u
->unit_file_preset
= unit_file_query_preset(
3149 u
->manager
->running_as
== SYSTEMD_SYSTEM
? UNIT_FILE_SYSTEM
: UNIT_FILE_USER
,
3150 NULL
, basename(u
->fragment_path
));
3152 return u
->unit_file_preset
;
3155 Unit
* unit_ref_set(UnitRef
*ref
, Unit
*u
) {
3160 unit_ref_unset(ref
);
3163 LIST_PREPEND(refs
, u
->refs
, ref
);
3167 void unit_ref_unset(UnitRef
*ref
) {
3173 LIST_REMOVE(refs
, ref
->unit
->refs
, ref
);
3177 int unit_patch_contexts(Unit
*u
) {
3185 /* Patch in the manager defaults into the exec and cgroup
3186 * contexts, _after_ the rest of the settings have been
3189 ec
= unit_get_exec_context(u
);
3191 /* This only copies in the ones that need memory */
3192 for (i
= 0; i
< _RLIMIT_MAX
; i
++)
3193 if (u
->manager
->rlimit
[i
] && !ec
->rlimit
[i
]) {
3194 ec
->rlimit
[i
] = newdup(struct rlimit
, u
->manager
->rlimit
[i
], 1);
3199 if (u
->manager
->running_as
== SYSTEMD_USER
&&
3200 !ec
->working_directory
) {
3202 r
= get_home_dir(&ec
->working_directory
);
3206 /* Allow user services to run, even if the
3207 * home directory is missing */
3208 ec
->working_directory_missing_ok
= true;
3211 if (u
->manager
->running_as
== SYSTEMD_USER
&&
3212 (ec
->syscall_whitelist
||
3213 !set_isempty(ec
->syscall_filter
) ||
3214 !set_isempty(ec
->syscall_archs
) ||
3215 ec
->address_families_whitelist
||
3216 !set_isempty(ec
->address_families
)))
3217 ec
->no_new_privileges
= true;
3219 if (ec
->private_devices
)
3220 ec
->capability_bounding_set_drop
|= (uint64_t) 1ULL << (uint64_t) CAP_MKNOD
;
3223 cc
= unit_get_cgroup_context(u
);
3227 ec
->private_devices
&&
3228 cc
->device_policy
== CGROUP_AUTO
)
3229 cc
->device_policy
= CGROUP_CLOSED
;
3235 ExecContext
*unit_get_exec_context(Unit
*u
) {
3242 offset
= UNIT_VTABLE(u
)->exec_context_offset
;
3246 return (ExecContext
*) ((uint8_t*) u
+ offset
);
3249 KillContext
*unit_get_kill_context(Unit
*u
) {
3256 offset
= UNIT_VTABLE(u
)->kill_context_offset
;
3260 return (KillContext
*) ((uint8_t*) u
+ offset
);
3263 CGroupContext
*unit_get_cgroup_context(Unit
*u
) {
3269 offset
= UNIT_VTABLE(u
)->cgroup_context_offset
;
3273 return (CGroupContext
*) ((uint8_t*) u
+ offset
);
3276 ExecRuntime
*unit_get_exec_runtime(Unit
*u
) {
3282 offset
= UNIT_VTABLE(u
)->exec_runtime_offset
;
3286 return *(ExecRuntime
**) ((uint8_t*) u
+ offset
);
3289 static int unit_drop_in_dir(Unit
*u
, UnitSetPropertiesMode mode
, bool transient
, char **dir
) {
3290 if (u
->manager
->running_as
== SYSTEMD_USER
) {
3293 if (mode
== UNIT_PERSISTENT
&& !transient
)
3294 r
= user_config_home(dir
);
3296 r
= user_runtime_dir(dir
);
3303 if (mode
== UNIT_PERSISTENT
&& !transient
)
3304 *dir
= strdup("/etc/systemd/system");
3306 *dir
= strdup("/run/systemd/system");
3313 static int unit_drop_in_file(Unit
*u
,
3314 UnitSetPropertiesMode mode
, const char *name
, char **p
, char **q
) {
3315 _cleanup_free_
char *dir
= NULL
;
3320 r
= unit_drop_in_dir(u
, mode
, u
->transient
, &dir
);
3324 return drop_in_file(dir
, u
->id
, 50, name
, p
, q
);
3327 int unit_write_drop_in(Unit
*u
, UnitSetPropertiesMode mode
, const char *name
, const char *data
) {
3329 _cleanup_free_
char *dir
= NULL
, *p
= NULL
, *q
= NULL
;
3334 if (!IN_SET(mode
, UNIT_PERSISTENT
, UNIT_RUNTIME
))
3337 r
= unit_drop_in_dir(u
, mode
, u
->transient
, &dir
);
3341 r
= write_drop_in(dir
, u
->id
, 50, name
, data
);
3345 r
= drop_in_file(dir
, u
->id
, 50, name
, &p
, &q
);
3349 r
= strv_extend(&u
->dropin_paths
, q
);
3353 strv_sort(u
->dropin_paths
);
3354 strv_uniq(u
->dropin_paths
);
3356 u
->dropin_mtime
= now(CLOCK_REALTIME
);
3361 int unit_write_drop_in_format(Unit
*u
, UnitSetPropertiesMode mode
, const char *name
, const char *format
, ...) {
3362 _cleanup_free_
char *p
= NULL
;
3370 if (!IN_SET(mode
, UNIT_PERSISTENT
, UNIT_RUNTIME
))
3373 va_start(ap
, format
);
3374 r
= vasprintf(&p
, format
, ap
);
3380 return unit_write_drop_in(u
, mode
, name
, p
);
3383 int unit_write_drop_in_private(Unit
*u
, UnitSetPropertiesMode mode
, const char *name
, const char *data
) {
3384 _cleanup_free_
char *ndata
= NULL
;
3390 if (!UNIT_VTABLE(u
)->private_section
)
3393 if (!IN_SET(mode
, UNIT_PERSISTENT
, UNIT_RUNTIME
))
3396 ndata
= strjoin("[", UNIT_VTABLE(u
)->private_section
, "]\n", data
, NULL
);
3400 return unit_write_drop_in(u
, mode
, name
, ndata
);
3403 int unit_write_drop_in_private_format(Unit
*u
, UnitSetPropertiesMode mode
, const char *name
, const char *format
, ...) {
3404 _cleanup_free_
char *p
= NULL
;
3412 if (!IN_SET(mode
, UNIT_PERSISTENT
, UNIT_RUNTIME
))
3415 va_start(ap
, format
);
3416 r
= vasprintf(&p
, format
, ap
);
3422 return unit_write_drop_in_private(u
, mode
, name
, p
);
3425 int unit_remove_drop_in(Unit
*u
, UnitSetPropertiesMode mode
, const char *name
) {
3426 _cleanup_free_
char *p
= NULL
, *q
= NULL
;
3431 if (!IN_SET(mode
, UNIT_PERSISTENT
, UNIT_RUNTIME
))
3434 r
= unit_drop_in_file(u
, mode
, name
, &p
, &q
);
3439 r
= errno
== ENOENT
? 0 : -errno
;
3447 int unit_make_transient(Unit
*u
) {
3452 u
->load_state
= UNIT_STUB
;
3454 u
->transient
= true;
3456 free(u
->fragment_path
);
3457 u
->fragment_path
= NULL
;
3459 if (u
->manager
->running_as
== SYSTEMD_USER
) {
3460 _cleanup_free_
char *c
= NULL
;
3462 r
= user_runtime_dir(&c
);
3468 u
->fragment_path
= strjoin(c
, "/", u
->id
, NULL
);
3469 if (!u
->fragment_path
)
3474 u
->fragment_path
= strappend("/run/systemd/system/", u
->id
);
3475 if (!u
->fragment_path
)
3478 mkdir_p("/run/systemd/system", 0755);
3481 return write_string_file_atomic_label(u
->fragment_path
, "# Transient stub");
3484 int unit_kill_context(
3490 bool main_pid_alien
) {
3492 int sig
, wait_for_exit
= false, r
;
3497 if (c
->kill_mode
== KILL_NONE
)
3507 case KILL_TERMINATE
:
3508 sig
= c
->kill_signal
;
3511 assert_not_reached("KillOperation unknown");
3515 r
= kill_and_sigcont(main_pid
, sig
);
3517 if (r
< 0 && r
!= -ESRCH
) {
3518 _cleanup_free_
char *comm
= NULL
;
3519 get_process_comm(main_pid
, &comm
);
3521 log_unit_warning_errno(u
->id
, r
, "Failed to kill main process " PID_FMT
" (%s): %m", main_pid
, strna(comm
));
3523 if (!main_pid_alien
)
3524 wait_for_exit
= true;
3526 if (c
->send_sighup
&& k
!= KILL_KILL
)
3527 kill(main_pid
, SIGHUP
);
3531 if (control_pid
> 0) {
3532 r
= kill_and_sigcont(control_pid
, sig
);
3534 if (r
< 0 && r
!= -ESRCH
) {
3535 _cleanup_free_
char *comm
= NULL
;
3536 get_process_comm(control_pid
, &comm
);
3538 log_unit_warning_errno(u
->id
, r
, "Failed to kill control process " PID_FMT
" (%s): %m", control_pid
, strna(comm
));
3540 wait_for_exit
= true;
3542 if (c
->send_sighup
&& k
!= KILL_KILL
)
3543 kill(control_pid
, SIGHUP
);
3547 if ((c
->kill_mode
== KILL_CONTROL_GROUP
|| (c
->kill_mode
== KILL_MIXED
&& k
== KILL_KILL
)) && u
->cgroup_path
) {
3548 _cleanup_set_free_ Set
*pid_set
= NULL
;
3550 /* Exclude the main/control pids from being killed via the cgroup */
3551 pid_set
= unit_pid_set(main_pid
, control_pid
);
3555 r
= cg_kill_recursive(SYSTEMD_CGROUP_CONTROLLER
, u
->cgroup_path
, sig
, true, true, false, pid_set
);
3557 if (r
!= -EAGAIN
&& r
!= -ESRCH
&& r
!= -ENOENT
)
3558 log_unit_warning_errno(u
->id
, r
, "Failed to kill control group: %m");
3561 /* FIXME: For now, we will not wait for the
3562 * cgroup members to die, simply because
3563 * cgroup notification is unreliable. It
3564 * doesn't work at all in containers, and
3565 * outside of containers it can be confused
3566 * easily by leaving directories in the
3569 /* wait_for_exit = true; */
3571 if (c
->send_sighup
&& k
!= KILL_KILL
) {
3574 pid_set
= unit_pid_set(main_pid
, control_pid
);
3578 cg_kill_recursive(SYSTEMD_CGROUP_CONTROLLER
, u
->cgroup_path
, SIGHUP
, false, true, false, pid_set
);
3583 return wait_for_exit
;
3586 int unit_require_mounts_for(Unit
*u
, const char *path
) {
3587 char prefix
[strlen(path
) + 1], *p
;
3593 /* Registers a unit for requiring a certain path and all its
3594 * prefixes. We keep a simple array of these paths in the
3595 * unit, since its usually short. However, we build a prefix
3596 * table for all possible prefixes so that new appearing mount
3597 * units can easily determine which units to make themselves a
3600 if (!path_is_absolute(path
))
3607 path_kill_slashes(p
);
3609 if (!path_is_safe(p
)) {
3614 if (strv_contains(u
->requires_mounts_for
, p
)) {
3619 r
= strv_consume(&u
->requires_mounts_for
, p
);
3623 PATH_FOREACH_PREFIX_MORE(prefix
, p
) {
3626 x
= hashmap_get(u
->manager
->units_requiring_mounts_for
, prefix
);
3630 if (!u
->manager
->units_requiring_mounts_for
) {
3631 u
->manager
->units_requiring_mounts_for
= hashmap_new(&string_hash_ops
);
3632 if (!u
->manager
->units_requiring_mounts_for
)
3646 r
= hashmap_put(u
->manager
->units_requiring_mounts_for
, q
, x
);
3662 int unit_setup_exec_runtime(Unit
*u
) {
3668 offset
= UNIT_VTABLE(u
)->exec_runtime_offset
;
3671 /* Check if there already is an ExecRuntime for this unit? */
3672 rt
= (ExecRuntime
**) ((uint8_t*) u
+ offset
);
3676 /* Try to get it from somebody else */
3677 SET_FOREACH(other
, u
->dependencies
[UNIT_JOINS_NAMESPACE_OF
], i
) {
3679 *rt
= unit_get_exec_runtime(other
);
3681 exec_runtime_ref(*rt
);
3686 return exec_runtime_make(rt
, unit_get_exec_context(u
), u
->id
);
3689 static const char* const unit_active_state_table
[_UNIT_ACTIVE_STATE_MAX
] = {
3690 [UNIT_ACTIVE
] = "active",
3691 [UNIT_RELOADING
] = "reloading",
3692 [UNIT_INACTIVE
] = "inactive",
3693 [UNIT_FAILED
] = "failed",
3694 [UNIT_ACTIVATING
] = "activating",
3695 [UNIT_DEACTIVATING
] = "deactivating"
3698 DEFINE_STRING_TABLE_LOOKUP(unit_active_state
, UnitActiveState
);