From: Tejun Heo Date: Fri, 13 Mar 2026 19:43:23 +0000 (-1000) Subject: sched_ext: Use schedule_deferred_locked() in schedule_dsq_reenq() X-Git-Url: http://git.ipfire.org/gitweb.cgi?a=commitdiff_plain;h=238eba8c210d02ec2908d9c9143db4edcea3bfa1;p=thirdparty%2Flinux.git sched_ext: Use schedule_deferred_locked() in schedule_dsq_reenq() schedule_dsq_reenq() always uses schedule_deferred() which falls back to irq_work. However, callers like schedule_reenq_local() already hold the target rq lock, and scx_bpf_dsq_reenq() may hold it via the ops callback. Add a locked_rq parameter so schedule_dsq_reenq() can use schedule_deferred_locked() when the target rq is already held. The locked variant can use cheaper paths (balance callbacks, wakeup hooks) instead of always bouncing through irq_work. Signed-off-by: Tejun Heo Reviewed-by: Andrea Righi --- diff --git a/kernel/sched/ext.c b/kernel/sched/ext.c index 0ce205e3e9993..a234e57a45555 100644 --- a/kernel/sched/ext.c +++ b/kernel/sched/ext.c @@ -1218,8 +1218,10 @@ static void schedule_deferred_locked(struct rq *rq) } static void schedule_dsq_reenq(struct scx_sched *sch, struct scx_dispatch_q *dsq, - u64 reenq_flags) + u64 reenq_flags, struct rq *locked_rq) { + struct rq *rq; + /* * Allowing reenqueues doesn't make sense while bypassing. This also * blocks from new reenqueues to be scheduled on dead scheds. @@ -1228,7 +1230,8 @@ static void schedule_dsq_reenq(struct scx_sched *sch, struct scx_dispatch_q *dsq return; if (dsq->id == SCX_DSQ_LOCAL) { - struct rq *rq = container_of(dsq, struct rq, scx.local_dsq); + rq = container_of(dsq, struct rq, scx.local_dsq); + struct scx_sched_pcpu *sch_pcpu = per_cpu_ptr(sch->pcpu, cpu_of(rq)); struct scx_deferred_reenq_local *drl = &sch_pcpu->deferred_reenq_local; @@ -1247,10 +1250,9 @@ static void schedule_dsq_reenq(struct scx_sched *sch, struct scx_dispatch_q *dsq list_move_tail(&drl->node, &rq->scx.deferred_reenq_locals); WRITE_ONCE(drl->flags, drl->flags | reenq_flags); } - - schedule_deferred(rq); } else if (!(dsq->id & SCX_DSQ_FLAG_BUILTIN)) { - struct rq *rq = this_rq(); + rq = this_rq(); + struct scx_dsq_pcpu *dsq_pcpu = per_cpu_ptr(dsq->pcpu, cpu_of(rq)); struct scx_deferred_reenq_user *dru = &dsq_pcpu->deferred_reenq_user; @@ -1269,11 +1271,15 @@ static void schedule_dsq_reenq(struct scx_sched *sch, struct scx_dispatch_q *dsq list_move_tail(&dru->node, &rq->scx.deferred_reenq_users); WRITE_ONCE(dru->flags, dru->flags | reenq_flags); } - - schedule_deferred(rq); } else { scx_error(sch, "DSQ 0x%llx not allowed for reenq", dsq->id); + return; } + + if (rq == locked_rq) + schedule_deferred_locked(rq); + else + schedule_deferred(rq); } static void schedule_reenq_local(struct rq *rq, u64 reenq_flags) @@ -1283,7 +1289,7 @@ static void schedule_reenq_local(struct rq *rq, u64 reenq_flags) if (WARN_ON_ONCE(!root)) return; - schedule_dsq_reenq(root, &rq->scx.local_dsq, reenq_flags); + schedule_dsq_reenq(root, &rq->scx.local_dsq, reenq_flags, rq); } /** @@ -8845,7 +8851,7 @@ __bpf_kfunc void scx_bpf_dsq_reenq(u64 dsq_id, u64 reenq_flags, reenq_flags |= SCX_REENQ_ANY; dsq = find_dsq_for_dispatch(sch, this_rq(), dsq_id, smp_processor_id()); - schedule_dsq_reenq(sch, dsq, reenq_flags); + schedule_dsq_reenq(sch, dsq, reenq_flags, scx_locked_rq()); } /**