]> git.ipfire.org Git - thirdparty/kernel/stable-queue.git/commitdiff
drop some rcu patches from 5.4 and 5.10
authorGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Tue, 2 Mar 2021 12:32:55 +0000 (13:32 +0100)
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Tue, 2 Mar 2021 12:32:55 +0000 (13:32 +0100)
queue-5.10/entry-explicitly-flush-pending-rcuog-wakeup-before-last-rescheduling-point.patch [deleted file]
queue-5.10/entry-kvm-explicitly-flush-pending-rcuog-wakeup-before-last-rescheduling-point.patch [deleted file]
queue-5.10/rcu-nocb-trigger-self-ipi-on-late-deferred-wake-up-before-user-resume.patch [deleted file]
queue-5.10/series
queue-5.4/rcu-nocb-trigger-self-ipi-on-late-deferred-wake-up-before-user-resume.patch [deleted file]
queue-5.4/series

diff --git a/queue-5.10/entry-explicitly-flush-pending-rcuog-wakeup-before-last-rescheduling-point.patch b/queue-5.10/entry-explicitly-flush-pending-rcuog-wakeup-before-last-rescheduling-point.patch
deleted file mode 100644 (file)
index 6862e19..0000000
+++ /dev/null
@@ -1,71 +0,0 @@
-From 47b8ff194c1fd73d58dc339b597d466fe48c8958 Mon Sep 17 00:00:00 2001
-From: Frederic Weisbecker <frederic@kernel.org>
-Date: Mon, 1 Feb 2021 00:05:47 +0100
-Subject: entry: Explicitly flush pending rcuog wakeup before last rescheduling point
-
-From: Frederic Weisbecker <frederic@kernel.org>
-
-commit 47b8ff194c1fd73d58dc339b597d466fe48c8958 upstream.
-
-Following the idle loop model, cleanly check for pending rcuog wakeup
-before the last rescheduling point on resuming to user mode. This
-way we can avoid to do it from rcu_user_enter() with the last resort
-self-IPI hack that enforces rescheduling.
-
-Signed-off-by: Frederic Weisbecker <frederic@kernel.org>
-Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
-Signed-off-by: Ingo Molnar <mingo@kernel.org>
-Cc: stable@vger.kernel.org
-Link: https://lkml.kernel.org/r/20210131230548.32970-5-frederic@kernel.org
-Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
----
- kernel/entry/common.c |    7 +++++++
- kernel/rcu/tree.c     |   12 +++++++-----
- 2 files changed, 14 insertions(+), 5 deletions(-)
-
---- a/kernel/entry/common.c
-+++ b/kernel/entry/common.c
-@@ -174,6 +174,10 @@ static unsigned long exit_to_user_mode_l
-                * enabled above.
-                */
-               local_irq_disable_exit_to_user();
-+
-+              /* Check if any of the above work has queued a deferred wakeup */
-+              rcu_nocb_flush_deferred_wakeup();
-+
-               ti_work = READ_ONCE(current_thread_info()->flags);
-       }
-@@ -187,6 +191,9 @@ static void exit_to_user_mode_prepare(st
-       lockdep_assert_irqs_disabled();
-+      /* Flush pending rcuog wakeup before the last need_resched() check */
-+      rcu_nocb_flush_deferred_wakeup();
-+
-       if (unlikely(ti_work & EXIT_TO_USER_MODE_WORK))
-               ti_work = exit_to_user_mode_loop(regs, ti_work);
---- a/kernel/rcu/tree.c
-+++ b/kernel/rcu/tree.c
-@@ -699,13 +699,15 @@ noinstr void rcu_user_enter(void)
-       lockdep_assert_irqs_disabled();
-       /*
--       * We may be past the last rescheduling opportunity in the entry code.
--       * Trigger a self IPI that will fire and reschedule once we resume to
--       * user/guest mode.
-+       * Other than generic entry implementation, we may be past the last
-+       * rescheduling opportunity in the entry code. Trigger a self IPI
-+       * that will fire and reschedule once we resume in user/guest mode.
-        */
-       instrumentation_begin();
--      if (do_nocb_deferred_wakeup(rdp) && need_resched())
--              irq_work_queue(this_cpu_ptr(&late_wakeup_work));
-+      if (!IS_ENABLED(CONFIG_GENERIC_ENTRY) || (current->flags & PF_VCPU)) {
-+              if (do_nocb_deferred_wakeup(rdp) && need_resched())
-+                      irq_work_queue(this_cpu_ptr(&late_wakeup_work));
-+      }
-       instrumentation_end();
-       rcu_eqs_enter(true);
diff --git a/queue-5.10/entry-kvm-explicitly-flush-pending-rcuog-wakeup-before-last-rescheduling-point.patch b/queue-5.10/entry-kvm-explicitly-flush-pending-rcuog-wakeup-before-last-rescheduling-point.patch
deleted file mode 100644 (file)
index 2deff61..0000000
+++ /dev/null
@@ -1,147 +0,0 @@
-From 4ae7dc97f726ea95c58ac58af71cc034ad22d7de Mon Sep 17 00:00:00 2001
-From: Frederic Weisbecker <frederic@kernel.org>
-Date: Mon, 1 Feb 2021 00:05:48 +0100
-Subject: entry/kvm: Explicitly flush pending rcuog wakeup before last rescheduling point
-
-From: Frederic Weisbecker <frederic@kernel.org>
-
-commit 4ae7dc97f726ea95c58ac58af71cc034ad22d7de upstream.
-
-Following the idle loop model, cleanly check for pending rcuog wakeup
-before the last rescheduling point upon resuming to guest mode. This
-way we can avoid to do it from rcu_user_enter() with the last resort
-self-IPI hack that enforces rescheduling.
-
-Suggested-by: Peter Zijlstra <peterz@infradead.org>
-Signed-off-by: Frederic Weisbecker <frederic@kernel.org>
-Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
-Signed-off-by: Ingo Molnar <mingo@kernel.org>
-Cc: stable@vger.kernel.org
-Link: https://lkml.kernel.org/r/20210131230548.32970-6-frederic@kernel.org
-Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
----
- arch/x86/kvm/x86.c        |    1 +
- include/linux/entry-kvm.h |   14 ++++++++++++++
- kernel/rcu/tree.c         |   44 ++++++++++++++++++++++++++++++++++----------
- kernel/rcu/tree_plugin.h  |    1 +
- 4 files changed, 50 insertions(+), 10 deletions(-)
-
---- a/arch/x86/kvm/x86.c
-+++ b/arch/x86/kvm/x86.c
-@@ -1776,6 +1776,7 @@ EXPORT_SYMBOL_GPL(kvm_emulate_wrmsr);
- bool kvm_vcpu_exit_request(struct kvm_vcpu *vcpu)
- {
-+      xfer_to_guest_mode_prepare();
-       return vcpu->mode == EXITING_GUEST_MODE || kvm_request_pending(vcpu) ||
-               xfer_to_guest_mode_work_pending();
- }
---- a/include/linux/entry-kvm.h
-+++ b/include/linux/entry-kvm.h
-@@ -47,6 +47,20 @@ static inline int arch_xfer_to_guest_mod
- int xfer_to_guest_mode_handle_work(struct kvm_vcpu *vcpu);
- /**
-+ * xfer_to_guest_mode_prepare - Perform last minute preparation work that
-+ *                            need to be handled while IRQs are disabled
-+ *                            upon entering to guest.
-+ *
-+ * Has to be invoked with interrupts disabled before the last call
-+ * to xfer_to_guest_mode_work_pending().
-+ */
-+static inline void xfer_to_guest_mode_prepare(void)
-+{
-+      lockdep_assert_irqs_disabled();
-+      rcu_nocb_flush_deferred_wakeup();
-+}
-+
-+/**
-  * __xfer_to_guest_mode_work_pending - Check if work is pending
-  *
-  * Returns: True if work pending, False otherwise.
---- a/kernel/rcu/tree.c
-+++ b/kernel/rcu/tree.c
-@@ -670,9 +670,10 @@ EXPORT_SYMBOL_GPL(rcu_idle_enter);
- #ifdef CONFIG_NO_HZ_FULL
-+#if !defined(CONFIG_GENERIC_ENTRY) || !defined(CONFIG_KVM_XFER_TO_GUEST_WORK)
- /*
-  * An empty function that will trigger a reschedule on
-- * IRQ tail once IRQs get re-enabled on userspace resume.
-+ * IRQ tail once IRQs get re-enabled on userspace/guest resume.
-  */
- static void late_wakeup_func(struct irq_work *work)
- {
-@@ -681,6 +682,37 @@ static void late_wakeup_func(struct irq_
- static DEFINE_PER_CPU(struct irq_work, late_wakeup_work) =
-       IRQ_WORK_INIT(late_wakeup_func);
-+/*
-+ * If either:
-+ *
-+ * 1) the task is about to enter in guest mode and $ARCH doesn't support KVM generic work
-+ * 2) the task is about to enter in user mode and $ARCH doesn't support generic entry.
-+ *
-+ * In these cases the late RCU wake ups aren't supported in the resched loops and our
-+ * last resort is to fire a local irq_work that will trigger a reschedule once IRQs
-+ * get re-enabled again.
-+ */
-+noinstr static void rcu_irq_work_resched(void)
-+{
-+      struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
-+
-+      if (IS_ENABLED(CONFIG_GENERIC_ENTRY) && !(current->flags & PF_VCPU))
-+              return;
-+
-+      if (IS_ENABLED(CONFIG_KVM_XFER_TO_GUEST_WORK) && (current->flags & PF_VCPU))
-+              return;
-+
-+      instrumentation_begin();
-+      if (do_nocb_deferred_wakeup(rdp) && need_resched()) {
-+              irq_work_queue(this_cpu_ptr(&late_wakeup_work));
-+      }
-+      instrumentation_end();
-+}
-+
-+#else
-+static inline void rcu_irq_work_resched(void) { }
-+#endif
-+
- /**
-  * rcu_user_enter - inform RCU that we are resuming userspace.
-  *
-@@ -694,8 +726,6 @@ static DEFINE_PER_CPU(struct irq_work, l
-  */
- noinstr void rcu_user_enter(void)
- {
--      struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
--
-       lockdep_assert_irqs_disabled();
-       /*
-@@ -703,13 +733,7 @@ noinstr void rcu_user_enter(void)
-        * rescheduling opportunity in the entry code. Trigger a self IPI
-        * that will fire and reschedule once we resume in user/guest mode.
-        */
--      instrumentation_begin();
--      if (!IS_ENABLED(CONFIG_GENERIC_ENTRY) || (current->flags & PF_VCPU)) {
--              if (do_nocb_deferred_wakeup(rdp) && need_resched())
--                      irq_work_queue(this_cpu_ptr(&late_wakeup_work));
--      }
--      instrumentation_end();
--
-+      rcu_irq_work_resched();
-       rcu_eqs_enter(true);
- }
---- a/kernel/rcu/tree_plugin.h
-+++ b/kernel/rcu/tree_plugin.h
-@@ -2197,6 +2197,7 @@ void rcu_nocb_flush_deferred_wakeup(void
- {
-       do_nocb_deferred_wakeup(this_cpu_ptr(&rcu_data));
- }
-+EXPORT_SYMBOL_GPL(rcu_nocb_flush_deferred_wakeup);
- void __init rcu_init_nohz(void)
- {
diff --git a/queue-5.10/rcu-nocb-trigger-self-ipi-on-late-deferred-wake-up-before-user-resume.patch b/queue-5.10/rcu-nocb-trigger-self-ipi-on-late-deferred-wake-up-before-user-resume.patch
deleted file mode 100644 (file)
index bcb71f9..0000000
+++ /dev/null
@@ -1,174 +0,0 @@
-From f8bb5cae9616224a39cbb399de382d36ac41df10 Mon Sep 17 00:00:00 2001
-From: Frederic Weisbecker <frederic@kernel.org>
-Date: Mon, 1 Feb 2021 00:05:46 +0100
-Subject: rcu/nocb: Trigger self-IPI on late deferred wake up before user resume
-
-From: Frederic Weisbecker <frederic@kernel.org>
-
-commit f8bb5cae9616224a39cbb399de382d36ac41df10 upstream.
-
-Entering RCU idle mode may cause a deferred wake up of an RCU NOCB_GP
-kthread (rcuog) to be serviced.
-
-Unfortunately the call to rcu_user_enter() is already past the last
-rescheduling opportunity before we resume to userspace or to guest mode.
-We may escape there with the woken task ignored.
-
-The ultimate resort to fix every callsites is to trigger a self-IPI
-(nohz_full depends on arch to implement arch_irq_work_raise()) that will
-trigger a reschedule on IRQ tail or guest exit.
-
-Eventually every site that want a saner treatment will need to carefully
-place a call to rcu_nocb_flush_deferred_wakeup() before the last explicit
-need_resched() check upon resume.
-
-Fixes: 96d3fd0d315a (rcu: Break call_rcu() deadlock involving scheduler and perf)
-Reported-by: Paul E. McKenney <paulmck@kernel.org>
-Signed-off-by: Frederic Weisbecker <frederic@kernel.org>
-Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
-Signed-off-by: Ingo Molnar <mingo@kernel.org>
-Cc: stable@vger.kernel.org
-Link: https://lkml.kernel.org/r/20210131230548.32970-4-frederic@kernel.org
-Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
----
- kernel/rcu/tree.c        |   21 ++++++++++++++++++++-
- kernel/rcu/tree.h        |    2 +-
- kernel/rcu/tree_plugin.h |   25 ++++++++++++++++---------
- 3 files changed, 37 insertions(+), 11 deletions(-)
-
---- a/kernel/rcu/tree.c
-+++ b/kernel/rcu/tree.c
-@@ -669,6 +669,18 @@ void rcu_idle_enter(void)
- EXPORT_SYMBOL_GPL(rcu_idle_enter);
- #ifdef CONFIG_NO_HZ_FULL
-+
-+/*
-+ * An empty function that will trigger a reschedule on
-+ * IRQ tail once IRQs get re-enabled on userspace resume.
-+ */
-+static void late_wakeup_func(struct irq_work *work)
-+{
-+}
-+
-+static DEFINE_PER_CPU(struct irq_work, late_wakeup_work) =
-+      IRQ_WORK_INIT(late_wakeup_func);
-+
- /**
-  * rcu_user_enter - inform RCU that we are resuming userspace.
-  *
-@@ -686,12 +698,19 @@ noinstr void rcu_user_enter(void)
-       lockdep_assert_irqs_disabled();
-+      /*
-+       * We may be past the last rescheduling opportunity in the entry code.
-+       * Trigger a self IPI that will fire and reschedule once we resume to
-+       * user/guest mode.
-+       */
-       instrumentation_begin();
--      do_nocb_deferred_wakeup(rdp);
-+      if (do_nocb_deferred_wakeup(rdp) && need_resched())
-+              irq_work_queue(this_cpu_ptr(&late_wakeup_work));
-       instrumentation_end();
-       rcu_eqs_enter(true);
- }
-+
- #endif /* CONFIG_NO_HZ_FULL */
- /**
---- a/kernel/rcu/tree.h
-+++ b/kernel/rcu/tree.h
-@@ -431,7 +431,7 @@ static bool rcu_nocb_try_bypass(struct r
- static void __call_rcu_nocb_wake(struct rcu_data *rdp, bool was_empty,
-                                unsigned long flags);
- static int rcu_nocb_need_deferred_wakeup(struct rcu_data *rdp);
--static void do_nocb_deferred_wakeup(struct rcu_data *rdp);
-+static bool do_nocb_deferred_wakeup(struct rcu_data *rdp);
- static void rcu_boot_init_nocb_percpu_data(struct rcu_data *rdp);
- static void rcu_spawn_cpu_nocb_kthread(int cpu);
- static void __init rcu_spawn_nocb_kthreads(void);
---- a/kernel/rcu/tree_plugin.h
-+++ b/kernel/rcu/tree_plugin.h
-@@ -1631,8 +1631,8 @@ bool rcu_is_nocb_cpu(int cpu)
-  * Kick the GP kthread for this NOCB group.  Caller holds ->nocb_lock
-  * and this function releases it.
-  */
--static void wake_nocb_gp(struct rcu_data *rdp, bool force,
--                         unsigned long flags)
-+static bool wake_nocb_gp(struct rcu_data *rdp, bool force,
-+                       unsigned long flags)
-       __releases(rdp->nocb_lock)
- {
-       bool needwake = false;
-@@ -1643,7 +1643,7 @@ static void wake_nocb_gp(struct rcu_data
-               trace_rcu_nocb_wake(rcu_state.name, rdp->cpu,
-                                   TPS("AlreadyAwake"));
-               rcu_nocb_unlock_irqrestore(rdp, flags);
--              return;
-+              return false;
-       }
-       del_timer(&rdp->nocb_timer);
-       rcu_nocb_unlock_irqrestore(rdp, flags);
-@@ -1656,6 +1656,8 @@ static void wake_nocb_gp(struct rcu_data
-       raw_spin_unlock_irqrestore(&rdp_gp->nocb_gp_lock, flags);
-       if (needwake)
-               wake_up_process(rdp_gp->nocb_gp_kthread);
-+
-+      return needwake;
- }
- /*
-@@ -2152,20 +2154,23 @@ static int rcu_nocb_need_deferred_wakeup
- }
- /* Do a deferred wakeup of rcu_nocb_kthread(). */
--static void do_nocb_deferred_wakeup_common(struct rcu_data *rdp)
-+static bool do_nocb_deferred_wakeup_common(struct rcu_data *rdp)
- {
-       unsigned long flags;
-       int ndw;
-+      int ret;
-       rcu_nocb_lock_irqsave(rdp, flags);
-       if (!rcu_nocb_need_deferred_wakeup(rdp)) {
-               rcu_nocb_unlock_irqrestore(rdp, flags);
--              return;
-+              return false;
-       }
-       ndw = READ_ONCE(rdp->nocb_defer_wakeup);
-       WRITE_ONCE(rdp->nocb_defer_wakeup, RCU_NOCB_WAKE_NOT);
--      wake_nocb_gp(rdp, ndw == RCU_NOCB_WAKE_FORCE, flags);
-+      ret = wake_nocb_gp(rdp, ndw == RCU_NOCB_WAKE_FORCE, flags);
-       trace_rcu_nocb_wake(rcu_state.name, rdp->cpu, TPS("DeferredWake"));
-+
-+      return ret;
- }
- /* Do a deferred wakeup of rcu_nocb_kthread() from a timer handler. */
-@@ -2181,10 +2186,11 @@ static void do_nocb_deferred_wakeup_time
-  * This means we do an inexact common-case check.  Note that if
-  * we miss, ->nocb_timer will eventually clean things up.
-  */
--static void do_nocb_deferred_wakeup(struct rcu_data *rdp)
-+static bool do_nocb_deferred_wakeup(struct rcu_data *rdp)
- {
-       if (rcu_nocb_need_deferred_wakeup(rdp))
--              do_nocb_deferred_wakeup_common(rdp);
-+              return do_nocb_deferred_wakeup_common(rdp);
-+      return false;
- }
- void rcu_nocb_flush_deferred_wakeup(void)
-@@ -2523,8 +2529,9 @@ static int rcu_nocb_need_deferred_wakeup
-       return false;
- }
--static void do_nocb_deferred_wakeup(struct rcu_data *rdp)
-+static bool do_nocb_deferred_wakeup(struct rcu_data *rdp)
- {
-+      return false;
- }
- static void rcu_spawn_cpu_nocb_kthread(int cpu)
index f05341d7447b9fc6adc65d8d8ddb934cf92996e6..fccbbf86b88ef148af7b7117c58a905ef5a33a3a 100644 (file)
@@ -569,9 +569,6 @@ x86-entry-fix-instrumentation-annotation.patch
 powerpc-prom-fix-ibm-arch-vec-5-platform-support-scan.patch
 rcu-pull-deferred-rcuog-wake-up-to-rcu_eqs_enter-callers.patch
 rcu-nocb-perform-deferred-wake-up-before-last-idle-s-need_resched-check.patch
-rcu-nocb-trigger-self-ipi-on-late-deferred-wake-up-before-user-resume.patch
-entry-explicitly-flush-pending-rcuog-wakeup-before-last-rescheduling-point.patch
-entry-kvm-explicitly-flush-pending-rcuog-wakeup-before-last-rescheduling-point.patch
 kprobes-fix-to-delay-the-kprobes-jump-optimization.patch
 arm64-extend-workaround-for-erratum-1024718-to-all-versions-of-cortex-a55.patch
 iommu-arm-smmu-qcom-fix-mask-extraction-for-bootloader-programmed-smrs.patch
diff --git a/queue-5.4/rcu-nocb-trigger-self-ipi-on-late-deferred-wake-up-before-user-resume.patch b/queue-5.4/rcu-nocb-trigger-self-ipi-on-late-deferred-wake-up-before-user-resume.patch
deleted file mode 100644 (file)
index 1f86662..0000000
+++ /dev/null
@@ -1,174 +0,0 @@
-From f8bb5cae9616224a39cbb399de382d36ac41df10 Mon Sep 17 00:00:00 2001
-From: Frederic Weisbecker <frederic@kernel.org>
-Date: Mon, 1 Feb 2021 00:05:46 +0100
-Subject: rcu/nocb: Trigger self-IPI on late deferred wake up before user resume
-
-From: Frederic Weisbecker <frederic@kernel.org>
-
-commit f8bb5cae9616224a39cbb399de382d36ac41df10 upstream.
-
-Entering RCU idle mode may cause a deferred wake up of an RCU NOCB_GP
-kthread (rcuog) to be serviced.
-
-Unfortunately the call to rcu_user_enter() is already past the last
-rescheduling opportunity before we resume to userspace or to guest mode.
-We may escape there with the woken task ignored.
-
-The ultimate resort to fix every callsites is to trigger a self-IPI
-(nohz_full depends on arch to implement arch_irq_work_raise()) that will
-trigger a reschedule on IRQ tail or guest exit.
-
-Eventually every site that want a saner treatment will need to carefully
-place a call to rcu_nocb_flush_deferred_wakeup() before the last explicit
-need_resched() check upon resume.
-
-Fixes: 96d3fd0d315a (rcu: Break call_rcu() deadlock involving scheduler and perf)
-Reported-by: Paul E. McKenney <paulmck@kernel.org>
-Signed-off-by: Frederic Weisbecker <frederic@kernel.org>
-Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
-Signed-off-by: Ingo Molnar <mingo@kernel.org>
-Cc: stable@vger.kernel.org
-Link: https://lkml.kernel.org/r/20210131230548.32970-4-frederic@kernel.org
-Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
----
- kernel/rcu/tree.c        |   21 ++++++++++++++++++++-
- kernel/rcu/tree.h        |    2 +-
- kernel/rcu/tree_plugin.h |   25 ++++++++++++++++---------
- 3 files changed, 37 insertions(+), 11 deletions(-)
-
---- a/kernel/rcu/tree.c
-+++ b/kernel/rcu/tree.c
-@@ -604,6 +604,18 @@ void rcu_idle_enter(void)
- }
- #ifdef CONFIG_NO_HZ_FULL
-+
-+/*
-+ * An empty function that will trigger a reschedule on
-+ * IRQ tail once IRQs get re-enabled on userspace resume.
-+ */
-+static void late_wakeup_func(struct irq_work *work)
-+{
-+}
-+
-+static DEFINE_PER_CPU(struct irq_work, late_wakeup_work) =
-+      IRQ_WORK_INIT(late_wakeup_func);
-+
- /**
-  * rcu_user_enter - inform RCU that we are resuming userspace.
-  *
-@@ -621,12 +633,19 @@ void rcu_user_enter(void)
-       lockdep_assert_irqs_disabled();
-+      /*
-+       * We may be past the last rescheduling opportunity in the entry code.
-+       * Trigger a self IPI that will fire and reschedule once we resume to
-+       * user/guest mode.
-+       */
-       instrumentation_begin();
--      do_nocb_deferred_wakeup(rdp);
-+      if (do_nocb_deferred_wakeup(rdp) && need_resched())
-+              irq_work_queue(this_cpu_ptr(&late_wakeup_work));
-       instrumentation_end();
-       rcu_eqs_enter(true);
- }
-+
- #endif /* CONFIG_NO_HZ_FULL */
- /*
---- a/kernel/rcu/tree.h
-+++ b/kernel/rcu/tree.h
-@@ -438,7 +438,7 @@ static bool rcu_nocb_try_bypass(struct r
- static void __call_rcu_nocb_wake(struct rcu_data *rdp, bool was_empty,
-                                unsigned long flags);
- static int rcu_nocb_need_deferred_wakeup(struct rcu_data *rdp);
--static void do_nocb_deferred_wakeup(struct rcu_data *rdp);
-+static bool do_nocb_deferred_wakeup(struct rcu_data *rdp);
- static void rcu_boot_init_nocb_percpu_data(struct rcu_data *rdp);
- static void rcu_spawn_cpu_nocb_kthread(int cpu);
- static void __init rcu_spawn_nocb_kthreads(void);
---- a/kernel/rcu/tree_plugin.h
-+++ b/kernel/rcu/tree_plugin.h
-@@ -1639,8 +1639,8 @@ bool rcu_is_nocb_cpu(int cpu)
-  * Kick the GP kthread for this NOCB group.  Caller holds ->nocb_lock
-  * and this function releases it.
-  */
--static void wake_nocb_gp(struct rcu_data *rdp, bool force,
--                         unsigned long flags)
-+static bool wake_nocb_gp(struct rcu_data *rdp, bool force,
-+                       unsigned long flags)
-       __releases(rdp->nocb_lock)
- {
-       bool needwake = false;
-@@ -1651,7 +1651,7 @@ static void wake_nocb_gp(struct rcu_data
-               trace_rcu_nocb_wake(rcu_state.name, rdp->cpu,
-                                   TPS("AlreadyAwake"));
-               rcu_nocb_unlock_irqrestore(rdp, flags);
--              return;
-+              return false;
-       }
-       del_timer(&rdp->nocb_timer);
-       rcu_nocb_unlock_irqrestore(rdp, flags);
-@@ -1664,6 +1664,8 @@ static void wake_nocb_gp(struct rcu_data
-       raw_spin_unlock_irqrestore(&rdp_gp->nocb_gp_lock, flags);
-       if (needwake)
-               wake_up_process(rdp_gp->nocb_gp_kthread);
-+
-+      return needwake;
- }
- /*
-@@ -2155,20 +2157,23 @@ static int rcu_nocb_need_deferred_wakeup
- }
- /* Do a deferred wakeup of rcu_nocb_kthread(). */
--static void do_nocb_deferred_wakeup_common(struct rcu_data *rdp)
-+static bool do_nocb_deferred_wakeup_common(struct rcu_data *rdp)
- {
-       unsigned long flags;
-       int ndw;
-+      int ret;
-       rcu_nocb_lock_irqsave(rdp, flags);
-       if (!rcu_nocb_need_deferred_wakeup(rdp)) {
-               rcu_nocb_unlock_irqrestore(rdp, flags);
--              return;
-+              return false;
-       }
-       ndw = READ_ONCE(rdp->nocb_defer_wakeup);
-       WRITE_ONCE(rdp->nocb_defer_wakeup, RCU_NOCB_WAKE_NOT);
--      wake_nocb_gp(rdp, ndw == RCU_NOCB_WAKE_FORCE, flags);
-+      ret = wake_nocb_gp(rdp, ndw == RCU_NOCB_WAKE_FORCE, flags);
-       trace_rcu_nocb_wake(rcu_state.name, rdp->cpu, TPS("DeferredWake"));
-+
-+      return ret;
- }
- /* Do a deferred wakeup of rcu_nocb_kthread() from a timer handler. */
-@@ -2184,10 +2189,11 @@ static void do_nocb_deferred_wakeup_time
-  * This means we do an inexact common-case check.  Note that if
-  * we miss, ->nocb_timer will eventually clean things up.
-  */
--static void do_nocb_deferred_wakeup(struct rcu_data *rdp)
-+static bool do_nocb_deferred_wakeup(struct rcu_data *rdp)
- {
-       if (rcu_nocb_need_deferred_wakeup(rdp))
--              do_nocb_deferred_wakeup_common(rdp);
-+              return do_nocb_deferred_wakeup_common(rdp);
-+      return false;
- }
- void rcu_nocb_flush_deferred_wakeup(void)
-@@ -2527,8 +2533,9 @@ static int rcu_nocb_need_deferred_wakeup
-       return false;
- }
--static void do_nocb_deferred_wakeup(struct rcu_data *rdp)
-+static bool do_nocb_deferred_wakeup(struct rcu_data *rdp)
- {
-+      return false;
- }
- static void rcu_spawn_cpu_nocb_kthread(int cpu)
index 12dc341157253e6c3bad0de16595129f4ef624b2..65c8b2d4981a0ecb9cc23c18d054035124266a65 100644 (file)
@@ -284,7 +284,6 @@ x86-reboot-force-all-cpus-to-exit-vmx-root-if-vmx-is-supported.patch
 powerpc-prom-fix-ibm-arch-vec-5-platform-support-scan.patch
 rcu-pull-deferred-rcuog-wake-up-to-rcu_eqs_enter-callers.patch
 rcu-nocb-perform-deferred-wake-up-before-last-idle-s-need_resched-check.patch
-rcu-nocb-trigger-self-ipi-on-late-deferred-wake-up-before-user-resume.patch
 floppy-reintroduce-o_ndelay-fix.patch
 arm64-kexec_file-fix-memory-leakage-in-create_dtb-when-fdt_open_into-fails.patch
 arm64-uprobe-return-eopnotsupp-for-aarch32-instruction-probing.patch