]> git.ipfire.org Git - thirdparty/kernel/stable-queue.git/commitdiff
4.9-stable patches
authorGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Sat, 20 Mar 2021 11:52:22 +0000 (12:52 +0100)
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Sat, 20 Mar 2021 11:52:22 +0000 (12:52 +0100)
added patches:
nfsd-repair-misuse-of-sv_lock-in-5.10.16-rt30.patch
nvmet-don-t-check-iosqes-iocqes-for-discovery-controllers.patch
sunrpc-fix-refcount-leak-for-rpc-auth-modules.patch
svcrdma-disable-timeouts-on-rdma-backchannel.patch

queue-4.9/nfsd-repair-misuse-of-sv_lock-in-5.10.16-rt30.patch [new file with mode: 0644]
queue-4.9/nvmet-don-t-check-iosqes-iocqes-for-discovery-controllers.patch [new file with mode: 0644]
queue-4.9/series
queue-4.9/sunrpc-fix-refcount-leak-for-rpc-auth-modules.patch [new file with mode: 0644]
queue-4.9/svcrdma-disable-timeouts-on-rdma-backchannel.patch [new file with mode: 0644]

diff --git a/queue-4.9/nfsd-repair-misuse-of-sv_lock-in-5.10.16-rt30.patch b/queue-4.9/nfsd-repair-misuse-of-sv_lock-in-5.10.16-rt30.patch
new file mode 100644 (file)
index 0000000..a127cb9
--- /dev/null
@@ -0,0 +1,152 @@
+From c7de87ff9dac5f396f62d584f3908f80ddc0e07b Mon Sep 17 00:00:00 2001
+From: Joe Korty <joe.korty@concurrent-rt.com>
+Date: Fri, 26 Feb 2021 09:38:20 -0500
+Subject: NFSD: Repair misuse of sv_lock in 5.10.16-rt30.
+
+From: Joe Korty <joe.korty@concurrent-rt.com>
+
+commit c7de87ff9dac5f396f62d584f3908f80ddc0e07b upstream.
+
+[ This problem is in mainline, but only rt has the chops to be
+able to detect it. ]
+
+Lockdep reports a circular lock dependency between serv->sv_lock and
+softirq_ctl.lock on system shutdown, when using a kernel built with
+CONFIG_PREEMPT_RT=y, and a nfs mount exists.
+
+This is due to the definition of spin_lock_bh on rt:
+
+       local_bh_disable();
+       rt_spin_lock(lock);
+
+which forces a softirq_ctl.lock -> serv->sv_lock dependency.  This is
+not a problem as long as _every_ lock of serv->sv_lock is a:
+
+       spin_lock_bh(&serv->sv_lock);
+
+but there is one of the form:
+
+       spin_lock(&serv->sv_lock);
+
+This is what is causing the circular dependency splat.  The spin_lock()
+grabs the lock without first grabbing softirq_ctl.lock via local_bh_disable.
+If later on in the critical region,  someone does a local_bh_disable, we
+get a serv->sv_lock -> softirq_ctrl.lock dependency established.  Deadlock.
+
+Fix is to make serv->sv_lock be locked with spin_lock_bh everywhere, no
+exceptions.
+
+[  OK  ] Stopped target NFS client services.
+         Stopping Logout off all iSCSI sessions on shutdown...
+         Stopping NFS server and services...
+[  109.442380]
+[  109.442385] ======================================================
+[  109.442386] WARNING: possible circular locking dependency detected
+[  109.442387] 5.10.16-rt30 #1 Not tainted
+[  109.442389] ------------------------------------------------------
+[  109.442390] nfsd/1032 is trying to acquire lock:
+[  109.442392] ffff994237617f60 ((softirq_ctrl.lock).lock){+.+.}-{2:2}, at: __local_bh_disable_ip+0xd9/0x270
+[  109.442405]
+[  109.442405] but task is already holding lock:
+[  109.442406] ffff994245cb00b0 (&serv->sv_lock){+.+.}-{0:0}, at: svc_close_list+0x1f/0x90
+[  109.442415]
+[  109.442415] which lock already depends on the new lock.
+[  109.442415]
+[  109.442416]
+[  109.442416] the existing dependency chain (in reverse order) is:
+[  109.442417]
+[  109.442417] -> #1 (&serv->sv_lock){+.+.}-{0:0}:
+[  109.442421]        rt_spin_lock+0x2b/0xc0
+[  109.442428]        svc_add_new_perm_xprt+0x42/0xa0
+[  109.442430]        svc_addsock+0x135/0x220
+[  109.442434]        write_ports+0x4b3/0x620
+[  109.442438]        nfsctl_transaction_write+0x45/0x80
+[  109.442440]        vfs_write+0xff/0x420
+[  109.442444]        ksys_write+0x4f/0xc0
+[  109.442446]        do_syscall_64+0x33/0x40
+[  109.442450]        entry_SYSCALL_64_after_hwframe+0x44/0xa9
+[  109.442454]
+[  109.442454] -> #0 ((softirq_ctrl.lock).lock){+.+.}-{2:2}:
+[  109.442457]        __lock_acquire+0x1264/0x20b0
+[  109.442463]        lock_acquire+0xc2/0x400
+[  109.442466]        rt_spin_lock+0x2b/0xc0
+[  109.442469]        __local_bh_disable_ip+0xd9/0x270
+[  109.442471]        svc_xprt_do_enqueue+0xc0/0x4d0
+[  109.442474]        svc_close_list+0x60/0x90
+[  109.442476]        svc_close_net+0x49/0x1a0
+[  109.442478]        svc_shutdown_net+0x12/0x40
+[  109.442480]        nfsd_destroy+0xc5/0x180
+[  109.442482]        nfsd+0x1bc/0x270
+[  109.442483]        kthread+0x194/0x1b0
+[  109.442487]        ret_from_fork+0x22/0x30
+[  109.442492]
+[  109.442492] other info that might help us debug this:
+[  109.442492]
+[  109.442493]  Possible unsafe locking scenario:
+[  109.442493]
+[  109.442493]        CPU0                    CPU1
+[  109.442494]        ----                    ----
+[  109.442495]   lock(&serv->sv_lock);
+[  109.442496]                                lock((softirq_ctrl.lock).lock);
+[  109.442498]                                lock(&serv->sv_lock);
+[  109.442499]   lock((softirq_ctrl.lock).lock);
+[  109.442501]
+[  109.442501]  *** DEADLOCK ***
+[  109.442501]
+[  109.442501] 3 locks held by nfsd/1032:
+[  109.442503]  #0: ffffffff93b49258 (nfsd_mutex){+.+.}-{3:3}, at: nfsd+0x19a/0x270
+[  109.442508]  #1: ffff994245cb00b0 (&serv->sv_lock){+.+.}-{0:0}, at: svc_close_list+0x1f/0x90
+[  109.442512]  #2: ffffffff93a81b20 (rcu_read_lock){....}-{1:2}, at: rt_spin_lock+0x5/0xc0
+[  109.442518]
+[  109.442518] stack backtrace:
+[  109.442519] CPU: 0 PID: 1032 Comm: nfsd Not tainted 5.10.16-rt30 #1
+[  109.442522] Hardware name: Supermicro X9DRL-3F/iF/X9DRL-3F/iF, BIOS 3.2 09/22/2015
+[  109.442524] Call Trace:
+[  109.442527]  dump_stack+0x77/0x97
+[  109.442533]  check_noncircular+0xdc/0xf0
+[  109.442546]  __lock_acquire+0x1264/0x20b0
+[  109.442553]  lock_acquire+0xc2/0x400
+[  109.442564]  rt_spin_lock+0x2b/0xc0
+[  109.442570]  __local_bh_disable_ip+0xd9/0x270
+[  109.442573]  svc_xprt_do_enqueue+0xc0/0x4d0
+[  109.442577]  svc_close_list+0x60/0x90
+[  109.442581]  svc_close_net+0x49/0x1a0
+[  109.442585]  svc_shutdown_net+0x12/0x40
+[  109.442588]  nfsd_destroy+0xc5/0x180
+[  109.442590]  nfsd+0x1bc/0x270
+[  109.442595]  kthread+0x194/0x1b0
+[  109.442600]  ret_from_fork+0x22/0x30
+[  109.518225] nfsd: last server has exited, flushing export cache
+[  OK  ] Stopped NFSv4 ID-name mapping service.
+[  OK  ] Stopped GSSAPI Proxy Daemon.
+[  OK  ] Stopped NFS Mount Daemon.
+[  OK  ] Stopped NFS status monitor for NFSv2/3 locking..
+
+Fixes: 719f8bcc883e ("svcrpc: fix xpt_list traversal locking on shutdown")
+Signed-off-by: Joe Korty <joe.korty@concurrent-rt.com>
+Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
+Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
+---
+ net/sunrpc/svc_xprt.c |    4 ++--
+ 1 file changed, 2 insertions(+), 2 deletions(-)
+
+--- a/net/sunrpc/svc_xprt.c
++++ b/net/sunrpc/svc_xprt.c
+@@ -1104,7 +1104,7 @@ static int svc_close_list(struct svc_ser
+       struct svc_xprt *xprt;
+       int ret = 0;
+-      spin_lock(&serv->sv_lock);
++      spin_lock_bh(&serv->sv_lock);
+       list_for_each_entry(xprt, xprt_list, xpt_list) {
+               if (xprt->xpt_net != net)
+                       continue;
+@@ -1112,7 +1112,7 @@ static int svc_close_list(struct svc_ser
+               set_bit(XPT_CLOSE, &xprt->xpt_flags);
+               svc_xprt_enqueue(xprt);
+       }
+-      spin_unlock(&serv->sv_lock);
++      spin_unlock_bh(&serv->sv_lock);
+       return ret;
+ }
diff --git a/queue-4.9/nvmet-don-t-check-iosqes-iocqes-for-discovery-controllers.patch b/queue-4.9/nvmet-don-t-check-iosqes-iocqes-for-discovery-controllers.patch
new file mode 100644 (file)
index 0000000..38780d3
--- /dev/null
@@ -0,0 +1,57 @@
+From d218a8a3003e84ab136e69a4e30dd4ec7dab2d22 Mon Sep 17 00:00:00 2001
+From: Sagi Grimberg <sagi@grimberg.me>
+Date: Mon, 15 Mar 2021 15:34:51 -0700
+Subject: nvmet: don't check iosqes,iocqes for discovery controllers
+
+From: Sagi Grimberg <sagi@grimberg.me>
+
+commit d218a8a3003e84ab136e69a4e30dd4ec7dab2d22 upstream.
+
+From the base spec, Figure 78:
+
+  "Controller Configuration, these fields are defined as parameters to
+   configure an "I/O Controller (IOC)" and not to configure a "Discovery
+   Controller (DC).
+
+   ...
+   If the controller does not support I/O queues, then this field shall
+   be read-only with a value of 0h
+
+Just perform this check for I/O controllers.
+
+Fixes: a07b4970f464 ("nvmet: add a generic NVMe target")
+Reported-by: Belanger, Martin <Martin.Belanger@dell.com>
+Signed-off-by: Sagi Grimberg <sagi@grimberg.me>
+Reviewed-by: Chaitanya Kulkarni <chaitanya.kulkarni@wdc.com>
+Signed-off-by: Christoph Hellwig <hch@lst.de>
+Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
+---
+ drivers/nvme/target/core.c |   17 ++++++++++++++---
+ 1 file changed, 14 insertions(+), 3 deletions(-)
+
+--- a/drivers/nvme/target/core.c
++++ b/drivers/nvme/target/core.c
+@@ -574,9 +574,20 @@ static void nvmet_start_ctrl(struct nvme
+ {
+       lockdep_assert_held(&ctrl->lock);
+-      if (nvmet_cc_iosqes(ctrl->cc) != NVME_NVM_IOSQES ||
+-          nvmet_cc_iocqes(ctrl->cc) != NVME_NVM_IOCQES ||
+-          nvmet_cc_mps(ctrl->cc) != 0 ||
++      /*
++       * Only I/O controllers should verify iosqes,iocqes.
++       * Strictly speaking, the spec says a discovery controller
++       * should verify iosqes,iocqes are zeroed, however that
++       * would break backwards compatibility, so don't enforce it.
++       */
++      if (ctrl->subsys->type != NVME_NQN_DISC &&
++          (nvmet_cc_iosqes(ctrl->cc) != NVME_NVM_IOSQES ||
++           nvmet_cc_iocqes(ctrl->cc) != NVME_NVM_IOCQES)) {
++              ctrl->csts = NVME_CSTS_CFS;
++              return;
++      }
++
++      if (nvmet_cc_mps(ctrl->cc) != 0 ||
+           nvmet_cc_ams(ctrl->cc) != 0 ||
+           nvmet_cc_css(ctrl->cc) != 0) {
+               ctrl->csts = NVME_CSTS_CFS;
index afaa14f789ed7397cc98b68e889418d805167768..230737a48ea056c13ab88bab0fecefda7b8b1709 100644 (file)
@@ -5,3 +5,7 @@ net-dsa-b53-support-setting-learning-on-port.patch
 ixgbe-check-for-tx-timestamp-timeouts-during-watchdog.patch
 ixgbe-prevent-ptp_rx_hang-from-running-when-in-filter_all-mode.patch
 btrfs-fix-race-when-cloning-extent-buffer-during-rewind-of-an-old-root.patch
+nvmet-don-t-check-iosqes-iocqes-for-discovery-controllers.patch
+nfsd-repair-misuse-of-sv_lock-in-5.10.16-rt30.patch
+svcrdma-disable-timeouts-on-rdma-backchannel.patch
+sunrpc-fix-refcount-leak-for-rpc-auth-modules.patch
diff --git a/queue-4.9/sunrpc-fix-refcount-leak-for-rpc-auth-modules.patch b/queue-4.9/sunrpc-fix-refcount-leak-for-rpc-auth-modules.patch
new file mode 100644 (file)
index 0000000..dfc230a
--- /dev/null
@@ -0,0 +1,53 @@
+From f1442d6349a2e7bb7a6134791bdc26cb776c79af Mon Sep 17 00:00:00 2001
+From: Daniel Kobras <kobras@puzzle-itc.de>
+Date: Sat, 27 Feb 2021 00:04:37 +0100
+Subject: sunrpc: fix refcount leak for rpc auth modules
+
+From: Daniel Kobras <kobras@puzzle-itc.de>
+
+commit f1442d6349a2e7bb7a6134791bdc26cb776c79af upstream.
+
+If an auth module's accept op returns SVC_CLOSE, svc_process_common()
+enters a call path that does not call svc_authorise() before leaving the
+function, and thus leaks a reference on the auth module's refcount. Hence,
+make sure calls to svc_authenticate() and svc_authorise() are paired for
+all call paths, to make sure rpc auth modules can be unloaded.
+
+Signed-off-by: Daniel Kobras <kobras@puzzle-itc.de>
+Fixes: 4d712ef1db05 ("svcauth_gss: Close connection when dropping an incoming message")
+Link: https://lore.kernel.org/linux-nfs/3F1B347F-B809-478F-A1E9-0BE98E22B0F0@oracle.com/T/#t
+Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
+Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
+---
+ net/sunrpc/svc.c |    6 ++++--
+ 1 file changed, 4 insertions(+), 2 deletions(-)
+
+--- a/net/sunrpc/svc.c
++++ b/net/sunrpc/svc.c
+@@ -1306,7 +1306,7 @@ svc_process_common(struct svc_rqst *rqst
+  sendit:
+       if (svc_authorise(rqstp))
+-              goto close;
++              goto close_xprt;
+       return 1;               /* Caller can now send it */
+  dropit:
+@@ -1315,6 +1315,8 @@ svc_process_common(struct svc_rqst *rqst
+       return 0;
+  close:
++      svc_authorise(rqstp);
++close_xprt:
+       if (rqstp->rq_xprt && test_bit(XPT_TEMP, &rqstp->rq_xprt->xpt_flags))
+               svc_close_xprt(rqstp->rq_xprt);
+       dprintk("svc: svc_process close\n");
+@@ -1323,7 +1325,7 @@ svc_process_common(struct svc_rqst *rqst
+ err_short_len:
+       svc_printk(rqstp, "short len %Zd, dropping request\n",
+                       argv->iov_len);
+-      goto close;
++      goto close_xprt;
+ err_bad_rpc:
+       serv->sv_stats->rpcbadfmt++;
diff --git a/queue-4.9/svcrdma-disable-timeouts-on-rdma-backchannel.patch b/queue-4.9/svcrdma-disable-timeouts-on-rdma-backchannel.patch
new file mode 100644 (file)
index 0000000..c410b00
--- /dev/null
@@ -0,0 +1,38 @@
+From 6820bf77864d5894ff67b5c00d7dba8f92011e3d Mon Sep 17 00:00:00 2001
+From: Timo Rothenpieler <timo@rothenpieler.org>
+Date: Tue, 23 Feb 2021 00:36:19 +0100
+Subject: svcrdma: disable timeouts on rdma backchannel
+
+From: Timo Rothenpieler <timo@rothenpieler.org>
+
+commit 6820bf77864d5894ff67b5c00d7dba8f92011e3d upstream.
+
+This brings it in line with the regular tcp backchannel, which also has
+all those timeouts disabled.
+
+Prevents the backchannel from timing out, getting some async operations
+like server side copying getting stuck indefinitely on the client side.
+
+Signed-off-by: Timo Rothenpieler <timo@rothenpieler.org>
+Fixes: 5d252f90a800 ("svcrdma: Add class for RDMA backwards direction transport")
+Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
+Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
+---
+ net/sunrpc/xprtrdma/svc_rdma_backchannel.c |    6 +++---
+ 1 file changed, 3 insertions(+), 3 deletions(-)
+
+--- a/net/sunrpc/xprtrdma/svc_rdma_backchannel.c
++++ b/net/sunrpc/xprtrdma/svc_rdma_backchannel.c
+@@ -328,9 +328,9 @@ xprt_setup_rdma_bc(struct xprt_create *a
+       xprt->timeout = &xprt_rdma_bc_timeout;
+       xprt_set_bound(xprt);
+       xprt_set_connected(xprt);
+-      xprt->bind_timeout = RPCRDMA_BIND_TO;
+-      xprt->reestablish_timeout = RPCRDMA_INIT_REEST_TO;
+-      xprt->idle_timeout = RPCRDMA_IDLE_DISC_TO;
++      xprt->bind_timeout = 0;
++      xprt->reestablish_timeout = 0;
++      xprt->idle_timeout = 0;
+       xprt->prot = XPRT_TRANSPORT_BC_RDMA;
+       xprt->tsh_size = RPCRDMA_HDRLEN_MIN / sizeof(__be32);