]> git.ipfire.org Git - thirdparty/kernel/stable.git/commitdiff
xen/blkfront: don't trust the backend response data blindly
authorJuergen Gross <jgross@suse.com>
Mon, 29 Nov 2021 13:18:20 +0000 (14:18 +0100)
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Wed, 8 Dec 2021 07:44:06 +0000 (08:44 +0100)
commit b94e4b147fd1992ad450e1fea1fdaa3738753373 upstream.

Today blkfront will trust the backend to send only sane response data.
In order to avoid privilege escalations or crashes in case of malicious
backends verify the data to be within expected limits. Especially make
sure that the response always references an outstanding request.

Introduce a new state of the ring BLKIF_STATE_ERROR which will be
switched to in case an inconsistency is being detected. Recovering from
this state is possible only via removing and adding the virtual device
again (e.g. via a suspend/resume cycle).

Make all warning messages issued due to valid error responses rate
limited in order to avoid message floods being triggered by a malicious
backend.

Signed-off-by: Juergen Gross <jgross@suse.com>
Reviewed-by: Jan Beulich <jbeulich@suse.com>
Acked-by: Roger Pau Monné <roger.pau@citrix.com>
Link: https://lore.kernel.org/r/20210730103854.12681-4-jgross@suse.com
Signed-off-by: Juergen Gross <jgross@suse.com>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
drivers/block/xen-blkfront.c

index af027288cc23b70cbbb0d93ea575db4ed29a38d4..b27917dfdcc05193944fe05e8a77dbdad20b2156 100644 (file)
@@ -64,6 +64,7 @@ enum blkif_state {
        BLKIF_STATE_DISCONNECTED,
        BLKIF_STATE_CONNECTED,
        BLKIF_STATE_SUSPENDED,
+       BLKIF_STATE_ERROR,
 };
 
 struct grant {
@@ -79,6 +80,7 @@ struct blk_shadow {
        struct grant **indirect_grants;
        struct scatterlist *sg;
        unsigned int num_sg;
+       bool inflight;
 };
 
 struct split_bio {
@@ -495,6 +497,7 @@ static int blkif_queue_discard_req(struct request *req)
 
        /* Copy the request to the ring page. */
        *final_ring_req = *ring_req;
+       info->shadow[id].inflight = true;
 
        return 0;
 }
@@ -712,6 +715,7 @@ static int blkif_queue_rw_req(struct request *req)
 
        /* Copy request(s) to the ring page. */
        *final_ring_req = *ring_req;
+       info->shadow[id].inflight = true;
 
        if (new_persistent_gnts)
                gnttab_free_grant_references(setup.gref_head);
@@ -1324,11 +1328,17 @@ static irqreturn_t blkif_interrupt(int irq, void *dev_id)
        }
 
  again:
-       rp = info->ring.sring->rsp_prod;
+       rp = READ_ONCE(info->ring.sring->rsp_prod);
        rmb(); /* Ensure we see queued responses up to 'rp'. */
+       if (RING_RESPONSE_PROD_OVERFLOW(&info->ring, rp)) {
+               pr_alert("%s: illegal number of responses %u\n",
+                        info->gd->disk_name, rp - info->ring.rsp_cons);
+               goto err;
+       }
 
        for (i = info->ring.rsp_cons; i != rp; i++) {
                unsigned long id;
+               unsigned int op;
 
                RING_COPY_RESPONSE(&info->ring, i, &bret);
                id   = bret.id;
@@ -1339,14 +1349,28 @@ static irqreturn_t blkif_interrupt(int irq, void *dev_id)
                 * look in get_id_from_freelist.
                 */
                if (id >= BLK_RING_SIZE(info)) {
-                       WARN(1, "%s: response to %s has incorrect id (%ld)\n",
-                            info->gd->disk_name, op_name(bret.operation), id);
-                       /* We can't safely get the 'struct request' as
-                        * the id is busted. */
-                       continue;
+                       pr_alert("%s: response has incorrect id (%ld)\n",
+                                info->gd->disk_name, id);
+                       goto err;
+               }
+               if (!info->shadow[id].inflight) {
+                       pr_alert("%s: response references no pending request\n",
+                                info->gd->disk_name);
+                       goto err;
                }
+
+               info->shadow[id].inflight = false;
                req  = info->shadow[id].request;
 
+               op = info->shadow[id].req.operation;
+               if (op == BLKIF_OP_INDIRECT)
+                       op = info->shadow[id].req.u.indirect.indirect_op;
+               if (bret.operation != op) {
+                       pr_alert("%s: response has wrong operation (%u instead of %u)\n",
+                                info->gd->disk_name, bret.operation, op);
+                       goto err;
+               }
+
                if (bret.operation != BLKIF_OP_DISCARD)
                        blkif_completion(&info->shadow[id], info, &bret);
 
@@ -1361,7 +1385,8 @@ static irqreturn_t blkif_interrupt(int irq, void *dev_id)
                case BLKIF_OP_DISCARD:
                        if (unlikely(bret.status == BLKIF_RSP_EOPNOTSUPP)) {
                                struct request_queue *rq = info->rq;
-                               printk(KERN_WARNING "blkfront: %s: %s op failed\n",
+
+                               pr_warn_ratelimited("blkfront: %s: %s op failed\n",
                                           info->gd->disk_name, op_name(bret.operation));
                                error = -EOPNOTSUPP;
                                info->feature_discard = 0;
@@ -1374,13 +1399,13 @@ static irqreturn_t blkif_interrupt(int irq, void *dev_id)
                case BLKIF_OP_FLUSH_DISKCACHE:
                case BLKIF_OP_WRITE_BARRIER:
                        if (unlikely(bret.status == BLKIF_RSP_EOPNOTSUPP)) {
-                               printk(KERN_WARNING "blkfront: %s: %s op failed\n",
+                               pr_warn_ratelimited("blkfront: %s: %s op failed\n",
                                       info->gd->disk_name, op_name(bret.operation));
                                error = -EOPNOTSUPP;
                        }
                        if (unlikely(bret.status == BLKIF_RSP_ERROR &&
                                     info->shadow[id].req.u.rw.nr_segments == 0)) {
-                               printk(KERN_WARNING "blkfront: %s: empty %s op failed\n",
+                               pr_warn_ratelimited("blkfront: %s: empty %s op failed\n",
                                       info->gd->disk_name, op_name(bret.operation));
                                error = -EOPNOTSUPP;
                        }
@@ -1394,8 +1419,9 @@ static irqreturn_t blkif_interrupt(int irq, void *dev_id)
                case BLKIF_OP_READ:
                case BLKIF_OP_WRITE:
                        if (unlikely(bret.status != BLKIF_RSP_OKAY))
-                               dev_dbg(&info->xbdev->dev, "Bad return from blkdev data "
-                                       "request: %x\n", bret.status);
+                               dev_dbg_ratelimited(&info->xbdev->dev,
+                                       "Bad return from blkdev data request: %x\n",
+                                       bret.status);
 
                        blk_mq_complete_request(req, error);
                        break;
@@ -1419,6 +1445,14 @@ static irqreturn_t blkif_interrupt(int irq, void *dev_id)
        spin_unlock_irqrestore(&info->io_lock, flags);
 
        return IRQ_HANDLED;
+
+ err:
+       info->connected = BLKIF_STATE_ERROR;
+
+       spin_unlock_irqrestore(&info->io_lock, flags);
+
+       pr_alert("%s disabled for further use\n", info->gd->disk_name);
+       return IRQ_HANDLED;
 }
 
 
@@ -1928,6 +1962,7 @@ out_of_memory:
                info->shadow[i].sg = NULL;
                kfree(info->shadow[i].indirect_grants);
                info->shadow[i].indirect_grants = NULL;
+               info->shadow[i].inflight = false;
        }
        if (!list_empty(&info->indirect_pages)) {
                struct page *indirect_page, *n;