]> git.ipfire.org Git - thirdparty/kernel/stable-queue.git/commitdiff
4.14-stable patches
authorGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Fri, 3 Apr 2020 09:31:50 +0000 (11:31 +0200)
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Fri, 3 Apr 2020 09:31:50 +0000 (11:31 +0200)
added patches:
blk-mq-sync-the-update-nr_hw_queues-with-blk_mq_queue_tag_busy_iter.patch
drm-etnaviv-replace-mmu-flush-marker-with-flush-sequence.patch

queue-4.14/blk-mq-sync-the-update-nr_hw_queues-with-blk_mq_queue_tag_busy_iter.patch [new file with mode: 0644]
queue-4.14/drm-etnaviv-replace-mmu-flush-marker-with-flush-sequence.patch [new file with mode: 0644]
queue-4.14/series

diff --git a/queue-4.14/blk-mq-sync-the-update-nr_hw_queues-with-blk_mq_queue_tag_busy_iter.patch b/queue-4.14/blk-mq-sync-the-update-nr_hw_queues-with-blk_mq_queue_tag_busy_iter.patch
new file mode 100644 (file)
index 0000000..823a992
--- /dev/null
@@ -0,0 +1,76 @@
+From f5bbbbe4d63577026f908a809f22f5fd5a90ea1f Mon Sep 17 00:00:00 2001
+From: Jianchao Wang <jianchao.w.wang@oracle.com>
+Date: Tue, 21 Aug 2018 15:15:04 +0800
+Subject: blk-mq: sync the update nr_hw_queues with blk_mq_queue_tag_busy_iter
+
+From: Jianchao Wang <jianchao.w.wang@oracle.com>
+
+commit f5bbbbe4d63577026f908a809f22f5fd5a90ea1f upstream.
+
+For blk-mq, part_in_flight/rw will invoke blk_mq_in_flight/rw to
+account the inflight requests. It will access the queue_hw_ctx and
+nr_hw_queues w/o any protection. When updating nr_hw_queues and
+blk_mq_in_flight/rw occur concurrently, panic comes up.
+
+Before update nr_hw_queues, the q will be frozen. So we could use
+q_usage_counter to avoid the race. percpu_ref_is_zero is used here
+so that we will not miss any in-flight request. The access to
+nr_hw_queues and queue_hw_ctx in blk_mq_queue_tag_busy_iter are
+under rcu critical section, __blk_mq_update_nr_hw_queues could use
+synchronize_rcu to ensure the zeroed q_usage_counter to be globally
+visible.
+
+Signed-off-by: Jianchao Wang <jianchao.w.wang@oracle.com>
+Reviewed-by: Ming Lei <ming.lei@redhat.com>
+Signed-off-by: Jens Axboe <axboe@kernel.dk>
+Cc: Giuliano Procida <gprocida@google.com>
+Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
+
+---
+ block/blk-mq-tag.c |   14 +++++++++++++-
+ block/blk-mq.c     |    4 ++++
+ 2 files changed, 17 insertions(+), 1 deletion(-)
+
+--- a/block/blk-mq-tag.c
++++ b/block/blk-mq-tag.c
+@@ -334,6 +334,18 @@ void blk_mq_queue_tag_busy_iter(struct r
+       struct blk_mq_hw_ctx *hctx;
+       int i;
++      /*
++       * __blk_mq_update_nr_hw_queues will update the nr_hw_queues and
++       * queue_hw_ctx after freeze the queue. So we could use q_usage_counter
++       * to avoid race with it. __blk_mq_update_nr_hw_queues will users
++       * synchronize_rcu to ensure all of the users go out of the critical
++       * section below and see zeroed q_usage_counter.
++       */
++      rcu_read_lock();
++      if (percpu_ref_is_zero(&q->q_usage_counter)) {
++              rcu_read_unlock();
++              return;
++      }
+       queue_for_each_hw_ctx(q, hctx, i) {
+               struct blk_mq_tags *tags = hctx->tags;
+@@ -349,7 +361,7 @@ void blk_mq_queue_tag_busy_iter(struct r
+                       bt_for_each(hctx, &tags->breserved_tags, fn, priv, true);
+               bt_for_each(hctx, &tags->bitmap_tags, fn, priv, false);
+       }
+-
++      rcu_read_unlock();
+ }
+ static int bt_alloc(struct sbitmap_queue *bt, unsigned int depth,
+--- a/block/blk-mq.c
++++ b/block/blk-mq.c
+@@ -2748,6 +2748,10 @@ static void __blk_mq_update_nr_hw_queues
+       list_for_each_entry(q, &set->tag_list, tag_set_list)
+               blk_mq_unfreeze_queue(q);
++      /*
++       * Sync with blk_mq_queue_tag_busy_iter.
++       */
++      synchronize_rcu();
+ }
+ void blk_mq_update_nr_hw_queues(struct blk_mq_tag_set *set, int nr_hw_queues)
diff --git a/queue-4.14/drm-etnaviv-replace-mmu-flush-marker-with-flush-sequence.patch b/queue-4.14/drm-etnaviv-replace-mmu-flush-marker-with-flush-sequence.patch
new file mode 100644 (file)
index 0000000..d662167
--- /dev/null
@@ -0,0 +1,147 @@
+From 4900dda90af2cb13bc1d4c12ce94b98acc8fe64e Mon Sep 17 00:00:00 2001
+From: Lucas Stach <l.stach@pengutronix.de>
+Date: Fri, 5 Jul 2019 19:17:23 +0200
+Subject: drm/etnaviv: replace MMU flush marker with flush sequence
+MIME-Version: 1.0
+Content-Type: text/plain; charset=UTF-8
+Content-Transfer-Encoding: 8bit
+
+From: Lucas Stach <l.stach@pengutronix.de>
+
+commit 4900dda90af2cb13bc1d4c12ce94b98acc8fe64e upstream.
+
+If a MMU is shared between multiple GPUs, all of them need to flush their
+TLBs, so a single marker that gets reset on the first flush won't do.
+Replace the flush marker with a sequence number, so that it's possible to
+check if the TLB is in sync with the current page table state for each GPU.
+
+Signed-off-by: Lucas Stach <l.stach@pengutronix.de>
+Reviewed-by: Philipp Zabel <p.zabel@pengutronix.de>
+Reviewed-by: Guido Günther <agx@sigxcpu.org>
+Signed-off-by: Robert Beckett <bob.beckett@collabora.com>
+Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
+
+---
+ drivers/gpu/drm/etnaviv/etnaviv_buffer.c |   10 ++++++----
+ drivers/gpu/drm/etnaviv/etnaviv_gpu.c    |    2 +-
+ drivers/gpu/drm/etnaviv/etnaviv_gpu.h    |    1 +
+ drivers/gpu/drm/etnaviv/etnaviv_mmu.c    |    8 ++++----
+ drivers/gpu/drm/etnaviv/etnaviv_mmu.h    |    2 +-
+ 5 files changed, 13 insertions(+), 10 deletions(-)
+
+--- a/drivers/gpu/drm/etnaviv/etnaviv_buffer.c
++++ b/drivers/gpu/drm/etnaviv/etnaviv_buffer.c
+@@ -258,6 +258,8 @@ void etnaviv_buffer_queue(struct etnaviv
+       unsigned int waitlink_offset = buffer->user_size - 16;
+       u32 return_target, return_dwords;
+       u32 link_target, link_dwords;
++      unsigned int new_flush_seq = READ_ONCE(gpu->mmu->flush_seq);
++      bool need_flush = gpu->flush_seq != new_flush_seq;
+       if (drm_debug & DRM_UT_DRIVER)
+               etnaviv_buffer_dump(gpu, buffer, 0, 0x50);
+@@ -270,14 +272,14 @@ void etnaviv_buffer_queue(struct etnaviv
+        * need to append a mmu flush load state, followed by a new
+        * link to this buffer - a total of four additional words.
+        */
+-      if (gpu->mmu->need_flush || gpu->switch_context) {
++      if (need_flush || gpu->switch_context) {
+               u32 target, extra_dwords;
+               /* link command */
+               extra_dwords = 1;
+               /* flush command */
+-              if (gpu->mmu->need_flush) {
++              if (need_flush) {
+                       if (gpu->mmu->version == ETNAVIV_IOMMU_V1)
+                               extra_dwords += 1;
+                       else
+@@ -290,7 +292,7 @@ void etnaviv_buffer_queue(struct etnaviv
+               target = etnaviv_buffer_reserve(gpu, buffer, extra_dwords);
+-              if (gpu->mmu->need_flush) {
++              if (need_flush) {
+                       /* Add the MMU flush */
+                       if (gpu->mmu->version == ETNAVIV_IOMMU_V1) {
+                               CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_MMU,
+@@ -310,7 +312,7 @@ void etnaviv_buffer_queue(struct etnaviv
+                                       SYNC_RECIPIENT_PE);
+                       }
+-                      gpu->mmu->need_flush = false;
++                      gpu->flush_seq = new_flush_seq;
+               }
+               if (gpu->switch_context) {
+--- a/drivers/gpu/drm/etnaviv/etnaviv_gpu.c
++++ b/drivers/gpu/drm/etnaviv/etnaviv_gpu.c
+@@ -1353,7 +1353,7 @@ int etnaviv_gpu_submit(struct etnaviv_gp
+       gpu->active_fence = submit->fence->seqno;
+       if (gpu->lastctx != cmdbuf->ctx) {
+-              gpu->mmu->need_flush = true;
++              gpu->mmu->flush_seq++;
+               gpu->switch_context = true;
+               gpu->lastctx = cmdbuf->ctx;
+       }
+--- a/drivers/gpu/drm/etnaviv/etnaviv_gpu.h
++++ b/drivers/gpu/drm/etnaviv/etnaviv_gpu.h
+@@ -138,6 +138,7 @@ struct etnaviv_gpu {
+       struct etnaviv_iommu *mmu;
+       struct etnaviv_cmdbuf_suballoc *cmdbuf_suballoc;
++      unsigned int flush_seq;
+       /* Power Control: */
+       struct clk *clk_bus;
+--- a/drivers/gpu/drm/etnaviv/etnaviv_mmu.c
++++ b/drivers/gpu/drm/etnaviv/etnaviv_mmu.c
+@@ -132,7 +132,7 @@ static int etnaviv_iommu_find_iova(struc
+                */
+               if (mmu->last_iova) {
+                       mmu->last_iova = 0;
+-                      mmu->need_flush = true;
++                      mmu->flush_seq++;
+                       continue;
+               }
+@@ -246,7 +246,7 @@ int etnaviv_iommu_map_gem(struct etnaviv
+       }
+       list_add_tail(&mapping->mmu_node, &mmu->mappings);
+-      mmu->need_flush = true;
++      mmu->flush_seq++;
+       mutex_unlock(&mmu->lock);
+       return ret;
+@@ -264,7 +264,7 @@ void etnaviv_iommu_unmap_gem(struct etna
+               etnaviv_iommu_remove_mapping(mmu, mapping);
+       list_del(&mapping->mmu_node);
+-      mmu->need_flush = true;
++      mmu->flush_seq++;
+       mutex_unlock(&mmu->lock);
+ }
+@@ -346,7 +346,7 @@ int etnaviv_iommu_get_suballoc_va(struct
+                       return ret;
+               }
+               mmu->last_iova = vram_node->start + size;
+-              gpu->mmu->need_flush = true;
++              mmu->flush_seq++;
+               mutex_unlock(&mmu->lock);
+               *iova = (u32)vram_node->start;
+--- a/drivers/gpu/drm/etnaviv/etnaviv_mmu.h
++++ b/drivers/gpu/drm/etnaviv/etnaviv_mmu.h
+@@ -44,7 +44,7 @@ struct etnaviv_iommu {
+       struct list_head mappings;
+       struct drm_mm mm;
+       u32 last_iova;
+-      bool need_flush;
++      unsigned int flush_seq;
+ };
+ struct etnaviv_gem_object;
index 7ba75505f379ef7ed3e6ffaa864284e2351959fd..0c6b2f5d17c5967e6c956b8a86f9c69b37913f8d 100644 (file)
@@ -5,3 +5,5 @@ sctp-fix-possibly-using-a-bad-saddr-with-a-given-dst.patch
 drm-bochs-downgrade-pci_request_region-failure-from-.patch
 initramfs-restore-default-compression-behavior.patch
 tools-power-turbostat-fix-gcc-build-warnings.patch
+drm-etnaviv-replace-mmu-flush-marker-with-flush-sequence.patch
+blk-mq-sync-the-update-nr_hw_queues-with-blk_mq_queue_tag_busy_iter.patch