From: Sasha Levin Date: Fri, 3 Jul 2020 00:21:58 +0000 (-0400) Subject: Fixes for 5.4 X-Git-Tag: v4.4.230~45 X-Git-Url: http://git.ipfire.org/gitweb.cgi?a=commitdiff_plain;h=dfc7990061f574c7ab205bd887f792f501699fcd;p=thirdparty%2Fkernel%2Fstable-queue.git Fixes for 5.4 Signed-off-by: Sasha Levin --- diff --git a/queue-5.4/edac-amd64-read-back-the-scrub-rate-pci-register-on-.patch b/queue-5.4/edac-amd64-read-back-the-scrub-rate-pci-register-on-.patch new file mode 100644 index 00000000000..e4906ce71d3 --- /dev/null +++ b/queue-5.4/edac-amd64-read-back-the-scrub-rate-pci-register-on-.patch @@ -0,0 +1,47 @@ +From bc1b4356188007a2b1821068df952e373eb18b68 Mon Sep 17 00:00:00 2001 +From: Sasha Levin +Date: Thu, 18 Jun 2020 20:25:25 +0200 +Subject: EDAC/amd64: Read back the scrub rate PCI register on F15h + +From: Borislav Petkov + +[ Upstream commit ee470bb25d0dcdf126f586ec0ae6dca66cb340a4 ] + +Commit: + + da92110dfdfa ("EDAC, amd64_edac: Extend scrub rate support to F15hM60h") + +added support for F15h, model 0x60 CPUs but in doing so, missed to read +back SCRCTRL PCI config register on F15h CPUs which are *not* model +0x60. Add that read so that doing + + $ cat /sys/devices/system/edac/mc/mc0/sdram_scrub_rate + +can show the previously set DRAM scrub rate. + +Fixes: da92110dfdfa ("EDAC, amd64_edac: Extend scrub rate support to F15hM60h") +Reported-by: Anders Andersson +Signed-off-by: Borislav Petkov +Cc: #v4.4.. +Link: https://lkml.kernel.org/r/CAKkunMbNWppx_i6xSdDHLseA2QQmGJqj_crY=NF-GZML5np4Vw@mail.gmail.com +Signed-off-by: Sasha Levin +--- + drivers/edac/amd64_edac.c | 2 ++ + 1 file changed, 2 insertions(+) + +diff --git a/drivers/edac/amd64_edac.c b/drivers/edac/amd64_edac.c +index ad7d2bce91cd8..125a44d5a69e3 100644 +--- a/drivers/edac/amd64_edac.c ++++ b/drivers/edac/amd64_edac.c +@@ -265,6 +265,8 @@ static int get_scrub_rate(struct mem_ctl_info *mci) + + if (pvt->model == 0x60) + amd64_read_pci_cfg(pvt->F2, F15H_M60H_SCRCTRL, &scrubval); ++ else ++ amd64_read_pci_cfg(pvt->F3, SCRCTRL, &scrubval); + break; + + case 0x17: +-- +2.25.1 + diff --git a/queue-5.4/mm-fix-swap-cache-node-allocation-mask.patch b/queue-5.4/mm-fix-swap-cache-node-allocation-mask.patch new file mode 100644 index 00000000000..a9d65f59199 --- /dev/null +++ b/queue-5.4/mm-fix-swap-cache-node-allocation-mask.patch @@ -0,0 +1,98 @@ +From bdfbe6c3ccc61d8b4b75ee65d21e1a83f4e2b2d4 Mon Sep 17 00:00:00 2001 +From: Sasha Levin +Date: Thu, 25 Jun 2020 20:29:59 -0700 +Subject: mm: fix swap cache node allocation mask + +From: Hugh Dickins + +[ Upstream commit 243bce09c91b0145aeaedd5afba799d81841c030 ] + +Chris Murphy reports that a slightly overcommitted load, testing swap +and zram along with i915, splats and keeps on splatting, when it had +better fail less noisily: + + gnome-shell: page allocation failure: order:0, + mode:0x400d0(__GFP_IO|__GFP_FS|__GFP_COMP|__GFP_RECLAIMABLE), + nodemask=(null),cpuset=/,mems_allowed=0 + CPU: 2 PID: 1155 Comm: gnome-shell Not tainted 5.7.0-1.fc33.x86_64 #1 + Call Trace: + dump_stack+0x64/0x88 + warn_alloc.cold+0x75/0xd9 + __alloc_pages_slowpath.constprop.0+0xcfa/0xd30 + __alloc_pages_nodemask+0x2df/0x320 + alloc_slab_page+0x195/0x310 + allocate_slab+0x3c5/0x440 + ___slab_alloc+0x40c/0x5f0 + __slab_alloc+0x1c/0x30 + kmem_cache_alloc+0x20e/0x220 + xas_nomem+0x28/0x70 + add_to_swap_cache+0x321/0x400 + __read_swap_cache_async+0x105/0x240 + swap_cluster_readahead+0x22c/0x2e0 + shmem_swapin+0x8e/0xc0 + shmem_swapin_page+0x196/0x740 + shmem_getpage_gfp+0x3a2/0xa60 + shmem_read_mapping_page_gfp+0x32/0x60 + shmem_get_pages+0x155/0x5e0 [i915] + __i915_gem_object_get_pages+0x68/0xa0 [i915] + i915_vma_pin+0x3fe/0x6c0 [i915] + eb_add_vma+0x10b/0x2c0 [i915] + i915_gem_do_execbuffer+0x704/0x3430 [i915] + i915_gem_execbuffer2_ioctl+0x1ea/0x3e0 [i915] + drm_ioctl_kernel+0x86/0xd0 [drm] + drm_ioctl+0x206/0x390 [drm] + ksys_ioctl+0x82/0xc0 + __x64_sys_ioctl+0x16/0x20 + do_syscall_64+0x5b/0xf0 + entry_SYSCALL_64_after_hwframe+0x44/0xa9 + +Reported on 5.7, but it goes back really to 3.1: when +shmem_read_mapping_page_gfp() was implemented for use by i915, and +allowed for __GFP_NORETRY and __GFP_NOWARN flags in most places, but +missed swapin's "& GFP_KERNEL" mask for page tree node allocation in +__read_swap_cache_async() - that was to mask off HIGHUSER_MOVABLE bits +from what page cache uses, but GFP_RECLAIM_MASK is now what's needed. + +Link: https://bugzilla.kernel.org/show_bug.cgi?id=208085 +Link: http://lkml.kernel.org/r/alpine.LSU.2.11.2006151330070.11064@eggly.anvils +Fixes: 68da9f055755 ("tmpfs: pass gfp to shmem_getpage_gfp") +Signed-off-by: Hugh Dickins +Reviewed-by: Vlastimil Babka +Reviewed-by: Matthew Wilcox (Oracle) +Reported-by: Chris Murphy +Analyzed-by: Vlastimil Babka +Analyzed-by: Matthew Wilcox +Tested-by: Chris Murphy +Cc: [3.1+] +Signed-off-by: Andrew Morton +Signed-off-by: Linus Torvalds +Signed-off-by: Sasha Levin +--- + mm/swap_state.c | 4 +++- + 1 file changed, 3 insertions(+), 1 deletion(-) + +diff --git a/mm/swap_state.c b/mm/swap_state.c +index 8e7ce9a9bc5eb..4ce014dc4571a 100644 +--- a/mm/swap_state.c ++++ b/mm/swap_state.c +@@ -23,6 +23,7 @@ + #include + + #include ++#include "internal.h" + + /* + * swapper_space is a fiction, retained to simplify the path through +@@ -418,7 +419,8 @@ struct page *__read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, + /* May fail (-ENOMEM) if XArray node allocation failed. */ + __SetPageLocked(new_page); + __SetPageSwapBacked(new_page); +- err = add_to_swap_cache(new_page, entry, gfp_mask & GFP_KERNEL); ++ err = add_to_swap_cache(new_page, entry, ++ gfp_mask & GFP_RECLAIM_MASK); + if (likely(!err)) { + /* Initiate read into locked page */ + SetPageWorkingset(new_page); +-- +2.25.1 + diff --git a/queue-5.4/series b/queue-5.4/series index 39598df23dc..d5bbfec73fa 100644 --- a/queue-5.4/series +++ b/queue-5.4/series @@ -1 +1,3 @@ io_uring-make-sure-async-workqueue-is-canceled-on-ex.patch +mm-fix-swap-cache-node-allocation-mask.patch +edac-amd64-read-back-the-scrub-rate-pci-register-on-.patch