+++ /dev/null
-From 3f9218db9eb28f31c8b11035df7eff1a911b4ea8 Mon Sep 17 00:00:00 2001
-From: Andrey Konovalov <andreyknvl@google.com>
-Date: Wed, 20 Feb 2019 22:20:25 -0800
-Subject: kasan, slab: fix conflicts with CONFIG_HARDENED_USERCOPY
-
-[ Upstream commit 219667c23c68eb3dbc0d5662b9246f28477fe529 ]
-
-Similarly to commit 96fedce27e13 ("kasan: make tag based mode work with
-CONFIG_HARDENED_USERCOPY"), we need to reset pointer tags in
-__check_heap_object() in mm/slab.c before doing any pointer math.
-
-Link: http://lkml.kernel.org/r/9a5c0f958db10e69df5ff9f2b997866b56b7effc.1550602886.git.andreyknvl@google.com
-Signed-off-by: Andrey Konovalov <andreyknvl@google.com>
-Tested-by: Qian Cai <cai@lca.pw>
-Cc: Alexander Potapenko <glider@google.com>
-Cc: Andrey Ryabinin <aryabinin@virtuozzo.com>
-Cc: Catalin Marinas <catalin.marinas@arm.com>
-Cc: Dmitry Vyukov <dvyukov@google.com>
-Cc: Evgeniy Stepanov <eugenis@google.com>
-Cc: Kostya Serebryany <kcc@google.com>
-Cc: Vincenzo Frascino <vincenzo.frascino@arm.com>
-Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
-Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-Signed-off-by: Sasha Levin <sashal@kernel.org>
----
- mm/slab.c | 2 ++
- 1 file changed, 2 insertions(+)
-
-diff --git a/mm/slab.c b/mm/slab.c
-index 09df506ae830..409631e49295 100644
---- a/mm/slab.c
-+++ b/mm/slab.c
-@@ -4415,6 +4415,8 @@ const char *__check_heap_object(const void *ptr, unsigned long n,
- unsigned int objnr;
- unsigned long offset;
-
-+ ptr = kasan_reset_tag(ptr);
-+
- /* Find and validate object. */
- cachep = page->slab_cache;
- objnr = obj_to_index(cachep, page, (void *)ptr);
---
-2.19.1
-
+++ /dev/null
-From a8146d43007331bebbe675fd3acbcd6a46342a37 Mon Sep 17 00:00:00 2001
-From: Andrey Konovalov <andreyknvl@google.com>
-Date: Wed, 20 Feb 2019 22:19:23 -0800
-Subject: kasan, slub: move kasan_poison_slab hook before page_address
-
-[ Upstream commit a71012242837fe5e67d8c999cfc357174ed5dba0 ]
-
-With tag based KASAN page_address() looks at the page flags to see whether
-the resulting pointer needs to have a tag set. Since we don't want to set
-a tag when page_address() is called on SLAB pages, we call
-page_kasan_tag_reset() in kasan_poison_slab(). However in allocate_slab()
-page_address() is called before kasan_poison_slab(). Fix it by changing
-the order.
-
-[andreyknvl@google.com: fix compilation error when CONFIG_SLUB_DEBUG=n]
- Link: http://lkml.kernel.org/r/ac27cc0bbaeb414ed77bcd6671a877cf3546d56e.1550066133.git.andreyknvl@google.com
-Link: http://lkml.kernel.org/r/cd895d627465a3f1c712647072d17f10883be2a1.1549921721.git.andreyknvl@google.com
-Signed-off-by: Andrey Konovalov <andreyknvl@google.com>
-Cc: Alexander Potapenko <glider@google.com>
-Cc: Andrey Ryabinin <aryabinin@virtuozzo.com>
-Cc: Catalin Marinas <catalin.marinas@arm.com>
-Cc: Christoph Lameter <cl@linux.com>
-Cc: David Rientjes <rientjes@google.com>
-Cc: Dmitry Vyukov <dvyukov@google.com>
-Cc: Evgeniy Stepanov <eugenis@google.com>
-Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
-Cc: Kostya Serebryany <kcc@google.com>
-Cc: Pekka Enberg <penberg@kernel.org>
-Cc: Qian Cai <cai@lca.pw>
-Cc: Vincenzo Frascino <vincenzo.frascino@arm.com>
-Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
-Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-Signed-off-by: Sasha Levin <sashal@kernel.org>
----
- mm/slub.c | 19 +++++++++++++++----
- 1 file changed, 15 insertions(+), 4 deletions(-)
-
-diff --git a/mm/slub.c b/mm/slub.c
-index 220d42e592ef..f14ef59c9e57 100644
---- a/mm/slub.c
-+++ b/mm/slub.c
-@@ -1087,6 +1087,16 @@ static void setup_object_debug(struct kmem_cache *s, struct page *page,
- init_tracking(s, object);
- }
-
-+static void setup_page_debug(struct kmem_cache *s, void *addr, int order)
-+{
-+ if (!(s->flags & SLAB_POISON))
-+ return;
-+
-+ metadata_access_enable();
-+ memset(addr, POISON_INUSE, PAGE_SIZE << order);
-+ metadata_access_disable();
-+}
-+
- static inline int alloc_consistency_checks(struct kmem_cache *s,
- struct page *page,
- void *object, unsigned long addr)
-@@ -1304,6 +1314,8 @@ unsigned long kmem_cache_flags(unsigned long object_size,
- #else /* !CONFIG_SLUB_DEBUG */
- static inline void setup_object_debug(struct kmem_cache *s,
- struct page *page, void *object) {}
-+static inline void setup_page_debug(struct kmem_cache *s,
-+ void *addr, int order) {}
-
- static inline int alloc_debug_processing(struct kmem_cache *s,
- struct page *page, void *object, unsigned long addr) { return 0; }
-@@ -1599,12 +1611,11 @@ static struct page *allocate_slab(struct kmem_cache *s, gfp_t flags, int node)
- if (page_is_pfmemalloc(page))
- SetPageSlabPfmemalloc(page);
-
-- start = page_address(page);
-+ kasan_poison_slab(page);
-
-- if (unlikely(s->flags & SLAB_POISON))
-- memset(start, POISON_INUSE, PAGE_SIZE << order);
-+ start = page_address(page);
-
-- kasan_poison_slab(page);
-+ setup_page_debug(s, start, order);
-
- shuffle = shuffle_freelist(s, page);
-
---
-2.19.1
-
qed-fix-iwarp-syn-packet-mac-address-validation.patch
arm64-relax-gic-version-check-during-early-boot.patch
net-marvell-mvneta-fix-dma-debug-warning.patch
-kasan-slub-move-kasan_poison_slab-hook-before-page_a.patch
tmpfs-fix-link-accounting-when-a-tmpfile-is-linked-i.patch
-kasan-slab-fix-conflicts-with-config_hardened_userco.patch
ixgbe-fix-older-devices-that-do-not-support-ixgbe_mr.patch
arcv2-lib-memcpy-fix-doing-prefetchw-outside-of-buff.patch
arc-uacces-remove-lp_start-lp_end-from-clobber-list.patch
+++ /dev/null
-From e59823590ea9e0081abaf0816ea774bed96757ac Mon Sep 17 00:00:00 2001
-From: Andrey Konovalov <andreyknvl@google.com>
-Date: Wed, 20 Feb 2019 22:20:25 -0800
-Subject: kasan, slab: fix conflicts with CONFIG_HARDENED_USERCOPY
-
-[ Upstream commit 219667c23c68eb3dbc0d5662b9246f28477fe529 ]
-
-Similarly to commit 96fedce27e13 ("kasan: make tag based mode work with
-CONFIG_HARDENED_USERCOPY"), we need to reset pointer tags in
-__check_heap_object() in mm/slab.c before doing any pointer math.
-
-Link: http://lkml.kernel.org/r/9a5c0f958db10e69df5ff9f2b997866b56b7effc.1550602886.git.andreyknvl@google.com
-Signed-off-by: Andrey Konovalov <andreyknvl@google.com>
-Tested-by: Qian Cai <cai@lca.pw>
-Cc: Alexander Potapenko <glider@google.com>
-Cc: Andrey Ryabinin <aryabinin@virtuozzo.com>
-Cc: Catalin Marinas <catalin.marinas@arm.com>
-Cc: Dmitry Vyukov <dvyukov@google.com>
-Cc: Evgeniy Stepanov <eugenis@google.com>
-Cc: Kostya Serebryany <kcc@google.com>
-Cc: Vincenzo Frascino <vincenzo.frascino@arm.com>
-Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
-Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-Signed-off-by: Sasha Levin <sashal@kernel.org>
----
- mm/slab.c | 2 ++
- 1 file changed, 2 insertions(+)
-
-diff --git a/mm/slab.c b/mm/slab.c
-index fad6839e8eab..813bd89c263b 100644
---- a/mm/slab.c
-+++ b/mm/slab.c
-@@ -4421,6 +4421,8 @@ void __check_heap_object(const void *ptr, unsigned long n, struct page *page,
- unsigned int objnr;
- unsigned long offset;
-
-+ ptr = kasan_reset_tag(ptr);
-+
- /* Find and validate object. */
- cachep = page->slab_cache;
- objnr = obj_to_index(cachep, page, (void *)ptr);
---
-2.19.1
-
+++ /dev/null
-From 84986e6bcd937ac6402c7061f4d0cd28f73e6087 Mon Sep 17 00:00:00 2001
-From: Andrey Konovalov <andreyknvl@google.com>
-Date: Wed, 20 Feb 2019 22:19:23 -0800
-Subject: kasan, slub: move kasan_poison_slab hook before page_address
-
-[ Upstream commit a71012242837fe5e67d8c999cfc357174ed5dba0 ]
-
-With tag based KASAN page_address() looks at the page flags to see whether
-the resulting pointer needs to have a tag set. Since we don't want to set
-a tag when page_address() is called on SLAB pages, we call
-page_kasan_tag_reset() in kasan_poison_slab(). However in allocate_slab()
-page_address() is called before kasan_poison_slab(). Fix it by changing
-the order.
-
-[andreyknvl@google.com: fix compilation error when CONFIG_SLUB_DEBUG=n]
- Link: http://lkml.kernel.org/r/ac27cc0bbaeb414ed77bcd6671a877cf3546d56e.1550066133.git.andreyknvl@google.com
-Link: http://lkml.kernel.org/r/cd895d627465a3f1c712647072d17f10883be2a1.1549921721.git.andreyknvl@google.com
-Signed-off-by: Andrey Konovalov <andreyknvl@google.com>
-Cc: Alexander Potapenko <glider@google.com>
-Cc: Andrey Ryabinin <aryabinin@virtuozzo.com>
-Cc: Catalin Marinas <catalin.marinas@arm.com>
-Cc: Christoph Lameter <cl@linux.com>
-Cc: David Rientjes <rientjes@google.com>
-Cc: Dmitry Vyukov <dvyukov@google.com>
-Cc: Evgeniy Stepanov <eugenis@google.com>
-Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
-Cc: Kostya Serebryany <kcc@google.com>
-Cc: Pekka Enberg <penberg@kernel.org>
-Cc: Qian Cai <cai@lca.pw>
-Cc: Vincenzo Frascino <vincenzo.frascino@arm.com>
-Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
-Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-Signed-off-by: Sasha Levin <sashal@kernel.org>
----
- mm/slub.c | 19 +++++++++++++++----
- 1 file changed, 15 insertions(+), 4 deletions(-)
-
-diff --git a/mm/slub.c b/mm/slub.c
-index 8da34a8af53d..7666cff34bfb 100644
---- a/mm/slub.c
-+++ b/mm/slub.c
-@@ -1075,6 +1075,16 @@ static void setup_object_debug(struct kmem_cache *s, struct page *page,
- init_tracking(s, object);
- }
-
-+static void setup_page_debug(struct kmem_cache *s, void *addr, int order)
-+{
-+ if (!(s->flags & SLAB_POISON))
-+ return;
-+
-+ metadata_access_enable();
-+ memset(addr, POISON_INUSE, PAGE_SIZE << order);
-+ metadata_access_disable();
-+}
-+
- static inline int alloc_consistency_checks(struct kmem_cache *s,
- struct page *page,
- void *object, unsigned long addr)
-@@ -1292,6 +1302,8 @@ slab_flags_t kmem_cache_flags(unsigned int object_size,
- #else /* !CONFIG_SLUB_DEBUG */
- static inline void setup_object_debug(struct kmem_cache *s,
- struct page *page, void *object) {}
-+static inline void setup_page_debug(struct kmem_cache *s,
-+ void *addr, int order) {}
-
- static inline int alloc_debug_processing(struct kmem_cache *s,
- struct page *page, void *object, unsigned long addr) { return 0; }
-@@ -1602,12 +1614,11 @@ static struct page *allocate_slab(struct kmem_cache *s, gfp_t flags, int node)
- if (page_is_pfmemalloc(page))
- SetPageSlabPfmemalloc(page);
-
-- start = page_address(page);
-+ kasan_poison_slab(page);
-
-- if (unlikely(s->flags & SLAB_POISON))
-- memset(start, POISON_INUSE, PAGE_SIZE << order);
-+ start = page_address(page);
-
-- kasan_poison_slab(page);
-+ setup_page_debug(s, start, order);
-
- shuffle = shuffle_freelist(s, page);
-
---
-2.19.1
-
arm64-relax-gic-version-check-during-early-boot.patch
arm-tegra-restore-dt-abi-on-tegra124-chromebooks.patch
net-marvell-mvneta-fix-dma-debug-warning.patch
-kasan-slub-move-kasan_poison_slab-hook-before-page_a.patch
mm-handle-lru_add_drain_all-for-up-properly.patch
tmpfs-fix-link-accounting-when-a-tmpfile-is-linked-i.patch
-kasan-slab-fix-conflicts-with-config_hardened_userco.patch
ixgbe-fix-older-devices-that-do-not-support-ixgbe_mr.patch
arcv2-lib-memcpy-fix-doing-prefetchw-outside-of-buff.patch
arc-uacces-remove-lp_start-lp_end-from-clobber-list.patch
+++ /dev/null
-From 6d7a655f67ed6a63372e82e343c84daff705542b Mon Sep 17 00:00:00 2001
-From: Andrey Konovalov <andreyknvl@google.com>
-Date: Wed, 20 Feb 2019 22:20:25 -0800
-Subject: kasan, slab: fix conflicts with CONFIG_HARDENED_USERCOPY
-
-[ Upstream commit 219667c23c68eb3dbc0d5662b9246f28477fe529 ]
-
-Similarly to commit 96fedce27e13 ("kasan: make tag based mode work with
-CONFIG_HARDENED_USERCOPY"), we need to reset pointer tags in
-__check_heap_object() in mm/slab.c before doing any pointer math.
-
-Link: http://lkml.kernel.org/r/9a5c0f958db10e69df5ff9f2b997866b56b7effc.1550602886.git.andreyknvl@google.com
-Signed-off-by: Andrey Konovalov <andreyknvl@google.com>
-Tested-by: Qian Cai <cai@lca.pw>
-Cc: Alexander Potapenko <glider@google.com>
-Cc: Andrey Ryabinin <aryabinin@virtuozzo.com>
-Cc: Catalin Marinas <catalin.marinas@arm.com>
-Cc: Dmitry Vyukov <dvyukov@google.com>
-Cc: Evgeniy Stepanov <eugenis@google.com>
-Cc: Kostya Serebryany <kcc@google.com>
-Cc: Vincenzo Frascino <vincenzo.frascino@arm.com>
-Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
-Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-Signed-off-by: Sasha Levin <sashal@kernel.org>
----
- mm/slab.c | 2 ++
- 1 file changed, 2 insertions(+)
-
-diff --git a/mm/slab.c b/mm/slab.c
-index 354a09deecff..b30b58de793b 100644
---- a/mm/slab.c
-+++ b/mm/slab.c
-@@ -4461,6 +4461,8 @@ const char *__check_heap_object(const void *ptr, unsigned long n,
- unsigned int objnr;
- unsigned long offset;
-
-+ ptr = kasan_reset_tag(ptr);
-+
- /* Find and validate object. */
- cachep = page->slab_cache;
- objnr = obj_to_index(cachep, page, (void *)ptr);
---
-2.19.1
-
+++ /dev/null
-From fc44c0fdcc0d414fadd94ea09d4cabf0cf8fa507 Mon Sep 17 00:00:00 2001
-From: Andrey Konovalov <andreyknvl@google.com>
-Date: Wed, 20 Feb 2019 22:19:23 -0800
-Subject: kasan, slub: move kasan_poison_slab hook before page_address
-
-[ Upstream commit a71012242837fe5e67d8c999cfc357174ed5dba0 ]
-
-With tag based KASAN page_address() looks at the page flags to see whether
-the resulting pointer needs to have a tag set. Since we don't want to set
-a tag when page_address() is called on SLAB pages, we call
-page_kasan_tag_reset() in kasan_poison_slab(). However in allocate_slab()
-page_address() is called before kasan_poison_slab(). Fix it by changing
-the order.
-
-[andreyknvl@google.com: fix compilation error when CONFIG_SLUB_DEBUG=n]
- Link: http://lkml.kernel.org/r/ac27cc0bbaeb414ed77bcd6671a877cf3546d56e.1550066133.git.andreyknvl@google.com
-Link: http://lkml.kernel.org/r/cd895d627465a3f1c712647072d17f10883be2a1.1549921721.git.andreyknvl@google.com
-Signed-off-by: Andrey Konovalov <andreyknvl@google.com>
-Cc: Alexander Potapenko <glider@google.com>
-Cc: Andrey Ryabinin <aryabinin@virtuozzo.com>
-Cc: Catalin Marinas <catalin.marinas@arm.com>
-Cc: Christoph Lameter <cl@linux.com>
-Cc: David Rientjes <rientjes@google.com>
-Cc: Dmitry Vyukov <dvyukov@google.com>
-Cc: Evgeniy Stepanov <eugenis@google.com>
-Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
-Cc: Kostya Serebryany <kcc@google.com>
-Cc: Pekka Enberg <penberg@kernel.org>
-Cc: Qian Cai <cai@lca.pw>
-Cc: Vincenzo Frascino <vincenzo.frascino@arm.com>
-Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
-Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-Signed-off-by: Sasha Levin <sashal@kernel.org>
----
- mm/slub.c | 19 +++++++++++++++----
- 1 file changed, 15 insertions(+), 4 deletions(-)
-
-diff --git a/mm/slub.c b/mm/slub.c
-index 131dee87a67c..979400b1a781 100644
---- a/mm/slub.c
-+++ b/mm/slub.c
-@@ -1052,6 +1052,16 @@ static void setup_object_debug(struct kmem_cache *s, struct page *page,
- init_tracking(s, object);
- }
-
-+static void setup_page_debug(struct kmem_cache *s, void *addr, int order)
-+{
-+ if (!(s->flags & SLAB_POISON))
-+ return;
-+
-+ metadata_access_enable();
-+ memset(addr, POISON_INUSE, PAGE_SIZE << order);
-+ metadata_access_disable();
-+}
-+
- static inline int alloc_consistency_checks(struct kmem_cache *s,
- struct page *page,
- void *object, unsigned long addr)
-@@ -1269,6 +1279,8 @@ unsigned long kmem_cache_flags(unsigned long object_size,
- #else /* !CONFIG_SLUB_DEBUG */
- static inline void setup_object_debug(struct kmem_cache *s,
- struct page *page, void *object) {}
-+static inline void setup_page_debug(struct kmem_cache *s,
-+ void *addr, int order) {}
-
- static inline int alloc_debug_processing(struct kmem_cache *s,
- struct page *page, void *object, unsigned long addr) { return 0; }
-@@ -1584,12 +1596,11 @@ static struct page *allocate_slab(struct kmem_cache *s, gfp_t flags, int node)
- if (page_is_pfmemalloc(page))
- SetPageSlabPfmemalloc(page);
-
-- start = page_address(page);
-+ kasan_poison_slab(page);
-
-- if (unlikely(s->flags & SLAB_POISON))
-- memset(start, POISON_INUSE, PAGE_SIZE << order);
-+ start = page_address(page);
-
-- kasan_poison_slab(page);
-+ setup_page_debug(s, start, order);
-
- shuffle = shuffle_freelist(s, page);
-
---
-2.19.1
-
asoc-topology-free-created-components-in-tplg-load-e.patch
arm64-relax-gic-version-check-during-early-boot.patch
net-marvell-mvneta-fix-dma-debug-warning.patch
-kasan-slub-move-kasan_poison_slab-hook-before-page_a.patch
tmpfs-fix-link-accounting-when-a-tmpfile-is-linked-i.patch
-kasan-slab-fix-conflicts-with-config_hardened_userco.patch
arcv2-lib-memcpy-fix-doing-prefetchw-outside-of-buff.patch
arc-uacces-remove-lp_start-lp_end-from-clobber-list.patch
phonet-fix-building-with-clang.patch