]> git.ipfire.org Git - thirdparty/kernel/stable-queue.git/blob
ed1d040f478559e3c6a7e8c337aadf567a1919ff
[thirdparty/kernel/stable-queue.git] /
1 From foo@baz Tue Jul 12 05:07:35 PM CEST 2022
2 From: Arnaldo Carvalho de Melo <acme@redhat.com>
3 Date: Sun, 9 May 2021 10:19:37 -0300
4 Subject: tools arch: Update arch/x86/lib/mem{cpy,set}_64.S copies used in 'perf bench mem memcpy'
5
6 From: Arnaldo Carvalho de Melo <acme@redhat.com>
7
8 commit 35cb8c713a496e8c114eed5e2a5a30b359876df2 upstream.
9
10 To bring in the change made in this cset:
11
12 f94909ceb1ed4bfd ("x86: Prepare asm files for straight-line-speculation")
13
14 It silences these perf tools build warnings, no change in the tools:
15
16 Warning: Kernel ABI header at 'tools/arch/x86/lib/memcpy_64.S' differs from latest version at 'arch/x86/lib/memcpy_64.S'
17 diff -u tools/arch/x86/lib/memcpy_64.S arch/x86/lib/memcpy_64.S
18 Warning: Kernel ABI header at 'tools/arch/x86/lib/memset_64.S' differs from latest version at 'arch/x86/lib/memset_64.S'
19 diff -u tools/arch/x86/lib/memset_64.S arch/x86/lib/memset_64.S
20
21 The code generated was checked before and after using 'objdump -d /tmp/build/perf/bench/mem-memcpy-x86-64-asm.o',
22 no changes.
23
24 Cc: Borislav Petkov <bp@suse.de>
25 Cc: Peter Zijlstra <peterz@infradead.org>
26 Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
27 Signed-off-by: Sasha Levin <sashal@kernel.org>
28 Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
29 Signed-off-by: Ben Hutchings <ben@decadent.org.uk>
30 Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
31 ---
32 tools/arch/x86/lib/memcpy_64.S | 12 ++++++------
33 tools/arch/x86/lib/memset_64.S | 6 +++---
34 2 files changed, 9 insertions(+), 9 deletions(-)
35
36 --- a/tools/arch/x86/lib/memcpy_64.S
37 +++ b/tools/arch/x86/lib/memcpy_64.S
38 @@ -39,7 +39,7 @@ SYM_FUNC_START_WEAK(memcpy)
39 rep movsq
40 movl %edx, %ecx
41 rep movsb
42 - ret
43 + RET
44 SYM_FUNC_END(memcpy)
45 SYM_FUNC_END_ALIAS(__memcpy)
46 EXPORT_SYMBOL(memcpy)
47 @@ -53,7 +53,7 @@ SYM_FUNC_START_LOCAL(memcpy_erms)
48 movq %rdi, %rax
49 movq %rdx, %rcx
50 rep movsb
51 - ret
52 + RET
53 SYM_FUNC_END(memcpy_erms)
54
55 SYM_FUNC_START_LOCAL(memcpy_orig)
56 @@ -137,7 +137,7 @@ SYM_FUNC_START_LOCAL(memcpy_orig)
57 movq %r9, 1*8(%rdi)
58 movq %r10, -2*8(%rdi, %rdx)
59 movq %r11, -1*8(%rdi, %rdx)
60 - retq
61 + RET
62 .p2align 4
63 .Lless_16bytes:
64 cmpl $8, %edx
65 @@ -149,7 +149,7 @@ SYM_FUNC_START_LOCAL(memcpy_orig)
66 movq -1*8(%rsi, %rdx), %r9
67 movq %r8, 0*8(%rdi)
68 movq %r9, -1*8(%rdi, %rdx)
69 - retq
70 + RET
71 .p2align 4
72 .Lless_8bytes:
73 cmpl $4, %edx
74 @@ -162,7 +162,7 @@ SYM_FUNC_START_LOCAL(memcpy_orig)
75 movl -4(%rsi, %rdx), %r8d
76 movl %ecx, (%rdi)
77 movl %r8d, -4(%rdi, %rdx)
78 - retq
79 + RET
80 .p2align 4
81 .Lless_3bytes:
82 subl $1, %edx
83 @@ -180,7 +180,7 @@ SYM_FUNC_START_LOCAL(memcpy_orig)
84 movb %cl, (%rdi)
85
86 .Lend:
87 - retq
88 + RET
89 SYM_FUNC_END(memcpy_orig)
90
91 .popsection
92 --- a/tools/arch/x86/lib/memset_64.S
93 +++ b/tools/arch/x86/lib/memset_64.S
94 @@ -40,7 +40,7 @@ SYM_FUNC_START(__memset)
95 movl %edx,%ecx
96 rep stosb
97 movq %r9,%rax
98 - ret
99 + RET
100 SYM_FUNC_END(__memset)
101 SYM_FUNC_END_ALIAS(memset)
102 EXPORT_SYMBOL(memset)
103 @@ -63,7 +63,7 @@ SYM_FUNC_START_LOCAL(memset_erms)
104 movq %rdx,%rcx
105 rep stosb
106 movq %r9,%rax
107 - ret
108 + RET
109 SYM_FUNC_END(memset_erms)
110
111 SYM_FUNC_START_LOCAL(memset_orig)
112 @@ -125,7 +125,7 @@ SYM_FUNC_START_LOCAL(memset_orig)
113
114 .Lende:
115 movq %r10,%rax
116 - ret
117 + RET
118
119 .Lbad_alignment:
120 cmpq $7,%rdx