From: H.J. Lu Date: Tue, 23 Aug 2022 20:32:01 +0000 (-0700) Subject: x86: Replace vmovdqu with movdqu in BF16 XMM ABI tests X-Git-Tag: basepoints/gcc-14~5046 X-Git-Url: http://git.ipfire.org/?a=commitdiff_plain;h=4ade41de1a6c80db6bb61399da6bff2126813d75;p=thirdparty%2Fgcc.git x86: Replace vmovdqu with movdqu in BF16 XMM ABI tests Since XMM BF16 tests only require SSE2, replace vmovdqu with movdqu in BF16 XMM ABI tests to support SSE2 machines without AVX. Tested on x86-64 machines with and without AVX. * gcc.target/x86_64/abi/bf16/asm-support.S: Replace vmovdqu with movdqu. --- diff --git a/gcc/testsuite/gcc.target/x86_64/abi/bf16/asm-support.S b/gcc/testsuite/gcc.target/x86_64/abi/bf16/asm-support.S index a8165d86317c..7559aa910c4d 100644 --- a/gcc/testsuite/gcc.target/x86_64/abi/bf16/asm-support.S +++ b/gcc/testsuite/gcc.target/x86_64/abi/bf16/asm-support.S @@ -20,22 +20,22 @@ snapshot: movq %r13, r13(%rip) movq %r14, r14(%rip) movq %r15, r15(%rip) - vmovdqu %xmm0, xmm_regs+0(%rip) - vmovdqu %xmm1, xmm_regs+16(%rip) - vmovdqu %xmm2, xmm_regs+32(%rip) - vmovdqu %xmm3, xmm_regs+48(%rip) - vmovdqu %xmm4, xmm_regs+64(%rip) - vmovdqu %xmm5, xmm_regs+80(%rip) - vmovdqu %xmm6, xmm_regs+96(%rip) - vmovdqu %xmm7, xmm_regs+112(%rip) - vmovdqu %xmm8, xmm_regs+128(%rip) - vmovdqu %xmm9, xmm_regs+144(%rip) - vmovdqu %xmm10, xmm_regs+160(%rip) - vmovdqu %xmm11, xmm_regs+176(%rip) - vmovdqu %xmm12, xmm_regs+192(%rip) - vmovdqu %xmm13, xmm_regs+208(%rip) - vmovdqu %xmm14, xmm_regs+224(%rip) - vmovdqu %xmm15, xmm_regs+240(%rip) + movdqu %xmm0, xmm_regs+0(%rip) + movdqu %xmm1, xmm_regs+16(%rip) + movdqu %xmm2, xmm_regs+32(%rip) + movdqu %xmm3, xmm_regs+48(%rip) + movdqu %xmm4, xmm_regs+64(%rip) + movdqu %xmm5, xmm_regs+80(%rip) + movdqu %xmm6, xmm_regs+96(%rip) + movdqu %xmm7, xmm_regs+112(%rip) + movdqu %xmm8, xmm_regs+128(%rip) + movdqu %xmm9, xmm_regs+144(%rip) + movdqu %xmm10, xmm_regs+160(%rip) + movdqu %xmm11, xmm_regs+176(%rip) + movdqu %xmm12, xmm_regs+192(%rip) + movdqu %xmm13, xmm_regs+208(%rip) + movdqu %xmm14, xmm_regs+224(%rip) + movdqu %xmm15, xmm_regs+240(%rip) jmp *callthis(%rip) .LFE3: .size snapshot, .-snapshot @@ -50,8 +50,8 @@ snapshot_ret: addq $8, %rsp movq %rax, rax(%rip) movq %rdx, rdx(%rip) - vmovdqu %xmm0, xmm_regs+0(%rip) - vmovdqu %xmm1, xmm_regs+16(%rip) + movdqu %xmm0, xmm_regs+0(%rip) + movdqu %xmm1, xmm_regs+16(%rip) fstpt x87_regs(%rip) fstpt x87_regs+16(%rip) fldt x87_regs+16(%rip)