diff --git a/llvm/test/CodeGen/X86/vec_fptrunc.ll b/llvm/test/CodeGen/X86/vec_fptrunc.ll index c2f418680e7c5..5b2dafaf3ac28 100644 --- a/llvm/test/CodeGen/X86/vec_fptrunc.ll +++ b/llvm/test/CodeGen/X86/vec_fptrunc.ll @@ -1,25 +1,25 @@ ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py -; RUN: llc < %s -mtriple=i686-unknown -mattr=+sse4.1 | FileCheck %s --check-prefix=X32-SSE -; RUN: llc < %s -mtriple=i686-unknown -mattr=+avx | FileCheck %s --check-prefix=X32-AVX +; RUN: llc < %s -mtriple=i686-unknown -mattr=+sse4.1 | FileCheck %s --check-prefix=X86-SSE +; RUN: llc < %s -mtriple=i686-unknown -mattr=+avx | FileCheck %s --check-prefix=X86-AVX ; RUN: llc < %s -mtriple=x86_64-unknown -mattr=+sse4.1 | FileCheck %s --check-prefix=X64-SSE ; RUN: llc < %s -mtriple=x86_64-unknown -mattr=+avx | FileCheck %s --check-prefix=X64-AVX define void @fptrunc_frommem2(ptr %in, ptr %out) { -; X32-SSE-LABEL: fptrunc_frommem2: -; X32-SSE: # %bb.0: # %entry -; X32-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax -; X32-SSE-NEXT: movl {{[0-9]+}}(%esp), %ecx -; X32-SSE-NEXT: cvtpd2ps (%ecx), %xmm0 -; X32-SSE-NEXT: movlpd %xmm0, (%eax) -; X32-SSE-NEXT: retl -; -; X32-AVX-LABEL: fptrunc_frommem2: -; X32-AVX: # %bb.0: # %entry -; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax -; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %ecx -; X32-AVX-NEXT: vcvtpd2psx (%ecx), %xmm0 -; X32-AVX-NEXT: vmovlpd %xmm0, (%eax) -; X32-AVX-NEXT: retl +; X86-SSE-LABEL: fptrunc_frommem2: +; X86-SSE: # %bb.0: # %entry +; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %ecx +; X86-SSE-NEXT: cvtpd2ps (%ecx), %xmm0 +; X86-SSE-NEXT: movlpd %xmm0, (%eax) +; X86-SSE-NEXT: retl +; +; X86-AVX-LABEL: fptrunc_frommem2: +; X86-AVX: # %bb.0: # %entry +; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %ecx +; X86-AVX-NEXT: vcvtpd2psx (%ecx), %xmm0 +; X86-AVX-NEXT: vmovlpd %xmm0, (%eax) +; X86-AVX-NEXT: retl ; ; X64-SSE-LABEL: fptrunc_frommem2: ; X64-SSE: # %bb.0: # %entry @@ -40,23 +40,23 @@ entry: } define void @fptrunc_frommem4(ptr %in, ptr %out) { -; X32-SSE-LABEL: fptrunc_frommem4: -; X32-SSE: # %bb.0: # %entry -; X32-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax -; X32-SSE-NEXT: movl {{[0-9]+}}(%esp), %ecx -; X32-SSE-NEXT: cvtpd2ps 16(%ecx), %xmm0 -; X32-SSE-NEXT: cvtpd2ps (%ecx), %xmm1 -; X32-SSE-NEXT: unpcklpd {{.*#+}} xmm1 = xmm1[0],xmm0[0] -; X32-SSE-NEXT: movupd %xmm1, (%eax) -; X32-SSE-NEXT: retl -; -; X32-AVX-LABEL: fptrunc_frommem4: -; X32-AVX: # %bb.0: # %entry -; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax -; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %ecx -; X32-AVX-NEXT: vcvtpd2psy (%ecx), %xmm0 -; X32-AVX-NEXT: vmovupd %xmm0, (%eax) -; X32-AVX-NEXT: retl +; X86-SSE-LABEL: fptrunc_frommem4: +; X86-SSE: # %bb.0: # %entry +; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %ecx +; X86-SSE-NEXT: cvtpd2ps 16(%ecx), %xmm0 +; X86-SSE-NEXT: cvtpd2ps (%ecx), %xmm1 +; X86-SSE-NEXT: unpcklpd {{.*#+}} xmm1 = xmm1[0],xmm0[0] +; X86-SSE-NEXT: movupd %xmm1, (%eax) +; X86-SSE-NEXT: retl +; +; X86-AVX-LABEL: fptrunc_frommem4: +; X86-AVX: # %bb.0: # %entry +; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %ecx +; X86-AVX-NEXT: vcvtpd2psy (%ecx), %xmm0 +; X86-AVX-NEXT: vmovupd %xmm0, (%eax) +; X86-AVX-NEXT: retl ; ; X64-SSE-LABEL: fptrunc_frommem4: ; X64-SSE: # %bb.0: # %entry @@ -79,29 +79,29 @@ entry: } define void @fptrunc_frommem8(ptr %in, ptr %out) { -; X32-SSE-LABEL: fptrunc_frommem8: -; X32-SSE: # %bb.0: # %entry -; X32-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax -; X32-SSE-NEXT: movl {{[0-9]+}}(%esp), %ecx -; X32-SSE-NEXT: cvtpd2ps 16(%ecx), %xmm0 -; X32-SSE-NEXT: cvtpd2ps (%ecx), %xmm1 -; X32-SSE-NEXT: unpcklpd {{.*#+}} xmm1 = xmm1[0],xmm0[0] -; X32-SSE-NEXT: cvtpd2ps 48(%ecx), %xmm0 -; X32-SSE-NEXT: cvtpd2ps 32(%ecx), %xmm2 -; X32-SSE-NEXT: unpcklpd {{.*#+}} xmm2 = xmm2[0],xmm0[0] -; X32-SSE-NEXT: movupd %xmm2, 16(%eax) -; X32-SSE-NEXT: movupd %xmm1, (%eax) -; X32-SSE-NEXT: retl -; -; X32-AVX-LABEL: fptrunc_frommem8: -; X32-AVX: # %bb.0: # %entry -; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax -; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %ecx -; X32-AVX-NEXT: vcvtpd2psy (%ecx), %xmm0 -; X32-AVX-NEXT: vcvtpd2psy 32(%ecx), %xmm1 -; X32-AVX-NEXT: vmovupd %xmm1, 16(%eax) -; X32-AVX-NEXT: vmovupd %xmm0, (%eax) -; X32-AVX-NEXT: retl +; X86-SSE-LABEL: fptrunc_frommem8: +; X86-SSE: # %bb.0: # %entry +; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %ecx +; X86-SSE-NEXT: cvtpd2ps 16(%ecx), %xmm0 +; X86-SSE-NEXT: cvtpd2ps (%ecx), %xmm1 +; X86-SSE-NEXT: unpcklpd {{.*#+}} xmm1 = xmm1[0],xmm0[0] +; X86-SSE-NEXT: cvtpd2ps 48(%ecx), %xmm0 +; X86-SSE-NEXT: cvtpd2ps 32(%ecx), %xmm2 +; X86-SSE-NEXT: unpcklpd {{.*#+}} xmm2 = xmm2[0],xmm0[0] +; X86-SSE-NEXT: movupd %xmm2, 16(%eax) +; X86-SSE-NEXT: movupd %xmm1, (%eax) +; X86-SSE-NEXT: retl +; +; X86-AVX-LABEL: fptrunc_frommem8: +; X86-AVX: # %bb.0: # %entry +; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %ecx +; X86-AVX-NEXT: vcvtpd2psy (%ecx), %xmm0 +; X86-AVX-NEXT: vcvtpd2psy 32(%ecx), %xmm1 +; X86-AVX-NEXT: vmovupd %xmm1, 16(%eax) +; X86-AVX-NEXT: vmovupd %xmm0, (%eax) +; X86-AVX-NEXT: retl ; ; X64-SSE-LABEL: fptrunc_frommem8: ; X64-SSE: # %bb.0: # %entry @@ -130,17 +130,17 @@ entry: } define <4 x float> @fptrunc_frommem2_zext(ptr %ld) { -; X32-SSE-LABEL: fptrunc_frommem2_zext: -; X32-SSE: # %bb.0: -; X32-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax -; X32-SSE-NEXT: cvtpd2ps (%eax), %xmm0 -; X32-SSE-NEXT: retl -; -; X32-AVX-LABEL: fptrunc_frommem2_zext: -; X32-AVX: # %bb.0: -; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax -; X32-AVX-NEXT: vcvtpd2psx (%eax), %xmm0 -; X32-AVX-NEXT: retl +; X86-SSE-LABEL: fptrunc_frommem2_zext: +; X86-SSE: # %bb.0: +; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-SSE-NEXT: cvtpd2ps (%eax), %xmm0 +; X86-SSE-NEXT: retl +; +; X86-AVX-LABEL: fptrunc_frommem2_zext: +; X86-AVX: # %bb.0: +; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-AVX-NEXT: vcvtpd2psx (%eax), %xmm0 +; X86-AVX-NEXT: retl ; ; X64-SSE-LABEL: fptrunc_frommem2_zext: ; X64-SSE: # %bb.0: @@ -158,15 +158,15 @@ define <4 x float> @fptrunc_frommem2_zext(ptr %ld) { } define <4 x float> @fptrunc_fromreg2_zext(<2 x double> %arg) { -; X32-SSE-LABEL: fptrunc_fromreg2_zext: -; X32-SSE: # %bb.0: -; X32-SSE-NEXT: cvtpd2ps %xmm0, %xmm0 -; X32-SSE-NEXT: retl +; X86-SSE-LABEL: fptrunc_fromreg2_zext: +; X86-SSE: # %bb.0: +; X86-SSE-NEXT: cvtpd2ps %xmm0, %xmm0 +; X86-SSE-NEXT: retl ; -; X32-AVX-LABEL: fptrunc_fromreg2_zext: -; X32-AVX: # %bb.0: -; X32-AVX-NEXT: vcvtpd2ps %xmm0, %xmm0 -; X32-AVX-NEXT: retl +; X86-AVX-LABEL: fptrunc_fromreg2_zext: +; X86-AVX: # %bb.0: +; X86-AVX-NEXT: vcvtpd2ps %xmm0, %xmm0 +; X86-AVX-NEXT: retl ; ; X64-SSE-LABEL: fptrunc_fromreg2_zext: ; X64-SSE: # %bb.0: @@ -184,17 +184,17 @@ define <4 x float> @fptrunc_fromreg2_zext(<2 x double> %arg) { ; FIXME: For exact truncations we should be able to fold this. define <4 x float> @fptrunc_fromconst() { -; X32-SSE-LABEL: fptrunc_fromconst: -; X32-SSE: # %bb.0: # %entry -; X32-SSE-NEXT: cvtpd2ps {{\.?LCPI[0-9]+_[0-9]+}}, %xmm1 -; X32-SSE-NEXT: cvtpd2ps {{\.?LCPI[0-9]+_[0-9]+}}, %xmm0 -; X32-SSE-NEXT: unpcklpd {{.*#+}} xmm0 = xmm0[0],xmm1[0] -; X32-SSE-NEXT: retl -; -; X32-AVX-LABEL: fptrunc_fromconst: -; X32-AVX: # %bb.0: # %entry -; X32-AVX-NEXT: vcvtpd2psy {{\.?LCPI[0-9]+_[0-9]+}}, %xmm0 -; X32-AVX-NEXT: retl +; X86-SSE-LABEL: fptrunc_fromconst: +; X86-SSE: # %bb.0: # %entry +; X86-SSE-NEXT: cvtpd2ps {{\.?LCPI[0-9]+_[0-9]+}}, %xmm1 +; X86-SSE-NEXT: cvtpd2ps {{\.?LCPI[0-9]+_[0-9]+}}, %xmm0 +; X86-SSE-NEXT: unpcklpd {{.*#+}} xmm0 = xmm0[0],xmm1[0] +; X86-SSE-NEXT: retl +; +; X86-AVX-LABEL: fptrunc_fromconst: +; X86-AVX: # %bb.0: # %entry +; X86-AVX-NEXT: vcvtpd2psy {{\.?LCPI[0-9]+_[0-9]+}}, %xmm0 +; X86-AVX-NEXT: retl ; ; X64-SSE-LABEL: fptrunc_fromconst: ; X64-SSE: # %bb.0: # %entry