diff --git a/llvm/test/CodeGen/Thumb2/mve-vmaxv.ll b/llvm/test/CodeGen/Thumb2/mve-vmaxv.ll index f96c2f422a3fa..36c201cced56c 100644 --- a/llvm/test/CodeGen/Thumb2/mve-vmaxv.ll +++ b/llvm/test/CodeGen/Thumb2/mve-vmaxv.ll @@ -14,8 +14,8 @@ declare i8 @llvm.experimental.vector.reduce.umin.v16i8(<16 x i8>) declare i16 @llvm.experimental.vector.reduce.umin.v8i16(<8 x i16>) declare i32 @llvm.experimental.vector.reduce.umin.v4i32(<4 x i32>) -define arm_aapcs_vfpcc i8 @vmaxv_s_v16i8_i32(<16 x i8> %s1) { -; CHECK-LABEL: vmaxv_s_v16i8_i32: +define arm_aapcs_vfpcc i8 @vmaxv_s_v16i8(<16 x i8> %s1) { +; CHECK-LABEL: vmaxv_s_v16i8: ; CHECK: @ %bb.0: ; CHECK-NEXT: mvn r0, #127 ; CHECK-NEXT: vmaxv.s8 r0, q0 @@ -24,8 +24,8 @@ define arm_aapcs_vfpcc i8 @vmaxv_s_v16i8_i32(<16 x i8> %s1) { ret i8 %r } -define arm_aapcs_vfpcc i16 @vmaxv_s_v8i16_i32(<8 x i16> %s1) { -; CHECK-LABEL: vmaxv_s_v8i16_i32: +define arm_aapcs_vfpcc i16 @vmaxv_s_v8i16(<8 x i16> %s1) { +; CHECK-LABEL: vmaxv_s_v8i16: ; CHECK: @ %bb.0: ; CHECK-NEXT: movw r0, #32768 ; CHECK-NEXT: movt r0, #65535 @@ -35,8 +35,8 @@ define arm_aapcs_vfpcc i16 @vmaxv_s_v8i16_i32(<8 x i16> %s1) { ret i16 %r } -define arm_aapcs_vfpcc i32 @vmaxv_s_v4i32_i32(<4 x i32> %s1) { -; CHECK-LABEL: vmaxv_s_v4i32_i32: +define arm_aapcs_vfpcc i32 @vmaxv_s_v4i32(<4 x i32> %s1) { +; CHECK-LABEL: vmaxv_s_v4i32: ; CHECK: @ %bb.0: ; CHECK-NEXT: mov.w r0, #-2147483648 ; CHECK-NEXT: vmaxv.s32 r0, q0 @@ -45,8 +45,8 @@ define arm_aapcs_vfpcc i32 @vmaxv_s_v4i32_i32(<4 x i32> %s1) { ret i32 %r } -define arm_aapcs_vfpcc i8 @vmaxv_u_v16i8_i32(<16 x i8> %s1) { -; CHECK-LABEL: vmaxv_u_v16i8_i32: +define arm_aapcs_vfpcc i8 @vmaxv_u_v16i8(<16 x i8> %s1) { +; CHECK-LABEL: vmaxv_u_v16i8: ; CHECK: @ %bb.0: ; CHECK-NEXT: movs r0, #0 ; CHECK-NEXT: vmaxv.u8 r0, q0 @@ -55,8 +55,8 @@ define arm_aapcs_vfpcc i8 @vmaxv_u_v16i8_i32(<16 x i8> %s1) { ret i8 %r } -define arm_aapcs_vfpcc i16 @vmaxv_u_v8i16_i32(<8 x i16> %s1) { -; CHECK-LABEL: vmaxv_u_v8i16_i32: +define arm_aapcs_vfpcc i16 @vmaxv_u_v8i16(<8 x i16> %s1) { +; CHECK-LABEL: vmaxv_u_v8i16: ; CHECK: @ %bb.0: ; CHECK-NEXT: movs r0, #0 ; CHECK-NEXT: vmaxv.u16 r0, q0 @@ -65,8 +65,8 @@ define arm_aapcs_vfpcc i16 @vmaxv_u_v8i16_i32(<8 x i16> %s1) { ret i16 %r } -define arm_aapcs_vfpcc i32 @vmaxv_u_v4i32_i32(<4 x i32> %s1) { -; CHECK-LABEL: vmaxv_u_v4i32_i32: +define arm_aapcs_vfpcc i32 @vmaxv_u_v4i32(<4 x i32> %s1) { +; CHECK-LABEL: vmaxv_u_v4i32: ; CHECK: @ %bb.0: ; CHECK-NEXT: movs r0, #0 ; CHECK-NEXT: vmaxv.u32 r0, q0 @@ -75,8 +75,8 @@ define arm_aapcs_vfpcc i32 @vmaxv_u_v4i32_i32(<4 x i32> %s1) { ret i32 %r } -define arm_aapcs_vfpcc i8 @vminv_s_v16i8_i32(<16 x i8> %s1) { -; CHECK-LABEL: vminv_s_v16i8_i32: +define arm_aapcs_vfpcc i8 @vminv_s_v16i8(<16 x i8> %s1) { +; CHECK-LABEL: vminv_s_v16i8: ; CHECK: @ %bb.0: ; CHECK-NEXT: movs r0, #127 ; CHECK-NEXT: vminv.s8 r0, q0 @@ -85,8 +85,8 @@ define arm_aapcs_vfpcc i8 @vminv_s_v16i8_i32(<16 x i8> %s1) { ret i8 %r } -define arm_aapcs_vfpcc i16 @vminv_s_v8i16_i32(<8 x i16> %s1) { -; CHECK-LABEL: vminv_s_v8i16_i32: +define arm_aapcs_vfpcc i16 @vminv_s_v8i16(<8 x i16> %s1) { +; CHECK-LABEL: vminv_s_v8i16: ; CHECK: @ %bb.0: ; CHECK-NEXT: movw r0, #32767 ; CHECK-NEXT: vminv.s16 r0, q0 @@ -95,8 +95,8 @@ define arm_aapcs_vfpcc i16 @vminv_s_v8i16_i32(<8 x i16> %s1) { ret i16 %r } -define arm_aapcs_vfpcc i32 @vminv_s_v4i32_i32(<4 x i32> %s1) { -; CHECK-LABEL: vminv_s_v4i32_i32: +define arm_aapcs_vfpcc i32 @vminv_s_v4i32(<4 x i32> %s1) { +; CHECK-LABEL: vminv_s_v4i32: ; CHECK: @ %bb.0: ; CHECK-NEXT: mvn r0, #-2147483648 ; CHECK-NEXT: vminv.s32 r0, q0 @@ -105,8 +105,8 @@ define arm_aapcs_vfpcc i32 @vminv_s_v4i32_i32(<4 x i32> %s1) { ret i32 %r } -define arm_aapcs_vfpcc i8 @vminv_u_v16i8_i32(<16 x i8> %s1) { -; CHECK-LABEL: vminv_u_v16i8_i32: +define arm_aapcs_vfpcc i8 @vminv_u_v16i8(<16 x i8> %s1) { +; CHECK-LABEL: vminv_u_v16i8: ; CHECK: @ %bb.0: ; CHECK-NEXT: movs r0, #255 ; CHECK-NEXT: vminv.u8 r0, q0 @@ -115,8 +115,8 @@ define arm_aapcs_vfpcc i8 @vminv_u_v16i8_i32(<16 x i8> %s1) { ret i8 %r } -define arm_aapcs_vfpcc i16 @vminv_u_v8i16_i32(<8 x i16> %s1) { -; CHECK-LABEL: vminv_u_v8i16_i32: +define arm_aapcs_vfpcc i16 @vminv_u_v8i16(<8 x i16> %s1) { +; CHECK-LABEL: vminv_u_v8i16: ; CHECK: @ %bb.0: ; CHECK-NEXT: movw r0, #65535 ; CHECK-NEXT: vminv.u16 r0, q0 @@ -125,8 +125,8 @@ define arm_aapcs_vfpcc i16 @vminv_u_v8i16_i32(<8 x i16> %s1) { ret i16 %r } -define arm_aapcs_vfpcc i32 @vminv_u_v4i32_i32(<4 x i32> %s1) { -; CHECK-LABEL: vminv_u_v4i32_i32: +define arm_aapcs_vfpcc i32 @vminv_u_v4i32(<4 x i32> %s1) { +; CHECK-LABEL: vminv_u_v4i32: ; CHECK: @ %bb.0: ; CHECK-NEXT: mov.w r0, #-1 ; CHECK-NEXT: vminv.u32 r0, q0 @@ -134,3 +134,339 @@ define arm_aapcs_vfpcc i32 @vminv_u_v4i32_i32(<4 x i32> %s1) { %r = call i32 @llvm.experimental.vector.reduce.umin.v4i32(<4 x i32> %s1) ret i32 %r } + + + +define arm_aapcs_vfpcc i8 @vmaxv_s_v16i8_i8(<16 x i8> %s1, i8 %s2) { +; CHECK-LABEL: vmaxv_s_v16i8_i8: +; CHECK: @ %bb.0: +; CHECK-NEXT: mvn r1, #127 +; CHECK-NEXT: sxtb r3, r0 +; CHECK-NEXT: vmaxv.s8 r1, q0 +; CHECK-NEXT: sxtb r2, r1 +; CHECK-NEXT: cmp r2, r3 +; CHECK-NEXT: it gt +; CHECK-NEXT: movgt r0, r1 +; CHECK-NEXT: bx lr + %r = call i8 @llvm.experimental.vector.reduce.smax.v16i8(<16 x i8> %s1) + %c = icmp sgt i8 %r, %s2 + %s = select i1 %c, i8 %r, i8 %s2 + ret i8 %s +} + +define arm_aapcs_vfpcc i32 @vmaxv_s_v16i8_i32(<16 x i8> %s1, i32 %s2) { +; CHECK-LABEL: vmaxv_s_v16i8_i32: +; CHECK: @ %bb.0: +; CHECK-NEXT: mvn r1, #127 +; CHECK-NEXT: vmaxv.s8 r1, q0 +; CHECK-NEXT: sxtb r1, r1 +; CHECK-NEXT: cmp r1, r0 +; CHECK-NEXT: it gt +; CHECK-NEXT: movgt r0, r1 +; CHECK-NEXT: bx lr + %r = call i8 @llvm.experimental.vector.reduce.smax.v16i8(<16 x i8> %s1) + %rs = sext i8 %r to i32 + %c = icmp sgt i32 %rs, %s2 + %s = select i1 %c, i32 %rs, i32 %s2 + ret i32 %s +} + +define arm_aapcs_vfpcc i16 @vmaxv_s_v8i16_i16(<8 x i16> %s1, i16 %s2) { +; CHECK-LABEL: vmaxv_s_v8i16_i16: +; CHECK: @ %bb.0: +; CHECK-NEXT: movw r1, #32768 +; CHECK-NEXT: sxth r3, r0 +; CHECK-NEXT: movt r1, #65535 +; CHECK-NEXT: vmaxv.s16 r1, q0 +; CHECK-NEXT: sxth r2, r1 +; CHECK-NEXT: cmp r2, r3 +; CHECK-NEXT: it gt +; CHECK-NEXT: movgt r0, r1 +; CHECK-NEXT: bx lr + %r = call i16 @llvm.experimental.vector.reduce.smax.v8i16(<8 x i16> %s1) + %c = icmp sgt i16 %r, %s2 + %s = select i1 %c, i16 %r, i16 %s2 + ret i16 %s +} + +define arm_aapcs_vfpcc i32 @vmaxv_s_v8i16_i32(<8 x i16> %s1, i32 %s2) { +; CHECK-LABEL: vmaxv_s_v8i16_i32: +; CHECK: @ %bb.0: +; CHECK-NEXT: movw r1, #32768 +; CHECK-NEXT: movt r1, #65535 +; CHECK-NEXT: vmaxv.s16 r1, q0 +; CHECK-NEXT: sxth r1, r1 +; CHECK-NEXT: cmp r1, r0 +; CHECK-NEXT: it gt +; CHECK-NEXT: movgt r0, r1 +; CHECK-NEXT: bx lr + %r = call i16 @llvm.experimental.vector.reduce.smax.v8i16(<8 x i16> %s1) + %rs = sext i16 %r to i32 + %c = icmp sgt i32 %rs, %s2 + %s = select i1 %c, i32 %rs, i32 %s2 + ret i32 %s +} + +define arm_aapcs_vfpcc i32 @vmaxv_s_v4i32_i32(<4 x i32> %s1, i32 %s2) { +; CHECK-LABEL: vmaxv_s_v4i32_i32: +; CHECK: @ %bb.0: +; CHECK-NEXT: mov.w r1, #-2147483648 +; CHECK-NEXT: vmaxv.s32 r1, q0 +; CHECK-NEXT: cmp r1, r0 +; CHECK-NEXT: it gt +; CHECK-NEXT: movgt r0, r1 +; CHECK-NEXT: bx lr + %r = call i32 @llvm.experimental.vector.reduce.smax.v4i32(<4 x i32> %s1) + %c = icmp sgt i32 %r, %s2 + %s = select i1 %c, i32 %r, i32 %s2 + ret i32 %s +} + +define arm_aapcs_vfpcc i8 @vmaxv_u_v16i8_i8(<16 x i8> %s1, i8 %s2) { +; CHECK-LABEL: vmaxv_u_v16i8_i8: +; CHECK: @ %bb.0: +; CHECK-NEXT: movs r1, #0 +; CHECK-NEXT: uxtb r3, r0 +; CHECK-NEXT: vmaxv.u8 r1, q0 +; CHECK-NEXT: uxtb r2, r1 +; CHECK-NEXT: cmp r2, r3 +; CHECK-NEXT: it hi +; CHECK-NEXT: movhi r0, r1 +; CHECK-NEXT: bx lr + %r = call i8 @llvm.experimental.vector.reduce.umax.v16i8(<16 x i8> %s1) + %c = icmp ugt i8 %r, %s2 + %s = select i1 %c, i8 %r, i8 %s2 + ret i8 %s +} + +define arm_aapcs_vfpcc i32 @vmaxv_u_v16i8_i32(<16 x i8> %s1, i32 %s2) { +; CHECK-LABEL: vmaxv_u_v16i8_i32: +; CHECK: @ %bb.0: +; CHECK-NEXT: movs r1, #0 +; CHECK-NEXT: vmaxv.u8 r1, q0 +; CHECK-NEXT: uxtb r1, r1 +; CHECK-NEXT: cmp r1, r0 +; CHECK-NEXT: it hi +; CHECK-NEXT: movhi r0, r1 +; CHECK-NEXT: bx lr + %r = call i8 @llvm.experimental.vector.reduce.umax.v16i8(<16 x i8> %s1) + %rs = zext i8 %r to i32 + %c = icmp ugt i32 %rs, %s2 + %s = select i1 %c, i32 %rs, i32 %s2 + ret i32 %s +} + +define arm_aapcs_vfpcc i16 @vmaxv_u_v8i16_i16(<8 x i16> %s1, i16 %s2) { +; CHECK-LABEL: vmaxv_u_v8i16_i16: +; CHECK: @ %bb.0: +; CHECK-NEXT: movs r1, #0 +; CHECK-NEXT: uxth r3, r0 +; CHECK-NEXT: vmaxv.u16 r1, q0 +; CHECK-NEXT: uxth r2, r1 +; CHECK-NEXT: cmp r2, r3 +; CHECK-NEXT: it hi +; CHECK-NEXT: movhi r0, r1 +; CHECK-NEXT: bx lr + %r = call i16 @llvm.experimental.vector.reduce.umax.v8i16(<8 x i16> %s1) + %c = icmp ugt i16 %r, %s2 + %s = select i1 %c, i16 %r, i16 %s2 + ret i16 %s +} + +define arm_aapcs_vfpcc i32 @vmaxv_u_v8i16_i32(<8 x i16> %s1, i32 %s2) { +; CHECK-LABEL: vmaxv_u_v8i16_i32: +; CHECK: @ %bb.0: +; CHECK-NEXT: movs r1, #0 +; CHECK-NEXT: vmaxv.u16 r1, q0 +; CHECK-NEXT: uxth r1, r1 +; CHECK-NEXT: cmp r1, r0 +; CHECK-NEXT: it hi +; CHECK-NEXT: movhi r0, r1 +; CHECK-NEXT: bx lr + %r = call i16 @llvm.experimental.vector.reduce.umax.v8i16(<8 x i16> %s1) + %rs = zext i16 %r to i32 + %c = icmp ugt i32 %rs, %s2 + %s = select i1 %c, i32 %rs, i32 %s2 + ret i32 %s +} + +define arm_aapcs_vfpcc i32 @vmaxv_u_v4i32_i32(<4 x i32> %s1, i32 %s2) { +; CHECK-LABEL: vmaxv_u_v4i32_i32: +; CHECK: @ %bb.0: +; CHECK-NEXT: movs r1, #0 +; CHECK-NEXT: vmaxv.u32 r1, q0 +; CHECK-NEXT: cmp r1, r0 +; CHECK-NEXT: it hi +; CHECK-NEXT: movhi r0, r1 +; CHECK-NEXT: bx lr + %r = call i32 @llvm.experimental.vector.reduce.umax.v4i32(<4 x i32> %s1) + %c = icmp ugt i32 %r, %s2 + %s = select i1 %c, i32 %r, i32 %s2 + ret i32 %s +} + +define arm_aapcs_vfpcc i8 @vminv_s_v16i8_i8(<16 x i8> %s1, i8 %s2) { +; CHECK-LABEL: vminv_s_v16i8_i8: +; CHECK: @ %bb.0: +; CHECK-NEXT: movs r1, #127 +; CHECK-NEXT: sxtb r3, r0 +; CHECK-NEXT: vminv.s8 r1, q0 +; CHECK-NEXT: sxtb r2, r1 +; CHECK-NEXT: cmp r2, r3 +; CHECK-NEXT: it lt +; CHECK-NEXT: movlt r0, r1 +; CHECK-NEXT: bx lr + %r = call i8 @llvm.experimental.vector.reduce.smin.v16i8(<16 x i8> %s1) + %c = icmp slt i8 %r, %s2 + %s = select i1 %c, i8 %r, i8 %s2 + ret i8 %s +} + +define arm_aapcs_vfpcc i32 @vminv_s_v16i8_i32(<16 x i8> %s1, i32 %s2) { +; CHECK-LABEL: vminv_s_v16i8_i32: +; CHECK: @ %bb.0: +; CHECK-NEXT: movs r1, #127 +; CHECK-NEXT: vminv.s8 r1, q0 +; CHECK-NEXT: sxtb r1, r1 +; CHECK-NEXT: cmp r1, r0 +; CHECK-NEXT: it lt +; CHECK-NEXT: movlt r0, r1 +; CHECK-NEXT: bx lr + %r = call i8 @llvm.experimental.vector.reduce.smin.v16i8(<16 x i8> %s1) + %rs = sext i8 %r to i32 + %c = icmp slt i32 %rs, %s2 + %s = select i1 %c, i32 %rs, i32 %s2 + ret i32 %s +} + +define arm_aapcs_vfpcc i16 @vminv_s_v8i16_i16(<8 x i16> %s1, i16 %s2) { +; CHECK-LABEL: vminv_s_v8i16_i16: +; CHECK: @ %bb.0: +; CHECK-NEXT: movw r1, #32767 +; CHECK-NEXT: sxth r3, r0 +; CHECK-NEXT: vminv.s16 r1, q0 +; CHECK-NEXT: sxth r2, r1 +; CHECK-NEXT: cmp r2, r3 +; CHECK-NEXT: it lt +; CHECK-NEXT: movlt r0, r1 +; CHECK-NEXT: bx lr + %r = call i16 @llvm.experimental.vector.reduce.smin.v8i16(<8 x i16> %s1) + %c = icmp slt i16 %r, %s2 + %s = select i1 %c, i16 %r, i16 %s2 + ret i16 %s +} + +define arm_aapcs_vfpcc i32 @vminv_s_v8i16_i32(<8 x i16> %s1, i32 %s2) { +; CHECK-LABEL: vminv_s_v8i16_i32: +; CHECK: @ %bb.0: +; CHECK-NEXT: movw r1, #32767 +; CHECK-NEXT: vminv.s16 r1, q0 +; CHECK-NEXT: sxth r1, r1 +; CHECK-NEXT: cmp r1, r0 +; CHECK-NEXT: it lt +; CHECK-NEXT: movlt r0, r1 +; CHECK-NEXT: bx lr + %r = call i16 @llvm.experimental.vector.reduce.smin.v8i16(<8 x i16> %s1) + %rs = sext i16 %r to i32 + %c = icmp slt i32 %rs, %s2 + %s = select i1 %c, i32 %rs, i32 %s2 + ret i32 %s +} + +define arm_aapcs_vfpcc i32 @vminv_s_v4i32_i32(<4 x i32> %s1, i32 %s2) { +; CHECK-LABEL: vminv_s_v4i32_i32: +; CHECK: @ %bb.0: +; CHECK-NEXT: mvn r1, #-2147483648 +; CHECK-NEXT: vminv.s32 r1, q0 +; CHECK-NEXT: cmp r1, r0 +; CHECK-NEXT: it lt +; CHECK-NEXT: movlt r0, r1 +; CHECK-NEXT: bx lr + %r = call i32 @llvm.experimental.vector.reduce.smin.v4i32(<4 x i32> %s1) + %c = icmp slt i32 %r, %s2 + %s = select i1 %c, i32 %r, i32 %s2 + ret i32 %s +} + +define arm_aapcs_vfpcc i8 @vminv_u_v16i8_i8(<16 x i8> %s1, i8 %s2) { +; CHECK-LABEL: vminv_u_v16i8_i8: +; CHECK: @ %bb.0: +; CHECK-NEXT: movs r1, #255 +; CHECK-NEXT: uxtb r3, r0 +; CHECK-NEXT: vminv.u8 r1, q0 +; CHECK-NEXT: uxtb r2, r1 +; CHECK-NEXT: cmp r2, r3 +; CHECK-NEXT: it lo +; CHECK-NEXT: movlo r0, r1 +; CHECK-NEXT: bx lr + %r = call i8 @llvm.experimental.vector.reduce.umin.v16i8(<16 x i8> %s1) + %c = icmp ult i8 %r, %s2 + %s = select i1 %c, i8 %r, i8 %s2 + ret i8 %s +} + +define arm_aapcs_vfpcc i32 @vminv_u_v16i8_i32(<16 x i8> %s1, i32 %s2) { +; CHECK-LABEL: vminv_u_v16i8_i32: +; CHECK: @ %bb.0: +; CHECK-NEXT: movs r1, #255 +; CHECK-NEXT: vminv.u8 r1, q0 +; CHECK-NEXT: uxtb r1, r1 +; CHECK-NEXT: cmp r1, r0 +; CHECK-NEXT: it lo +; CHECK-NEXT: movlo r0, r1 +; CHECK-NEXT: bx lr + %r = call i8 @llvm.experimental.vector.reduce.umin.v16i8(<16 x i8> %s1) + %rs = zext i8 %r to i32 + %c = icmp ult i32 %rs, %s2 + %s = select i1 %c, i32 %rs, i32 %s2 + ret i32 %s +} + +define arm_aapcs_vfpcc i16 @vminv_u_v8i16_i16(<8 x i16> %s1, i16 %s2) { +; CHECK-LABEL: vminv_u_v8i16_i16: +; CHECK: @ %bb.0: +; CHECK-NEXT: movw r1, #65535 +; CHECK-NEXT: uxth r3, r0 +; CHECK-NEXT: vminv.u16 r1, q0 +; CHECK-NEXT: uxth r2, r1 +; CHECK-NEXT: cmp r2, r3 +; CHECK-NEXT: it lo +; CHECK-NEXT: movlo r0, r1 +; CHECK-NEXT: bx lr + %r = call i16 @llvm.experimental.vector.reduce.umin.v8i16(<8 x i16> %s1) + %c = icmp ult i16 %r, %s2 + %s = select i1 %c, i16 %r, i16 %s2 + ret i16 %s +} + +define arm_aapcs_vfpcc i32 @vminv_u_v8i16_i32(<8 x i16> %s1, i32 %s2) { +; CHECK-LABEL: vminv_u_v8i16_i32: +; CHECK: @ %bb.0: +; CHECK-NEXT: movw r1, #65535 +; CHECK-NEXT: vminv.u16 r1, q0 +; CHECK-NEXT: uxth r1, r1 +; CHECK-NEXT: cmp r1, r0 +; CHECK-NEXT: it lo +; CHECK-NEXT: movlo r0, r1 +; CHECK-NEXT: bx lr + %r = call i16 @llvm.experimental.vector.reduce.umin.v8i16(<8 x i16> %s1) + %rs = zext i16 %r to i32 + %c = icmp ult i32 %rs, %s2 + %s = select i1 %c, i32 %rs, i32 %s2 + ret i32 %s +} + +define arm_aapcs_vfpcc i32 @vminv_u_v4i32_i32(<4 x i32> %s1, i32 %s2) { +; CHECK-LABEL: vminv_u_v4i32_i32: +; CHECK: @ %bb.0: +; CHECK-NEXT: mov.w r1, #-1 +; CHECK-NEXT: vminv.u32 r1, q0 +; CHECK-NEXT: cmp r1, r0 +; CHECK-NEXT: it lo +; CHECK-NEXT: movlo r0, r1 +; CHECK-NEXT: bx lr + %r = call i32 @llvm.experimental.vector.reduce.umin.v4i32(<4 x i32> %s1) + %c = icmp ult i32 %r, %s2 + %s = select i1 %c, i32 %r, i32 %s2 + ret i32 %s +}