From 4a9086b537371627252832eb4ce56065c220a787 Mon Sep 17 00:00:00 2001 From: Oliver Stannard Date: Tue, 5 Mar 2019 10:42:34 +0000 Subject: [PATCH] [ARM] Fix select_cc lowering for fp16 When lowering a select_cc node where the true and false values are of type f16, we can't use a general conditional move because the FP16 instructions do not support conditional execution. Instead, we must ensure that the condition code is one of the four supported by the VSEL instruction. Differential revision: https://reviews.llvm.org/D58813 llvm-svn: 355385 --- llvm/lib/Target/ARM/ARMISelLowering.cpp | 18 +- llvm/test/CodeGen/ARM/vsel-fp16.ll | 676 ++++++++++++++++++++++++ 2 files changed, 687 insertions(+), 7 deletions(-) create mode 100644 llvm/test/CodeGen/ARM/vsel-fp16.ll diff --git a/llvm/lib/Target/ARM/ARMISelLowering.cpp b/llvm/lib/Target/ARM/ARMISelLowering.cpp index 601da9f598104..de4f894ff1107 100644 --- a/llvm/lib/Target/ARM/ARMISelLowering.cpp +++ b/llvm/lib/Target/ARM/ARMISelLowering.cpp @@ -4494,7 +4494,8 @@ SDValue ARMTargetLowering::LowerSELECT_CC(SDValue Op, SelectionDAG &DAG) const { // inverting the compare condition, swapping 'less' and 'greater') and // sometimes need to swap the operands to the VSEL (which inverts the // condition in the sense of firing whenever the previous condition didn't) - if (Subtarget->hasFPARMv8() && (TrueVal.getValueType() == MVT::f32 || + if (Subtarget->hasFPARMv8() && (TrueVal.getValueType() == MVT::f16 || + TrueVal.getValueType() == MVT::f32 || TrueVal.getValueType() == MVT::f64)) { ARMCC::CondCodes CondCode = IntCCToARMCC(CC); if (CondCode == ARMCC::LT || CondCode == ARMCC::LE || @@ -4514,12 +4515,15 @@ SDValue ARMTargetLowering::LowerSELECT_CC(SDValue Op, SelectionDAG &DAG) const { bool InvalidOnQNaN; FPCCToARMCC(CC, CondCode, CondCode2, InvalidOnQNaN); - // Normalize the fp compare. If RHS is zero we keep it there so we match - // CMPFPw0 instead of CMPFP. - if (Subtarget->hasFPARMv8() && !isFloatingPointZero(RHS) && - (TrueVal.getValueType() == MVT::f16 || - TrueVal.getValueType() == MVT::f32 || - TrueVal.getValueType() == MVT::f64)) { + // Normalize the fp compare. If RHS is zero we prefer to keep it there so we + // match CMPFPw0 instead of CMPFP, though we don't do this for f16 because we + // must use VSEL (limited condition codes), due to not having conditional f16 + // moves. + if (Subtarget->hasFPARMv8() && + !(isFloatingPointZero(RHS) && TrueVal.getValueType() != MVT::f16) && + (TrueVal.getValueType() == MVT::f16 || + TrueVal.getValueType() == MVT::f32 || + TrueVal.getValueType() == MVT::f64)) { bool swpCmpOps = false; bool swpVselOps = false; checkVSELConstraints(CC, CondCode, swpCmpOps, swpVselOps); diff --git a/llvm/test/CodeGen/ARM/vsel-fp16.ll b/llvm/test/CodeGen/ARM/vsel-fp16.ll new file mode 100644 index 0000000000000..9ccc6f4272830 --- /dev/null +++ b/llvm/test/CodeGen/ARM/vsel-fp16.ll @@ -0,0 +1,676 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mtriple=armv8a--none-eabi -mattr=+fullfp16 -float-abi=hard | FileCheck %s + +@varhalf = global half 0.0 +@vardouble = global double 0.0 +define void @test_vsel32sgt(i32 %lhs, i32 %rhs, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32sgt: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: cmp r0, r1 +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vselgt.f16 s0, s0, s2 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %tst1 = icmp sgt i32 %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32sge(i32 %lhs, i32 %rhs, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32sge: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: cmp r0, r1 +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vselge.f16 s0, s0, s2 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %tst1 = icmp sge i32 %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32eq(i32 %lhs, i32 %rhs, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32eq: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: cmp r0, r1 +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vseleq.f16 s0, s0, s2 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %tst1 = icmp eq i32 %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32slt(i32 %lhs, i32 %rhs, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32slt: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: cmp r0, r1 +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vselge.f16 s0, s2, s0 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %tst1 = icmp slt i32 %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32sle(i32 %lhs, i32 %rhs, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32sle: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: cmp r0, r1 +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vselgt.f16 s0, s2, s0 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %tst1 = icmp sle i32 %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32ogt(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32ogt: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselgt.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp ogt half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32oge(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32oge: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselge.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp oge half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32oeq(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32oeq: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmp.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vseleq.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp oeq half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32ugt(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32ugt: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s6, s4 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselge.f16 s0, s2, s0 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp ugt half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32uge(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32uge: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s6, s4 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselgt.f16 s0, s2, s0 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp uge half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32olt(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32olt: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s6, s4 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselgt.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp olt half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32ult(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32ult: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselge.f16 s0, s2, s0 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp ult half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32ole(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32ole: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s6, s4 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselge.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp ole half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32ule(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32ule: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselgt.f16 s0, s2, s0 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp ule half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32ord(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32ord: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselvs.f16 s0, s2, s0 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp ord half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32une(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32une: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmp.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vseleq.f16 s0, s2, s0 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp une half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32uno(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32uno: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselvs.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp uno half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + + +define void @test_vsel32ogt_nnan(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32ogt_nnan: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselgt.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp nnan ogt half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32oge_nnan(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32oge_nnan: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselge.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp nnan oge half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32oeq_nnan(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32oeq_nnan: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmp.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vseleq.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp nnan oeq half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32ugt_nnan(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32ugt_nnan: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselgt.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp nnan ugt half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32uge_nnan(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32uge_nnan: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselge.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp nnan uge half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32olt_nnan(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32olt_nnan: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s6, s4 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselgt.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp nnan olt half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32ult_nnan(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32ult_nnan: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s6, s4 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselgt.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp nnan ult half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32ole_nnan(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32ole_nnan: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s6, s4 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselge.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp nnan ole half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32ule_nnan(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32ule_nnan: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s6, s4 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselge.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp nnan ule half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32ord_nnan(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32ord_nnan: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselvs.f16 s0, s2, s0 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp nnan ord half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32une_nnan(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32une_nnan: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmp.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vseleq.f16 s0, s2, s0 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp nnan une half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +} + +define void @test_vsel32uno_nnan(half* %lhs_ptr, half* %rhs_ptr, half* %a_ptr, half* %b_ptr) { +; CHECK-LABEL: test_vsel32uno_nnan: +; CHECK: @ %bb.0: +; CHECK-NEXT: vldr.16 s0, [r2] +; CHECK-NEXT: vldr.16 s2, [r3] +; CHECK-NEXT: vldr.16 s4, [r0] +; CHECK-NEXT: vldr.16 s6, [r1] +; CHECK-NEXT: movw r0, :lower16:varhalf +; CHECK-NEXT: vcmpe.f16 s4, s6 +; CHECK-NEXT: movt r0, :upper16:varhalf +; CHECK-NEXT: vmrs APSR_nzcv, fpscr +; CHECK-NEXT: vselvs.f16 s0, s0, s2 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr + %a = load volatile half, half* %a_ptr + %b = load volatile half, half* %b_ptr + %lhs = load volatile half, half* %lhs_ptr + %rhs = load volatile half, half* %rhs_ptr + %tst1 = fcmp nnan uno half %lhs, %rhs + %val1 = select i1 %tst1, half %a, half %b + store half %val1, half* @varhalf + ret void +}