diff --git a/llvm/test/CodeGen/RISCV/condops.ll b/llvm/test/CodeGen/RISCV/condops.ll index f5907321078f7..e43fa4f9acc69 100644 --- a/llvm/test/CodeGen/RISCV/condops.ll +++ b/llvm/test/CodeGen/RISCV/condops.ll @@ -1,10 +1,25 @@ -; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 3 +; RUN: llc -mtriple=riscv32 < %s | FileCheck %s -check-prefix=RV32I +; RUN: llc -mtriple=riscv64 < %s | FileCheck %s -check-prefix=RV64I ; RUN: llc -mtriple=riscv64 -mattr=+xventanacondops < %s | FileCheck %s -check-prefix=RV64XVENTANACONDOPS ; RUN: llc -mtriple=riscv64 -mattr=+xtheadcondmov < %s | FileCheck %s -check-prefix=RV64XTHEADCONDMOV -; RUN: llc -mtriple=riscv32 -mattr=+experimental-zicond < %s | FileCheck %s -check-prefixes=RV32ZICOND -; RUN: llc -mtriple=riscv64 -mattr=+experimental-zicond < %s | FileCheck %s -check-prefixes=RV64ZICOND +; RUN: llc -mtriple=riscv32 -mattr=+experimental-zicond < %s | FileCheck %s -check-prefix=RV32ZICOND +; RUN: llc -mtriple=riscv64 -mattr=+experimental-zicond < %s | FileCheck %s -check-prefix=RV64ZICOND define i64 @zero1(i64 %rs1, i1 zeroext %rc) { +; RV32I-LABEL: zero1: +; RV32I: # %bb.0: +; RV32I-NEXT: neg a2, a2 +; RV32I-NEXT: and a0, a2, a0 +; RV32I-NEXT: and a1, a2, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero1: +; RV64I: # %bb.0: +; RV64I-NEXT: neg a1, a1 +; RV64I-NEXT: and a0, a1, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero1: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a0, a1 @@ -30,6 +45,19 @@ define i64 @zero1(i64 %rs1, i1 zeroext %rc) { } define i64 @zero2(i64 %rs1, i1 zeroext %rc) { +; RV32I-LABEL: zero2: +; RV32I: # %bb.0: +; RV32I-NEXT: addi a2, a2, -1 +; RV32I-NEXT: and a0, a2, a0 +; RV32I-NEXT: and a1, a2, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero2: +; RV64I: # %bb.0: +; RV64I-NEXT: addi a1, a1, -1 +; RV64I-NEXT: and a0, a1, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero2: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a0, a1 @@ -55,6 +83,24 @@ define i64 @zero2(i64 %rs1, i1 zeroext %rc) { } define i64 @add1(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: add1: +; RV32I: # %bb.0: +; RV32I-NEXT: neg a0, a0 +; RV32I-NEXT: and a4, a0, a4 +; RV32I-NEXT: add a2, a2, a4 +; RV32I-NEXT: and a0, a0, a3 +; RV32I-NEXT: add a0, a1, a0 +; RV32I-NEXT: sltu a1, a0, a1 +; RV32I-NEXT: add a1, a2, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: add1: +; RV64I: # %bb.0: +; RV64I-NEXT: neg a0, a0 +; RV64I-NEXT: and a0, a0, a2 +; RV64I-NEXT: add a0, a1, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: add1: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a2, a0 @@ -88,6 +134,24 @@ define i64 @add1(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @add2(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: add2: +; RV32I: # %bb.0: +; RV32I-NEXT: neg a0, a0 +; RV32I-NEXT: and a2, a0, a2 +; RV32I-NEXT: add a2, a4, a2 +; RV32I-NEXT: and a0, a0, a1 +; RV32I-NEXT: add a0, a3, a0 +; RV32I-NEXT: sltu a1, a0, a3 +; RV32I-NEXT: add a1, a2, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: add2: +; RV64I: # %bb.0: +; RV64I-NEXT: neg a0, a0 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: add a0, a2, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: add2: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a1, a0 @@ -121,6 +185,24 @@ define i64 @add2(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @add3(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: add3: +; RV32I: # %bb.0: +; RV32I-NEXT: addi a0, a0, -1 +; RV32I-NEXT: and a4, a0, a4 +; RV32I-NEXT: add a2, a2, a4 +; RV32I-NEXT: and a0, a0, a3 +; RV32I-NEXT: add a0, a1, a0 +; RV32I-NEXT: sltu a1, a0, a1 +; RV32I-NEXT: add a1, a2, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: add3: +; RV64I: # %bb.0: +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a2 +; RV64I-NEXT: add a0, a1, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: add3: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a2, a0 @@ -154,6 +236,24 @@ define i64 @add3(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @add4(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: add4: +; RV32I: # %bb.0: +; RV32I-NEXT: addi a0, a0, -1 +; RV32I-NEXT: and a2, a0, a2 +; RV32I-NEXT: add a2, a4, a2 +; RV32I-NEXT: and a0, a0, a1 +; RV32I-NEXT: add a0, a3, a0 +; RV32I-NEXT: sltu a1, a0, a3 +; RV32I-NEXT: add a1, a2, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: add4: +; RV64I: # %bb.0: +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: add a0, a2, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: add4: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a1, a0 @@ -187,6 +287,25 @@ define i64 @add4(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @sub1(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: sub1: +; RV32I: # %bb.0: +; RV32I-NEXT: neg a0, a0 +; RV32I-NEXT: and a3, a0, a3 +; RV32I-NEXT: sltu a5, a1, a3 +; RV32I-NEXT: and a0, a0, a4 +; RV32I-NEXT: sub a2, a2, a0 +; RV32I-NEXT: sub a2, a2, a5 +; RV32I-NEXT: sub a0, a1, a3 +; RV32I-NEXT: mv a1, a2 +; RV32I-NEXT: ret +; +; RV64I-LABEL: sub1: +; RV64I: # %bb.0: +; RV64I-NEXT: neg a0, a0 +; RV64I-NEXT: and a0, a0, a2 +; RV64I-NEXT: sub a0, a1, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: sub1: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a2, a0 @@ -221,6 +340,25 @@ define i64 @sub1(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @sub2(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: sub2: +; RV32I: # %bb.0: +; RV32I-NEXT: addi a0, a0, -1 +; RV32I-NEXT: and a3, a0, a3 +; RV32I-NEXT: sltu a5, a1, a3 +; RV32I-NEXT: and a0, a0, a4 +; RV32I-NEXT: sub a2, a2, a0 +; RV32I-NEXT: sub a2, a2, a5 +; RV32I-NEXT: sub a0, a1, a3 +; RV32I-NEXT: mv a1, a2 +; RV32I-NEXT: ret +; +; RV64I-LABEL: sub2: +; RV64I: # %bb.0: +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a2 +; RV64I-NEXT: sub a0, a1, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: sub2: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a2, a0 @@ -255,6 +393,22 @@ define i64 @sub2(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @or1(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: or1: +; RV32I: # %bb.0: +; RV32I-NEXT: neg a5, a0 +; RV32I-NEXT: and a0, a5, a3 +; RV32I-NEXT: or a0, a1, a0 +; RV32I-NEXT: and a1, a5, a4 +; RV32I-NEXT: or a1, a2, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: or1: +; RV64I: # %bb.0: +; RV64I-NEXT: neg a0, a0 +; RV64I-NEXT: and a0, a0, a2 +; RV64I-NEXT: or a0, a1, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: or1: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a2, a0 @@ -287,6 +441,22 @@ define i64 @or1(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @or2(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: or2: +; RV32I: # %bb.0: +; RV32I-NEXT: neg a5, a0 +; RV32I-NEXT: and a0, a5, a1 +; RV32I-NEXT: or a0, a3, a0 +; RV32I-NEXT: and a1, a5, a2 +; RV32I-NEXT: or a1, a4, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: or2: +; RV64I: # %bb.0: +; RV64I-NEXT: neg a0, a0 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: or a0, a2, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: or2: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a1, a0 @@ -319,6 +489,22 @@ define i64 @or2(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @or3(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: or3: +; RV32I: # %bb.0: +; RV32I-NEXT: addi a5, a0, -1 +; RV32I-NEXT: and a0, a5, a3 +; RV32I-NEXT: or a0, a1, a0 +; RV32I-NEXT: and a1, a5, a4 +; RV32I-NEXT: or a1, a2, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: or3: +; RV64I: # %bb.0: +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a2 +; RV64I-NEXT: or a0, a1, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: or3: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a2, a0 @@ -351,6 +537,22 @@ define i64 @or3(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @or4(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: or4: +; RV32I: # %bb.0: +; RV32I-NEXT: addi a5, a0, -1 +; RV32I-NEXT: and a0, a5, a1 +; RV32I-NEXT: or a0, a3, a0 +; RV32I-NEXT: and a1, a5, a2 +; RV32I-NEXT: or a1, a4, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: or4: +; RV64I: # %bb.0: +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: or a0, a2, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: or4: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a1, a0 @@ -383,6 +585,22 @@ define i64 @or4(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @xor1(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: xor1: +; RV32I: # %bb.0: +; RV32I-NEXT: neg a5, a0 +; RV32I-NEXT: and a0, a5, a3 +; RV32I-NEXT: xor a0, a1, a0 +; RV32I-NEXT: and a1, a5, a4 +; RV32I-NEXT: xor a1, a2, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: xor1: +; RV64I: # %bb.0: +; RV64I-NEXT: neg a0, a0 +; RV64I-NEXT: and a0, a0, a2 +; RV64I-NEXT: xor a0, a1, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: xor1: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a2, a0 @@ -415,6 +633,22 @@ define i64 @xor1(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @xor2(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: xor2: +; RV32I: # %bb.0: +; RV32I-NEXT: neg a5, a0 +; RV32I-NEXT: and a0, a5, a1 +; RV32I-NEXT: xor a0, a3, a0 +; RV32I-NEXT: and a1, a5, a2 +; RV32I-NEXT: xor a1, a4, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: xor2: +; RV64I: # %bb.0: +; RV64I-NEXT: neg a0, a0 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: xor a0, a2, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: xor2: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a1, a0 @@ -447,6 +681,22 @@ define i64 @xor2(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @xor3(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: xor3: +; RV32I: # %bb.0: +; RV32I-NEXT: addi a5, a0, -1 +; RV32I-NEXT: and a0, a5, a3 +; RV32I-NEXT: xor a0, a1, a0 +; RV32I-NEXT: and a1, a5, a4 +; RV32I-NEXT: xor a1, a2, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: xor3: +; RV64I: # %bb.0: +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a2 +; RV64I-NEXT: xor a0, a1, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: xor3: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a2, a0 @@ -479,6 +729,22 @@ define i64 @xor3(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @xor4(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: xor4: +; RV32I: # %bb.0: +; RV32I-NEXT: addi a5, a0, -1 +; RV32I-NEXT: and a0, a5, a1 +; RV32I-NEXT: xor a0, a3, a0 +; RV32I-NEXT: and a1, a5, a2 +; RV32I-NEXT: xor a1, a4, a1 +; RV32I-NEXT: ret +; +; RV64I-LABEL: xor4: +; RV64I: # %bb.0: +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: xor a0, a2, a0 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: xor4: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a1, a0 @@ -511,6 +777,26 @@ define i64 @xor4(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @and1(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: and1: +; RV32I: # %bb.0: +; RV32I-NEXT: beqz a0, .LBB16_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: and a2, a2, a4 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: .LBB16_2: +; RV32I-NEXT: mv a0, a1 +; RV32I-NEXT: mv a1, a2 +; RV32I-NEXT: ret +; +; RV64I-LABEL: and1: +; RV64I: # %bb.0: +; RV64I-NEXT: beqz a0, .LBB16_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: and a1, a1, a2 +; RV64I-NEXT: .LBB16_2: +; RV64I-NEXT: mv a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: and1: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a1, a0 @@ -548,6 +834,26 @@ define i64 @and1(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @and2(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: and2: +; RV32I: # %bb.0: +; RV32I-NEXT: beqz a0, .LBB17_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: and a4, a2, a4 +; RV32I-NEXT: and a3, a1, a3 +; RV32I-NEXT: .LBB17_2: +; RV32I-NEXT: mv a0, a3 +; RV32I-NEXT: mv a1, a4 +; RV32I-NEXT: ret +; +; RV64I-LABEL: and2: +; RV64I: # %bb.0: +; RV64I-NEXT: beqz a0, .LBB17_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: and a2, a1, a2 +; RV64I-NEXT: .LBB17_2: +; RV64I-NEXT: mv a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: and2: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a2, a0 @@ -585,6 +891,26 @@ define i64 @and2(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @and3(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: and3: +; RV32I: # %bb.0: +; RV32I-NEXT: bnez a0, .LBB18_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: and a2, a2, a4 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: .LBB18_2: +; RV32I-NEXT: mv a0, a1 +; RV32I-NEXT: mv a1, a2 +; RV32I-NEXT: ret +; +; RV64I-LABEL: and3: +; RV64I: # %bb.0: +; RV64I-NEXT: bnez a0, .LBB18_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: and a1, a1, a2 +; RV64I-NEXT: .LBB18_2: +; RV64I-NEXT: mv a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: and3: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a1, a0 @@ -622,6 +948,26 @@ define i64 @and3(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @and4(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: and4: +; RV32I: # %bb.0: +; RV32I-NEXT: bnez a0, .LBB19_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: and a4, a2, a4 +; RV32I-NEXT: and a3, a1, a3 +; RV32I-NEXT: .LBB19_2: +; RV32I-NEXT: mv a0, a3 +; RV32I-NEXT: mv a1, a4 +; RV32I-NEXT: ret +; +; RV64I-LABEL: and4: +; RV64I: # %bb.0: +; RV64I-NEXT: bnez a0, .LBB19_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: and a2, a1, a2 +; RV64I-NEXT: .LBB19_2: +; RV64I-NEXT: mv a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: and4: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a2, a0 @@ -659,6 +1005,26 @@ define i64 @and4(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @basic(i1 zeroext %rc, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: basic: +; RV32I: # %bb.0: +; RV32I-NEXT: bnez a0, .LBB20_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: mv a1, a3 +; RV32I-NEXT: mv a2, a4 +; RV32I-NEXT: .LBB20_2: +; RV32I-NEXT: mv a0, a1 +; RV32I-NEXT: mv a1, a2 +; RV32I-NEXT: ret +; +; RV64I-LABEL: basic: +; RV64I: # %bb.0: +; RV64I-NEXT: bnez a0, .LBB20_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a1, a2 +; RV64I-NEXT: .LBB20_2: +; RV64I-NEXT: mv a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: basic: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a2, a2, a0 @@ -694,6 +1060,29 @@ define i64 @basic(i1 zeroext %rc, i64 %rs1, i64 %rs2) { } define i64 @seteq(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: seteq: +; RV32I: # %bb.0: +; RV32I-NEXT: xor a1, a1, a3 +; RV32I-NEXT: xor a0, a0, a2 +; RV32I-NEXT: or a2, a0, a1 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: beqz a2, .LBB21_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: mv a0, a6 +; RV32I-NEXT: mv a1, a7 +; RV32I-NEXT: .LBB21_2: +; RV32I-NEXT: ret +; +; RV64I-LABEL: seteq: +; RV64I: # %bb.0: +; RV64I-NEXT: beq a0, a1, .LBB21_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a2, a3 +; RV64I-NEXT: .LBB21_2: +; RV64I-NEXT: mv a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: seteq: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: xor a0, a0, a1 @@ -735,6 +1124,29 @@ define i64 @seteq(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { } define i64 @setne(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: setne: +; RV32I: # %bb.0: +; RV32I-NEXT: xor a1, a1, a3 +; RV32I-NEXT: xor a0, a0, a2 +; RV32I-NEXT: or a2, a0, a1 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: bnez a2, .LBB22_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: mv a0, a6 +; RV32I-NEXT: mv a1, a7 +; RV32I-NEXT: .LBB22_2: +; RV32I-NEXT: ret +; +; RV64I-LABEL: setne: +; RV64I: # %bb.0: +; RV64I-NEXT: bne a0, a1, .LBB22_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a2, a3 +; RV64I-NEXT: .LBB22_2: +; RV64I-NEXT: mv a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: setne: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: xor a0, a0, a1 @@ -776,6 +1188,33 @@ define i64 @setne(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { } define i64 @setgt(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: setgt: +; RV32I: # %bb.0: +; RV32I-NEXT: beq a1, a3, .LBB23_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: slt a0, a3, a1 +; RV32I-NEXT: beqz a0, .LBB23_3 +; RV32I-NEXT: j .LBB23_4 +; RV32I-NEXT: .LBB23_2: +; RV32I-NEXT: sltu a0, a2, a0 +; RV32I-NEXT: bnez a0, .LBB23_4 +; RV32I-NEXT: .LBB23_3: +; RV32I-NEXT: mv a4, a6 +; RV32I-NEXT: mv a5, a7 +; RV32I-NEXT: .LBB23_4: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: ret +; +; RV64I-LABEL: setgt: +; RV64I: # %bb.0: +; RV64I-NEXT: blt a1, a0, .LBB23_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a2, a3 +; RV64I-NEXT: .LBB23_2: +; RV64I-NEXT: mv a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: setgt: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: slt a0, a1, a0 @@ -820,6 +1259,35 @@ define i64 @setgt(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { } define i64 @setge(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: setge: +; RV32I: # %bb.0: +; RV32I-NEXT: beq a1, a3, .LBB24_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: slt a0, a1, a3 +; RV32I-NEXT: xori a0, a0, 1 +; RV32I-NEXT: beqz a0, .LBB24_3 +; RV32I-NEXT: j .LBB24_4 +; RV32I-NEXT: .LBB24_2: +; RV32I-NEXT: sltu a0, a0, a2 +; RV32I-NEXT: xori a0, a0, 1 +; RV32I-NEXT: bnez a0, .LBB24_4 +; RV32I-NEXT: .LBB24_3: +; RV32I-NEXT: mv a4, a6 +; RV32I-NEXT: mv a5, a7 +; RV32I-NEXT: .LBB24_4: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: ret +; +; RV64I-LABEL: setge: +; RV64I: # %bb.0: +; RV64I-NEXT: bge a0, a1, .LBB24_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a2, a3 +; RV64I-NEXT: .LBB24_2: +; RV64I-NEXT: mv a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: setge: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: slt a0, a0, a1 @@ -867,6 +1335,33 @@ define i64 @setge(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { } define i64 @setlt(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: setlt: +; RV32I: # %bb.0: +; RV32I-NEXT: beq a1, a3, .LBB25_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: slt a0, a1, a3 +; RV32I-NEXT: beqz a0, .LBB25_3 +; RV32I-NEXT: j .LBB25_4 +; RV32I-NEXT: .LBB25_2: +; RV32I-NEXT: sltu a0, a0, a2 +; RV32I-NEXT: bnez a0, .LBB25_4 +; RV32I-NEXT: .LBB25_3: +; RV32I-NEXT: mv a4, a6 +; RV32I-NEXT: mv a5, a7 +; RV32I-NEXT: .LBB25_4: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: ret +; +; RV64I-LABEL: setlt: +; RV64I: # %bb.0: +; RV64I-NEXT: blt a0, a1, .LBB25_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a2, a3 +; RV64I-NEXT: .LBB25_2: +; RV64I-NEXT: mv a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: setlt: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: slt a0, a0, a1 @@ -911,6 +1406,35 @@ define i64 @setlt(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { } define i64 @setle(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: setle: +; RV32I: # %bb.0: +; RV32I-NEXT: beq a1, a3, .LBB26_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: slt a0, a3, a1 +; RV32I-NEXT: xori a0, a0, 1 +; RV32I-NEXT: beqz a0, .LBB26_3 +; RV32I-NEXT: j .LBB26_4 +; RV32I-NEXT: .LBB26_2: +; RV32I-NEXT: sltu a0, a2, a0 +; RV32I-NEXT: xori a0, a0, 1 +; RV32I-NEXT: bnez a0, .LBB26_4 +; RV32I-NEXT: .LBB26_3: +; RV32I-NEXT: mv a4, a6 +; RV32I-NEXT: mv a5, a7 +; RV32I-NEXT: .LBB26_4: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: ret +; +; RV64I-LABEL: setle: +; RV64I: # %bb.0: +; RV64I-NEXT: bge a1, a0, .LBB26_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a2, a3 +; RV64I-NEXT: .LBB26_2: +; RV64I-NEXT: mv a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: setle: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: slt a0, a1, a0 @@ -958,6 +1482,33 @@ define i64 @setle(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { } define i64 @setugt(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: setugt: +; RV32I: # %bb.0: +; RV32I-NEXT: beq a1, a3, .LBB27_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: sltu a0, a3, a1 +; RV32I-NEXT: beqz a0, .LBB27_3 +; RV32I-NEXT: j .LBB27_4 +; RV32I-NEXT: .LBB27_2: +; RV32I-NEXT: sltu a0, a2, a0 +; RV32I-NEXT: bnez a0, .LBB27_4 +; RV32I-NEXT: .LBB27_3: +; RV32I-NEXT: mv a4, a6 +; RV32I-NEXT: mv a5, a7 +; RV32I-NEXT: .LBB27_4: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: ret +; +; RV64I-LABEL: setugt: +; RV64I: # %bb.0: +; RV64I-NEXT: bltu a1, a0, .LBB27_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a2, a3 +; RV64I-NEXT: .LBB27_2: +; RV64I-NEXT: mv a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: setugt: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: sltu a0, a1, a0 @@ -1002,6 +1553,35 @@ define i64 @setugt(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { } define i64 @setuge(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: setuge: +; RV32I: # %bb.0: +; RV32I-NEXT: beq a1, a3, .LBB28_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: sltu a0, a1, a3 +; RV32I-NEXT: xori a0, a0, 1 +; RV32I-NEXT: beqz a0, .LBB28_3 +; RV32I-NEXT: j .LBB28_4 +; RV32I-NEXT: .LBB28_2: +; RV32I-NEXT: sltu a0, a0, a2 +; RV32I-NEXT: xori a0, a0, 1 +; RV32I-NEXT: bnez a0, .LBB28_4 +; RV32I-NEXT: .LBB28_3: +; RV32I-NEXT: mv a4, a6 +; RV32I-NEXT: mv a5, a7 +; RV32I-NEXT: .LBB28_4: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: ret +; +; RV64I-LABEL: setuge: +; RV64I: # %bb.0: +; RV64I-NEXT: bgeu a0, a1, .LBB28_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a2, a3 +; RV64I-NEXT: .LBB28_2: +; RV64I-NEXT: mv a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: setuge: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: sltu a0, a0, a1 @@ -1049,6 +1629,33 @@ define i64 @setuge(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { } define i64 @setult(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: setult: +; RV32I: # %bb.0: +; RV32I-NEXT: beq a1, a3, .LBB29_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: sltu a0, a1, a3 +; RV32I-NEXT: beqz a0, .LBB29_3 +; RV32I-NEXT: j .LBB29_4 +; RV32I-NEXT: .LBB29_2: +; RV32I-NEXT: sltu a0, a0, a2 +; RV32I-NEXT: bnez a0, .LBB29_4 +; RV32I-NEXT: .LBB29_3: +; RV32I-NEXT: mv a4, a6 +; RV32I-NEXT: mv a5, a7 +; RV32I-NEXT: .LBB29_4: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: ret +; +; RV64I-LABEL: setult: +; RV64I: # %bb.0: +; RV64I-NEXT: bltu a0, a1, .LBB29_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a2, a3 +; RV64I-NEXT: .LBB29_2: +; RV64I-NEXT: mv a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: setult: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: sltu a0, a0, a1 @@ -1093,6 +1700,35 @@ define i64 @setult(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { } define i64 @setule(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: setule: +; RV32I: # %bb.0: +; RV32I-NEXT: beq a1, a3, .LBB30_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: sltu a0, a3, a1 +; RV32I-NEXT: xori a0, a0, 1 +; RV32I-NEXT: beqz a0, .LBB30_3 +; RV32I-NEXT: j .LBB30_4 +; RV32I-NEXT: .LBB30_2: +; RV32I-NEXT: sltu a0, a2, a0 +; RV32I-NEXT: xori a0, a0, 1 +; RV32I-NEXT: bnez a0, .LBB30_4 +; RV32I-NEXT: .LBB30_3: +; RV32I-NEXT: mv a4, a6 +; RV32I-NEXT: mv a5, a7 +; RV32I-NEXT: .LBB30_4: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: ret +; +; RV64I-LABEL: setule: +; RV64I: # %bb.0: +; RV64I-NEXT: bgeu a1, a0, .LBB30_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a2, a3 +; RV64I-NEXT: .LBB30_2: +; RV64I-NEXT: mv a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: setule: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: sltu a0, a1, a0 @@ -1140,6 +1776,27 @@ define i64 @setule(i64 %a, i64 %b, i64 %rs1, i64 %rs2) { } define i64 @seteq_zero(i64 %a, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: seteq_zero: +; RV32I: # %bb.0: +; RV32I-NEXT: or a6, a0, a1 +; RV32I-NEXT: mv a1, a3 +; RV32I-NEXT: mv a0, a2 +; RV32I-NEXT: beqz a6, .LBB31_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: .LBB31_2: +; RV32I-NEXT: ret +; +; RV64I-LABEL: seteq_zero: +; RV64I: # %bb.0: +; RV64I-NEXT: beqz a0, .LBB31_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a1, a2 +; RV64I-NEXT: .LBB31_2: +; RV64I-NEXT: mv a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: seteq_zero: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a1, a1, a0 @@ -1176,6 +1833,27 @@ define i64 @seteq_zero(i64 %a, i64 %rs1, i64 %rs2) { } define i64 @setne_zero(i64 %a, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: setne_zero: +; RV32I: # %bb.0: +; RV32I-NEXT: or a6, a0, a1 +; RV32I-NEXT: mv a1, a3 +; RV32I-NEXT: mv a0, a2 +; RV32I-NEXT: bnez a6, .LBB32_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: .LBB32_2: +; RV32I-NEXT: ret +; +; RV64I-LABEL: setne_zero: +; RV64I: # %bb.0: +; RV64I-NEXT: bnez a0, .LBB32_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a1, a2 +; RV64I-NEXT: .LBB32_2: +; RV64I-NEXT: mv a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: setne_zero: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a2, a2, a0 @@ -1212,6 +1890,29 @@ define i64 @setne_zero(i64 %a, i64 %rs1, i64 %rs2) { } define i64 @seteq_constant(i64 %a, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: seteq_constant: +; RV32I: # %bb.0: +; RV32I-NEXT: xori a0, a0, 123 +; RV32I-NEXT: or a6, a0, a1 +; RV32I-NEXT: mv a1, a3 +; RV32I-NEXT: mv a0, a2 +; RV32I-NEXT: beqz a6, .LBB33_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: .LBB33_2: +; RV32I-NEXT: ret +; +; RV64I-LABEL: seteq_constant: +; RV64I: # %bb.0: +; RV64I-NEXT: li a3, 123 +; RV64I-NEXT: beq a0, a3, .LBB33_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a1, a2 +; RV64I-NEXT: .LBB33_2: +; RV64I-NEXT: mv a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: seteq_constant: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: addi a0, a0, -123 @@ -1252,6 +1953,29 @@ define i64 @seteq_constant(i64 %a, i64 %rs1, i64 %rs2) { } define i64 @setne_constant(i64 %a, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: setne_constant: +; RV32I: # %bb.0: +; RV32I-NEXT: xori a0, a0, 456 +; RV32I-NEXT: or a6, a0, a1 +; RV32I-NEXT: mv a1, a3 +; RV32I-NEXT: mv a0, a2 +; RV32I-NEXT: bnez a6, .LBB34_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: .LBB34_2: +; RV32I-NEXT: ret +; +; RV64I-LABEL: setne_constant: +; RV64I: # %bb.0: +; RV64I-NEXT: li a3, 456 +; RV64I-NEXT: bne a0, a3, .LBB34_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a1, a2 +; RV64I-NEXT: .LBB34_2: +; RV64I-NEXT: mv a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: setne_constant: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: addi a0, a0, -456 @@ -1292,6 +2016,32 @@ define i64 @setne_constant(i64 %a, i64 %rs1, i64 %rs2) { } define i64 @seteq_2048(i64 %a, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: seteq_2048: +; RV32I: # %bb.0: +; RV32I-NEXT: li a6, 1 +; RV32I-NEXT: slli a6, a6, 11 +; RV32I-NEXT: xor a0, a0, a6 +; RV32I-NEXT: or a6, a0, a1 +; RV32I-NEXT: mv a1, a3 +; RV32I-NEXT: mv a0, a2 +; RV32I-NEXT: beqz a6, .LBB35_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: .LBB35_2: +; RV32I-NEXT: ret +; +; RV64I-LABEL: seteq_2048: +; RV64I: # %bb.0: +; RV64I-NEXT: li a3, 1 +; RV64I-NEXT: slli a3, a3, 11 +; RV64I-NEXT: beq a0, a3, .LBB35_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a1, a2 +; RV64I-NEXT: .LBB35_2: +; RV64I-NEXT: mv a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: seteq_2048: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: addi a0, a0, -2048 @@ -1334,6 +2084,30 @@ define i64 @seteq_2048(i64 %a, i64 %rs1, i64 %rs2) { } define i64 @seteq_neg2048(i64 %a, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: seteq_neg2048: +; RV32I: # %bb.0: +; RV32I-NEXT: not a1, a1 +; RV32I-NEXT: xori a0, a0, -2048 +; RV32I-NEXT: or a6, a0, a1 +; RV32I-NEXT: mv a1, a3 +; RV32I-NEXT: mv a0, a2 +; RV32I-NEXT: beqz a6, .LBB36_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: .LBB36_2: +; RV32I-NEXT: ret +; +; RV64I-LABEL: seteq_neg2048: +; RV64I: # %bb.0: +; RV64I-NEXT: li a3, -2048 +; RV64I-NEXT: beq a0, a3, .LBB36_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a1, a2 +; RV64I-NEXT: .LBB36_2: +; RV64I-NEXT: mv a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: seteq_neg2048: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: xori a0, a0, -2048 @@ -1375,6 +2149,30 @@ define i64 @seteq_neg2048(i64 %a, i64 %rs1, i64 %rs2) { } define i64 @setne_neg2048(i64 %a, i64 %rs1, i64 %rs2) { +; RV32I-LABEL: setne_neg2048: +; RV32I: # %bb.0: +; RV32I-NEXT: not a1, a1 +; RV32I-NEXT: xori a0, a0, -2048 +; RV32I-NEXT: or a6, a0, a1 +; RV32I-NEXT: mv a1, a3 +; RV32I-NEXT: mv a0, a2 +; RV32I-NEXT: bnez a6, .LBB37_2 +; RV32I-NEXT: # %bb.1: +; RV32I-NEXT: mv a0, a4 +; RV32I-NEXT: mv a1, a5 +; RV32I-NEXT: .LBB37_2: +; RV32I-NEXT: ret +; +; RV64I-LABEL: setne_neg2048: +; RV64I: # %bb.0: +; RV64I-NEXT: li a3, -2048 +; RV64I-NEXT: bne a0, a3, .LBB37_2 +; RV64I-NEXT: # %bb.1: +; RV64I-NEXT: mv a1, a2 +; RV64I-NEXT: .LBB37_2: +; RV64I-NEXT: mv a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: setne_neg2048: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: xori a0, a0, -2048 @@ -1416,6 +2214,25 @@ define i64 @setne_neg2048(i64 %a, i64 %rs1, i64 %rs2) { } define i64 @zero1_seteq(i64 %a, i64 %b, i64 %rs1) { +; RV32I-LABEL: zero1_seteq: +; RV32I: # %bb.0: +; RV32I-NEXT: xor a1, a1, a3 +; RV32I-NEXT: xor a0, a0, a2 +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: snez a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a4 +; RV32I-NEXT: and a1, a1, a5 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero1_seteq: +; RV64I: # %bb.0: +; RV64I-NEXT: xor a0, a0, a1 +; RV64I-NEXT: snez a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero1_seteq: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: xor a0, a0, a1 @@ -1449,6 +2266,25 @@ define i64 @zero1_seteq(i64 %a, i64 %b, i64 %rs1) { } define i64 @zero2_seteq(i64 %a, i64 %b, i64 %rs1) { +; RV32I-LABEL: zero2_seteq: +; RV32I: # %bb.0: +; RV32I-NEXT: xor a1, a1, a3 +; RV32I-NEXT: xor a0, a0, a2 +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: seqz a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a4 +; RV32I-NEXT: and a1, a1, a5 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero2_seteq: +; RV64I: # %bb.0: +; RV64I-NEXT: xor a0, a0, a1 +; RV64I-NEXT: seqz a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero2_seteq: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: xor a0, a0, a1 @@ -1482,6 +2318,25 @@ define i64 @zero2_seteq(i64 %a, i64 %b, i64 %rs1) { } define i64 @zero1_setne(i64 %a, i64 %b, i64 %rs1) { +; RV32I-LABEL: zero1_setne: +; RV32I: # %bb.0: +; RV32I-NEXT: xor a1, a1, a3 +; RV32I-NEXT: xor a0, a0, a2 +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: seqz a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a4 +; RV32I-NEXT: and a1, a1, a5 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero1_setne: +; RV64I: # %bb.0: +; RV64I-NEXT: xor a0, a0, a1 +; RV64I-NEXT: seqz a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero1_setne: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: xor a0, a0, a1 @@ -1515,6 +2370,25 @@ define i64 @zero1_setne(i64 %a, i64 %b, i64 %rs1) { } define i64 @zero2_setne(i64 %a, i64 %b, i64 %rs1) { +; RV32I-LABEL: zero2_setne: +; RV32I: # %bb.0: +; RV32I-NEXT: xor a1, a1, a3 +; RV32I-NEXT: xor a0, a0, a2 +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: snez a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a4 +; RV32I-NEXT: and a1, a1, a5 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero2_setne: +; RV64I: # %bb.0: +; RV64I-NEXT: xor a0, a0, a1 +; RV64I-NEXT: snez a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a2 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero2_setne: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: xor a0, a0, a1 @@ -1548,6 +2422,22 @@ define i64 @zero2_setne(i64 %a, i64 %b, i64 %rs1) { } define i64 @zero1_seteq_zero(i64 %a, i64 %rs1) { +; RV32I-LABEL: zero1_seteq_zero: +; RV32I: # %bb.0: +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: snez a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a2 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero1_seteq_zero: +; RV64I: # %bb.0: +; RV64I-NEXT: snez a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero1_seteq_zero: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a1, a0 @@ -1576,6 +2466,22 @@ define i64 @zero1_seteq_zero(i64 %a, i64 %rs1) { } define i64 @zero2_seteq_zero(i64 %a, i64 %rs1) { +; RV32I-LABEL: zero2_seteq_zero: +; RV32I: # %bb.0: +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: seqz a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a2 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero2_seteq_zero: +; RV64I: # %bb.0: +; RV64I-NEXT: seqz a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero2_seteq_zero: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a1, a0 @@ -1604,6 +2510,22 @@ define i64 @zero2_seteq_zero(i64 %a, i64 %rs1) { } define i64 @zero1_setne_zero(i64 %a, i64 %rs1) { +; RV32I-LABEL: zero1_setne_zero: +; RV32I: # %bb.0: +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: seqz a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a2 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero1_setne_zero: +; RV64I: # %bb.0: +; RV64I-NEXT: seqz a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero1_setne_zero: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a1, a0 @@ -1632,6 +2554,22 @@ define i64 @zero1_setne_zero(i64 %a, i64 %rs1) { } define i64 @zero2_setne_zero(i64 %a, i64 %rs1) { +; RV32I-LABEL: zero2_setne_zero: +; RV32I: # %bb.0: +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: snez a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a2 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero2_setne_zero: +; RV64I: # %bb.0: +; RV64I-NEXT: snez a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero2_setne_zero: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a1, a0 @@ -1660,6 +2598,25 @@ define i64 @zero2_setne_zero(i64 %a, i64 %rs1) { } define i64 @zero1_seteq_constant(i64 %a, i64 %rs1) { +; RV32I-LABEL: zero1_seteq_constant: +; RV32I: # %bb.0: +; RV32I-NEXT: not a1, a1 +; RV32I-NEXT: xori a0, a0, -231 +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: snez a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a2 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero1_seteq_constant: +; RV64I: # %bb.0: +; RV64I-NEXT: addi a0, a0, 231 +; RV64I-NEXT: snez a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero1_seteq_constant: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: addi a0, a0, 231 @@ -1693,6 +2650,24 @@ define i64 @zero1_seteq_constant(i64 %a, i64 %rs1) { } define i64 @zero2_seteq_constant(i64 %a, i64 %rs1) { +; RV32I-LABEL: zero2_seteq_constant: +; RV32I: # %bb.0: +; RV32I-NEXT: xori a0, a0, 546 +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: seqz a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a2 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero2_seteq_constant: +; RV64I: # %bb.0: +; RV64I-NEXT: addi a0, a0, -546 +; RV64I-NEXT: seqz a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero2_seteq_constant: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: addi a0, a0, -546 @@ -1725,6 +2700,24 @@ define i64 @zero2_seteq_constant(i64 %a, i64 %rs1) { } define i64 @zero1_setne_constant(i64 %a, i64 %rs1) { +; RV32I-LABEL: zero1_setne_constant: +; RV32I: # %bb.0: +; RV32I-NEXT: xori a0, a0, 321 +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: seqz a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a2 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero1_setne_constant: +; RV64I: # %bb.0: +; RV64I-NEXT: addi a0, a0, -321 +; RV64I-NEXT: seqz a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero1_setne_constant: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: addi a0, a0, -321 @@ -1757,6 +2750,25 @@ define i64 @zero1_setne_constant(i64 %a, i64 %rs1) { } define i64 @zero2_setne_constant(i64 %a, i64 %rs1) { +; RV32I-LABEL: zero2_setne_constant: +; RV32I: # %bb.0: +; RV32I-NEXT: not a1, a1 +; RV32I-NEXT: xori a0, a0, -654 +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: snez a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a2 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero2_setne_constant: +; RV64I: # %bb.0: +; RV64I-NEXT: addi a0, a0, 654 +; RV64I-NEXT: snez a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero2_setne_constant: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: addi a0, a0, 654 @@ -1790,6 +2802,25 @@ define i64 @zero2_setne_constant(i64 %a, i64 %rs1) { } define i64 @zero1_seteq_neg2048(i64 %a, i64 %rs1) { +; RV32I-LABEL: zero1_seteq_neg2048: +; RV32I: # %bb.0: +; RV32I-NEXT: not a1, a1 +; RV32I-NEXT: xori a0, a0, -2048 +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: snez a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a2 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero1_seteq_neg2048: +; RV64I: # %bb.0: +; RV64I-NEXT: xori a0, a0, -2048 +; RV64I-NEXT: snez a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero1_seteq_neg2048: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: xori a0, a0, -2048 @@ -1823,6 +2854,25 @@ define i64 @zero1_seteq_neg2048(i64 %a, i64 %rs1) { } define i64 @zero2_seteq_neg2048(i64 %a, i64 %rs1) { +; RV32I-LABEL: zero2_seteq_neg2048: +; RV32I: # %bb.0: +; RV32I-NEXT: not a1, a1 +; RV32I-NEXT: xori a0, a0, -2048 +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: seqz a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a2 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero2_seteq_neg2048: +; RV64I: # %bb.0: +; RV64I-NEXT: xori a0, a0, -2048 +; RV64I-NEXT: seqz a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero2_seteq_neg2048: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: xori a0, a0, -2048 @@ -1856,6 +2906,25 @@ define i64 @zero2_seteq_neg2048(i64 %a, i64 %rs1) { } define i64 @zero1_setne_neg2048(i64 %a, i64 %rs1) { +; RV32I-LABEL: zero1_setne_neg2048: +; RV32I: # %bb.0: +; RV32I-NEXT: not a1, a1 +; RV32I-NEXT: xori a0, a0, -2048 +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: seqz a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a2 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero1_setne_neg2048: +; RV64I: # %bb.0: +; RV64I-NEXT: xori a0, a0, -2048 +; RV64I-NEXT: seqz a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero1_setne_neg2048: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: xori a0, a0, -2048 @@ -1889,6 +2958,25 @@ define i64 @zero1_setne_neg2048(i64 %a, i64 %rs1) { } define i64 @zero2_setne_neg2048(i64 %a, i64 %rs1) { +; RV32I-LABEL: zero2_setne_neg2048: +; RV32I: # %bb.0: +; RV32I-NEXT: not a1, a1 +; RV32I-NEXT: xori a0, a0, -2048 +; RV32I-NEXT: or a0, a0, a1 +; RV32I-NEXT: snez a0, a0 +; RV32I-NEXT: addi a1, a0, -1 +; RV32I-NEXT: and a0, a1, a2 +; RV32I-NEXT: and a1, a1, a3 +; RV32I-NEXT: ret +; +; RV64I-LABEL: zero2_setne_neg2048: +; RV64I: # %bb.0: +; RV64I-NEXT: xori a0, a0, -2048 +; RV64I-NEXT: snez a0, a0 +; RV64I-NEXT: addi a0, a0, -1 +; RV64I-NEXT: and a0, a0, a1 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: zero2_setne_neg2048: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: xori a0, a0, -2048 @@ -1921,8 +3009,53 @@ define i64 @zero2_setne_neg2048(i64 %a, i64 %rs1) { ret i64 %sel } -; Test that we are able to convert the sext.w int he loop to mv. define void @sextw_removal_maskc(i1 %c, i32 signext %arg, i32 signext %arg1) nounwind { +; RV32I-LABEL: sextw_removal_maskc: +; RV32I: # %bb.0: # %bb +; RV32I-NEXT: addi sp, sp, -16 +; RV32I-NEXT: sw ra, 12(sp) # 4-byte Folded Spill +; RV32I-NEXT: sw s0, 8(sp) # 4-byte Folded Spill +; RV32I-NEXT: sw s1, 4(sp) # 4-byte Folded Spill +; RV32I-NEXT: mv s0, a2 +; RV32I-NEXT: slli a0, a0, 31 +; RV32I-NEXT: srai a0, a0, 31 +; RV32I-NEXT: and s1, a0, a1 +; RV32I-NEXT: .LBB54_1: # %bb2 +; RV32I-NEXT: # =>This Inner Loop Header: Depth=1 +; RV32I-NEXT: mv a0, s1 +; RV32I-NEXT: call bar@plt +; RV32I-NEXT: sll s1, s1, s0 +; RV32I-NEXT: bnez a0, .LBB54_1 +; RV32I-NEXT: # %bb.2: # %bb7 +; RV32I-NEXT: lw ra, 12(sp) # 4-byte Folded Reload +; RV32I-NEXT: lw s0, 8(sp) # 4-byte Folded Reload +; RV32I-NEXT: lw s1, 4(sp) # 4-byte Folded Reload +; RV32I-NEXT: addi sp, sp, 16 +; RV32I-NEXT: ret +; +; RV64I-LABEL: sextw_removal_maskc: +; RV64I: # %bb.0: # %bb +; RV64I-NEXT: addi sp, sp, -32 +; RV64I-NEXT: sd ra, 24(sp) # 8-byte Folded Spill +; RV64I-NEXT: sd s0, 16(sp) # 8-byte Folded Spill +; RV64I-NEXT: sd s1, 8(sp) # 8-byte Folded Spill +; RV64I-NEXT: mv s0, a2 +; RV64I-NEXT: slli a0, a0, 63 +; RV64I-NEXT: srai a0, a0, 63 +; RV64I-NEXT: and s1, a0, a1 +; RV64I-NEXT: .LBB54_1: # %bb2 +; RV64I-NEXT: # =>This Inner Loop Header: Depth=1 +; RV64I-NEXT: mv a0, s1 +; RV64I-NEXT: call bar@plt +; RV64I-NEXT: sllw s1, s1, s0 +; RV64I-NEXT: bnez a0, .LBB54_1 +; RV64I-NEXT: # %bb.2: # %bb7 +; RV64I-NEXT: ld ra, 24(sp) # 8-byte Folded Reload +; RV64I-NEXT: ld s0, 16(sp) # 8-byte Folded Reload +; RV64I-NEXT: ld s1, 8(sp) # 8-byte Folded Reload +; RV64I-NEXT: addi sp, sp, 32 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: sextw_removal_maskc: ; RV64XVENTANACONDOPS: # %bb.0: # %bb ; RV64XVENTANACONDOPS-NEXT: addi sp, sp, -32 @@ -2028,6 +3161,52 @@ bb7: ; preds = %bb2 declare signext i32 @bar(i32 signext) define void @sextw_removal_maskcn(i1 %c, i32 signext %arg, i32 signext %arg1) nounwind { +; RV32I-LABEL: sextw_removal_maskcn: +; RV32I: # %bb.0: # %bb +; RV32I-NEXT: addi sp, sp, -16 +; RV32I-NEXT: sw ra, 12(sp) # 4-byte Folded Spill +; RV32I-NEXT: sw s0, 8(sp) # 4-byte Folded Spill +; RV32I-NEXT: sw s1, 4(sp) # 4-byte Folded Spill +; RV32I-NEXT: mv s0, a2 +; RV32I-NEXT: andi a0, a0, 1 +; RV32I-NEXT: addi a0, a0, -1 +; RV32I-NEXT: and s1, a0, a1 +; RV32I-NEXT: .LBB55_1: # %bb2 +; RV32I-NEXT: # =>This Inner Loop Header: Depth=1 +; RV32I-NEXT: mv a0, s1 +; RV32I-NEXT: call bar@plt +; RV32I-NEXT: sll s1, s1, s0 +; RV32I-NEXT: bnez a0, .LBB55_1 +; RV32I-NEXT: # %bb.2: # %bb7 +; RV32I-NEXT: lw ra, 12(sp) # 4-byte Folded Reload +; RV32I-NEXT: lw s0, 8(sp) # 4-byte Folded Reload +; RV32I-NEXT: lw s1, 4(sp) # 4-byte Folded Reload +; RV32I-NEXT: addi sp, sp, 16 +; RV32I-NEXT: ret +; +; RV64I-LABEL: sextw_removal_maskcn: +; RV64I: # %bb.0: # %bb +; RV64I-NEXT: addi sp, sp, -32 +; RV64I-NEXT: sd ra, 24(sp) # 8-byte Folded Spill +; RV64I-NEXT: sd s0, 16(sp) # 8-byte Folded Spill +; RV64I-NEXT: sd s1, 8(sp) # 8-byte Folded Spill +; RV64I-NEXT: mv s0, a2 +; RV64I-NEXT: andi a0, a0, 1 +; RV64I-NEXT: addiw a0, a0, -1 +; RV64I-NEXT: and s1, a0, a1 +; RV64I-NEXT: .LBB55_1: # %bb2 +; RV64I-NEXT: # =>This Inner Loop Header: Depth=1 +; RV64I-NEXT: mv a0, s1 +; RV64I-NEXT: call bar@plt +; RV64I-NEXT: sllw s1, s1, s0 +; RV64I-NEXT: bnez a0, .LBB55_1 +; RV64I-NEXT: # %bb.2: # %bb7 +; RV64I-NEXT: ld ra, 24(sp) # 8-byte Folded Reload +; RV64I-NEXT: ld s0, 16(sp) # 8-byte Folded Reload +; RV64I-NEXT: ld s1, 8(sp) # 8-byte Folded Reload +; RV64I-NEXT: addi sp, sp, 32 +; RV64I-NEXT: ret +; ; RV64XVENTANACONDOPS-LABEL: sextw_removal_maskcn: ; RV64XVENTANACONDOPS: # %bb.0: # %bb ; RV64XVENTANACONDOPS-NEXT: addi sp, sp, -32