30 changes: 30 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfcvt-rtz-x-f.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x i16> @llvm.riscv.vfcvt.rtz.x.f.v.nxv1i16.nxv1f16(
<vscale x 1 x i16>,
<vscale x 1 x half>,
iXLen);

Expand All @@ -15,6 +16,7 @@ define <vscale x 1 x i16> @intrinsic_vfcvt_rtz.x.f.v_nxv1i16_nxv1f16(<vscale x 1
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i16> @llvm.riscv.vfcvt.rtz.x.f.v.nxv1i16.nxv1f16(
<vscale x 1 x i16> undef,
<vscale x 1 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -45,6 +47,7 @@ entry:
}

declare <vscale x 2 x i16> @llvm.riscv.vfcvt.rtz.x.f.v.nxv2i16.nxv2f16(
<vscale x 2 x i16>,
<vscale x 2 x half>,
iXLen);

Expand All @@ -56,6 +59,7 @@ define <vscale x 2 x i16> @intrinsic_vfcvt_rtz.x.f.v_nxv2i16_nxv2f16(<vscale x 2
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i16> @llvm.riscv.vfcvt.rtz.x.f.v.nxv2i16.nxv2f16(
<vscale x 2 x i16> undef,
<vscale x 2 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -86,6 +90,7 @@ entry:
}

declare <vscale x 4 x i16> @llvm.riscv.vfcvt.rtz.x.f.v.nxv4i16.nxv4f16(
<vscale x 4 x i16>,
<vscale x 4 x half>,
iXLen);

Expand All @@ -97,6 +102,7 @@ define <vscale x 4 x i16> @intrinsic_vfcvt_rtz.x.f.v_nxv4i16_nxv4f16(<vscale x 4
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i16> @llvm.riscv.vfcvt.rtz.x.f.v.nxv4i16.nxv4f16(
<vscale x 4 x i16> undef,
<vscale x 4 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -127,6 +133,7 @@ entry:
}

declare <vscale x 8 x i16> @llvm.riscv.vfcvt.rtz.x.f.v.nxv8i16.nxv8f16(
<vscale x 8 x i16>,
<vscale x 8 x half>,
iXLen);

Expand All @@ -138,6 +145,7 @@ define <vscale x 8 x i16> @intrinsic_vfcvt_rtz.x.f.v_nxv8i16_nxv8f16(<vscale x 8
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i16> @llvm.riscv.vfcvt.rtz.x.f.v.nxv8i16.nxv8f16(
<vscale x 8 x i16> undef,
<vscale x 8 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -168,6 +176,7 @@ entry:
}

declare <vscale x 16 x i16> @llvm.riscv.vfcvt.rtz.x.f.v.nxv16i16.nxv16f16(
<vscale x 16 x i16>,
<vscale x 16 x half>,
iXLen);

Expand All @@ -179,6 +188,7 @@ define <vscale x 16 x i16> @intrinsic_vfcvt_rtz.x.f.v_nxv16i16_nxv16f16(<vscale
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i16> @llvm.riscv.vfcvt.rtz.x.f.v.nxv16i16.nxv16f16(
<vscale x 16 x i16> undef,
<vscale x 16 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -209,6 +219,7 @@ entry:
}

declare <vscale x 32 x i16> @llvm.riscv.vfcvt.rtz.x.f.v.nxv32i16.nxv32f16(
<vscale x 32 x i16>,
<vscale x 32 x half>,
iXLen);

Expand All @@ -220,6 +231,7 @@ define <vscale x 32 x i16> @intrinsic_vfcvt_rtz.x.f.v_nxv32i16_nxv32f16(<vscale
; CHECK-NEXT: ret
entry:
%a = call <vscale x 32 x i16> @llvm.riscv.vfcvt.rtz.x.f.v.nxv32i16.nxv32f16(
<vscale x 32 x i16> undef,
<vscale x 32 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -250,6 +262,7 @@ entry:
}

declare <vscale x 1 x i32> @llvm.riscv.vfcvt.rtz.x.f.v.nxv1i32.nxv1f32(
<vscale x 1 x i32>,
<vscale x 1 x float>,
iXLen);

Expand All @@ -261,6 +274,7 @@ define <vscale x 1 x i32> @intrinsic_vfcvt_rtz.x.f.v_nxv1i32_nxv1f32(<vscale x 1
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i32> @llvm.riscv.vfcvt.rtz.x.f.v.nxv1i32.nxv1f32(
<vscale x 1 x i32> undef,
<vscale x 1 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -291,6 +305,7 @@ entry:
}

declare <vscale x 2 x i32> @llvm.riscv.vfcvt.rtz.x.f.v.nxv2i32.nxv2f32(
<vscale x 2 x i32>,
<vscale x 2 x float>,
iXLen);

Expand All @@ -302,6 +317,7 @@ define <vscale x 2 x i32> @intrinsic_vfcvt_rtz.x.f.v_nxv2i32_nxv2f32(<vscale x 2
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i32> @llvm.riscv.vfcvt.rtz.x.f.v.nxv2i32.nxv2f32(
<vscale x 2 x i32> undef,
<vscale x 2 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -332,6 +348,7 @@ entry:
}

declare <vscale x 4 x i32> @llvm.riscv.vfcvt.rtz.x.f.v.nxv4i32.nxv4f32(
<vscale x 4 x i32>,
<vscale x 4 x float>,
iXLen);

Expand All @@ -343,6 +360,7 @@ define <vscale x 4 x i32> @intrinsic_vfcvt_rtz.x.f.v_nxv4i32_nxv4f32(<vscale x 4
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i32> @llvm.riscv.vfcvt.rtz.x.f.v.nxv4i32.nxv4f32(
<vscale x 4 x i32> undef,
<vscale x 4 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -373,6 +391,7 @@ entry:
}

declare <vscale x 8 x i32> @llvm.riscv.vfcvt.rtz.x.f.v.nxv8i32.nxv8f32(
<vscale x 8 x i32>,
<vscale x 8 x float>,
iXLen);

Expand All @@ -384,6 +403,7 @@ define <vscale x 8 x i32> @intrinsic_vfcvt_rtz.x.f.v_nxv8i32_nxv8f32(<vscale x 8
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i32> @llvm.riscv.vfcvt.rtz.x.f.v.nxv8i32.nxv8f32(
<vscale x 8 x i32> undef,
<vscale x 8 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -414,6 +434,7 @@ entry:
}

declare <vscale x 16 x i32> @llvm.riscv.vfcvt.rtz.x.f.v.nxv16i32.nxv16f32(
<vscale x 16 x i32>,
<vscale x 16 x float>,
iXLen);

Expand All @@ -425,6 +446,7 @@ define <vscale x 16 x i32> @intrinsic_vfcvt_rtz.x.f.v_nxv16i32_nxv16f32(<vscale
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i32> @llvm.riscv.vfcvt.rtz.x.f.v.nxv16i32.nxv16f32(
<vscale x 16 x i32> undef,
<vscale x 16 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -455,6 +477,7 @@ entry:
}

declare <vscale x 1 x i64> @llvm.riscv.vfcvt.rtz.x.f.v.nxv1i64.nxv1f64(
<vscale x 1 x i64>,
<vscale x 1 x double>,
iXLen);

Expand All @@ -466,6 +489,7 @@ define <vscale x 1 x i64> @intrinsic_vfcvt_rtz.x.f.v_nxv1i64_nxv1f64(<vscale x 1
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i64> @llvm.riscv.vfcvt.rtz.x.f.v.nxv1i64.nxv1f64(
<vscale x 1 x i64> undef,
<vscale x 1 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -496,6 +520,7 @@ entry:
}

declare <vscale x 2 x i64> @llvm.riscv.vfcvt.rtz.x.f.v.nxv2i64.nxv2f64(
<vscale x 2 x i64>,
<vscale x 2 x double>,
iXLen);

Expand All @@ -507,6 +532,7 @@ define <vscale x 2 x i64> @intrinsic_vfcvt_rtz.x.f.v_nxv2i64_nxv2f64(<vscale x 2
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i64> @llvm.riscv.vfcvt.rtz.x.f.v.nxv2i64.nxv2f64(
<vscale x 2 x i64> undef,
<vscale x 2 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -537,6 +563,7 @@ entry:
}

declare <vscale x 4 x i64> @llvm.riscv.vfcvt.rtz.x.f.v.nxv4i64.nxv4f64(
<vscale x 4 x i64>,
<vscale x 4 x double>,
iXLen);

Expand All @@ -548,6 +575,7 @@ define <vscale x 4 x i64> @intrinsic_vfcvt_rtz.x.f.v_nxv4i64_nxv4f64(<vscale x 4
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i64> @llvm.riscv.vfcvt.rtz.x.f.v.nxv4i64.nxv4f64(
<vscale x 4 x i64> undef,
<vscale x 4 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -578,6 +606,7 @@ entry:
}

declare <vscale x 8 x i64> @llvm.riscv.vfcvt.rtz.x.f.v.nxv8i64.nxv8f64(
<vscale x 8 x i64>,
<vscale x 8 x double>,
iXLen);

Expand All @@ -589,6 +618,7 @@ define <vscale x 8 x i64> @intrinsic_vfcvt_rtz.x.f.v_nxv8i64_nxv8f64(<vscale x 8
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i64> @llvm.riscv.vfcvt.rtz.x.f.v.nxv8i64.nxv8f64(
<vscale x 8 x i64> undef,
<vscale x 8 x double> %0,
iXLen %1)

Expand Down
30 changes: 30 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfcvt-rtz-xu-f.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x i16> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv1i16.nxv1f16(
<vscale x 1 x i16>,
<vscale x 1 x half>,
iXLen);

Expand All @@ -15,6 +16,7 @@ define <vscale x 1 x i16> @intrinsic_vfcvt_rtz.xu.f.v_nxv1i16_nxv1f16(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i16> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv1i16.nxv1f16(
<vscale x 1 x i16> undef,
<vscale x 1 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -45,6 +47,7 @@ entry:
}

declare <vscale x 2 x i16> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv2i16.nxv2f16(
<vscale x 2 x i16>,
<vscale x 2 x half>,
iXLen);

Expand All @@ -56,6 +59,7 @@ define <vscale x 2 x i16> @intrinsic_vfcvt_rtz.xu.f.v_nxv2i16_nxv2f16(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i16> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv2i16.nxv2f16(
<vscale x 2 x i16> undef,
<vscale x 2 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -86,6 +90,7 @@ entry:
}

declare <vscale x 4 x i16> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv4i16.nxv4f16(
<vscale x 4 x i16>,
<vscale x 4 x half>,
iXLen);

Expand All @@ -97,6 +102,7 @@ define <vscale x 4 x i16> @intrinsic_vfcvt_rtz.xu.f.v_nxv4i16_nxv4f16(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i16> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv4i16.nxv4f16(
<vscale x 4 x i16> undef,
<vscale x 4 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -127,6 +133,7 @@ entry:
}

declare <vscale x 8 x i16> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv8i16.nxv8f16(
<vscale x 8 x i16>,
<vscale x 8 x half>,
iXLen);

Expand All @@ -138,6 +145,7 @@ define <vscale x 8 x i16> @intrinsic_vfcvt_rtz.xu.f.v_nxv8i16_nxv8f16(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i16> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv8i16.nxv8f16(
<vscale x 8 x i16> undef,
<vscale x 8 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -168,6 +176,7 @@ entry:
}

declare <vscale x 16 x i16> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv16i16.nxv16f16(
<vscale x 16 x i16>,
<vscale x 16 x half>,
iXLen);

Expand All @@ -179,6 +188,7 @@ define <vscale x 16 x i16> @intrinsic_vfcvt_rtz.xu.f.v_nxv16i16_nxv16f16(<vscale
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i16> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv16i16.nxv16f16(
<vscale x 16 x i16> undef,
<vscale x 16 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -209,6 +219,7 @@ entry:
}

declare <vscale x 32 x i16> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv32i16.nxv32f16(
<vscale x 32 x i16>,
<vscale x 32 x half>,
iXLen);

Expand All @@ -220,6 +231,7 @@ define <vscale x 32 x i16> @intrinsic_vfcvt_rtz.xu.f.v_nxv32i16_nxv32f16(<vscale
; CHECK-NEXT: ret
entry:
%a = call <vscale x 32 x i16> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv32i16.nxv32f16(
<vscale x 32 x i16> undef,
<vscale x 32 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -250,6 +262,7 @@ entry:
}

declare <vscale x 1 x i32> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv1i32.nxv1f32(
<vscale x 1 x i32>,
<vscale x 1 x float>,
iXLen);

Expand All @@ -261,6 +274,7 @@ define <vscale x 1 x i32> @intrinsic_vfcvt_rtz.xu.f.v_nxv1i32_nxv1f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i32> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv1i32.nxv1f32(
<vscale x 1 x i32> undef,
<vscale x 1 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -291,6 +305,7 @@ entry:
}

declare <vscale x 2 x i32> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv2i32.nxv2f32(
<vscale x 2 x i32>,
<vscale x 2 x float>,
iXLen);

Expand All @@ -302,6 +317,7 @@ define <vscale x 2 x i32> @intrinsic_vfcvt_rtz.xu.f.v_nxv2i32_nxv2f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i32> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv2i32.nxv2f32(
<vscale x 2 x i32> undef,
<vscale x 2 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -332,6 +348,7 @@ entry:
}

declare <vscale x 4 x i32> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv4i32.nxv4f32(
<vscale x 4 x i32>,
<vscale x 4 x float>,
iXLen);

Expand All @@ -343,6 +360,7 @@ define <vscale x 4 x i32> @intrinsic_vfcvt_rtz.xu.f.v_nxv4i32_nxv4f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i32> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv4i32.nxv4f32(
<vscale x 4 x i32> undef,
<vscale x 4 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -373,6 +391,7 @@ entry:
}

declare <vscale x 8 x i32> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv8i32.nxv8f32(
<vscale x 8 x i32>,
<vscale x 8 x float>,
iXLen);

Expand All @@ -384,6 +403,7 @@ define <vscale x 8 x i32> @intrinsic_vfcvt_rtz.xu.f.v_nxv8i32_nxv8f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i32> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv8i32.nxv8f32(
<vscale x 8 x i32> undef,
<vscale x 8 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -414,6 +434,7 @@ entry:
}

declare <vscale x 16 x i32> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv16i32.nxv16f32(
<vscale x 16 x i32>,
<vscale x 16 x float>,
iXLen);

Expand All @@ -425,6 +446,7 @@ define <vscale x 16 x i32> @intrinsic_vfcvt_rtz.xu.f.v_nxv16i32_nxv16f32(<vscale
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i32> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv16i32.nxv16f32(
<vscale x 16 x i32> undef,
<vscale x 16 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -455,6 +477,7 @@ entry:
}

declare <vscale x 1 x i64> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv1i64.nxv1f64(
<vscale x 1 x i64>,
<vscale x 1 x double>,
iXLen);

Expand All @@ -466,6 +489,7 @@ define <vscale x 1 x i64> @intrinsic_vfcvt_rtz.xu.f.v_nxv1i64_nxv1f64(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i64> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv1i64.nxv1f64(
<vscale x 1 x i64> undef,
<vscale x 1 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -496,6 +520,7 @@ entry:
}

declare <vscale x 2 x i64> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv2i64.nxv2f64(
<vscale x 2 x i64>,
<vscale x 2 x double>,
iXLen);

Expand All @@ -507,6 +532,7 @@ define <vscale x 2 x i64> @intrinsic_vfcvt_rtz.xu.f.v_nxv2i64_nxv2f64(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i64> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv2i64.nxv2f64(
<vscale x 2 x i64> undef,
<vscale x 2 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -537,6 +563,7 @@ entry:
}

declare <vscale x 4 x i64> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv4i64.nxv4f64(
<vscale x 4 x i64>,
<vscale x 4 x double>,
iXLen);

Expand All @@ -548,6 +575,7 @@ define <vscale x 4 x i64> @intrinsic_vfcvt_rtz.xu.f.v_nxv4i64_nxv4f64(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i64> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv4i64.nxv4f64(
<vscale x 4 x i64> undef,
<vscale x 4 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -578,6 +606,7 @@ entry:
}

declare <vscale x 8 x i64> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv8i64.nxv8f64(
<vscale x 8 x i64>,
<vscale x 8 x double>,
iXLen);

Expand All @@ -589,6 +618,7 @@ define <vscale x 8 x i64> @intrinsic_vfcvt_rtz.xu.f.v_nxv8i64_nxv8f64(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i64> @llvm.riscv.vfcvt.rtz.xu.f.v.nxv8i64.nxv8f64(
<vscale x 8 x i64> undef,
<vscale x 8 x double> %0,
iXLen %1)

Expand Down
30 changes: 30 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfcvt-x-f.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x i16> @llvm.riscv.vfcvt.x.f.v.nxv1i16.nxv1f16(
<vscale x 1 x i16>,
<vscale x 1 x half>,
iXLen);

Expand All @@ -15,6 +16,7 @@ define <vscale x 1 x i16> @intrinsic_vfcvt_x.f.v_nxv1i16_nxv1f16(<vscale x 1 x h
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i16> @llvm.riscv.vfcvt.x.f.v.nxv1i16.nxv1f16(
<vscale x 1 x i16> undef,
<vscale x 1 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -45,6 +47,7 @@ entry:
}

declare <vscale x 2 x i16> @llvm.riscv.vfcvt.x.f.v.nxv2i16.nxv2f16(
<vscale x 2 x i16>,
<vscale x 2 x half>,
iXLen);

Expand All @@ -56,6 +59,7 @@ define <vscale x 2 x i16> @intrinsic_vfcvt_x.f.v_nxv2i16_nxv2f16(<vscale x 2 x h
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i16> @llvm.riscv.vfcvt.x.f.v.nxv2i16.nxv2f16(
<vscale x 2 x i16> undef,
<vscale x 2 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -86,6 +90,7 @@ entry:
}

declare <vscale x 4 x i16> @llvm.riscv.vfcvt.x.f.v.nxv4i16.nxv4f16(
<vscale x 4 x i16>,
<vscale x 4 x half>,
iXLen);

Expand All @@ -97,6 +102,7 @@ define <vscale x 4 x i16> @intrinsic_vfcvt_x.f.v_nxv4i16_nxv4f16(<vscale x 4 x h
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i16> @llvm.riscv.vfcvt.x.f.v.nxv4i16.nxv4f16(
<vscale x 4 x i16> undef,
<vscale x 4 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -127,6 +133,7 @@ entry:
}

declare <vscale x 8 x i16> @llvm.riscv.vfcvt.x.f.v.nxv8i16.nxv8f16(
<vscale x 8 x i16>,
<vscale x 8 x half>,
iXLen);

Expand All @@ -138,6 +145,7 @@ define <vscale x 8 x i16> @intrinsic_vfcvt_x.f.v_nxv8i16_nxv8f16(<vscale x 8 x h
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i16> @llvm.riscv.vfcvt.x.f.v.nxv8i16.nxv8f16(
<vscale x 8 x i16> undef,
<vscale x 8 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -168,6 +176,7 @@ entry:
}

declare <vscale x 16 x i16> @llvm.riscv.vfcvt.x.f.v.nxv16i16.nxv16f16(
<vscale x 16 x i16>,
<vscale x 16 x half>,
iXLen);

Expand All @@ -179,6 +188,7 @@ define <vscale x 16 x i16> @intrinsic_vfcvt_x.f.v_nxv16i16_nxv16f16(<vscale x 16
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i16> @llvm.riscv.vfcvt.x.f.v.nxv16i16.nxv16f16(
<vscale x 16 x i16> undef,
<vscale x 16 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -209,6 +219,7 @@ entry:
}

declare <vscale x 32 x i16> @llvm.riscv.vfcvt.x.f.v.nxv32i16.nxv32f16(
<vscale x 32 x i16>,
<vscale x 32 x half>,
iXLen);

Expand All @@ -220,6 +231,7 @@ define <vscale x 32 x i16> @intrinsic_vfcvt_x.f.v_nxv32i16_nxv32f16(<vscale x 32
; CHECK-NEXT: ret
entry:
%a = call <vscale x 32 x i16> @llvm.riscv.vfcvt.x.f.v.nxv32i16.nxv32f16(
<vscale x 32 x i16> undef,
<vscale x 32 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -250,6 +262,7 @@ entry:
}

declare <vscale x 1 x i32> @llvm.riscv.vfcvt.x.f.v.nxv1i32.nxv1f32(
<vscale x 1 x i32>,
<vscale x 1 x float>,
iXLen);

Expand All @@ -261,6 +274,7 @@ define <vscale x 1 x i32> @intrinsic_vfcvt_x.f.v_nxv1i32_nxv1f32(<vscale x 1 x f
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i32> @llvm.riscv.vfcvt.x.f.v.nxv1i32.nxv1f32(
<vscale x 1 x i32> undef,
<vscale x 1 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -291,6 +305,7 @@ entry:
}

declare <vscale x 2 x i32> @llvm.riscv.vfcvt.x.f.v.nxv2i32.nxv2f32(
<vscale x 2 x i32>,
<vscale x 2 x float>,
iXLen);

Expand All @@ -302,6 +317,7 @@ define <vscale x 2 x i32> @intrinsic_vfcvt_x.f.v_nxv2i32_nxv2f32(<vscale x 2 x f
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i32> @llvm.riscv.vfcvt.x.f.v.nxv2i32.nxv2f32(
<vscale x 2 x i32> undef,
<vscale x 2 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -332,6 +348,7 @@ entry:
}

declare <vscale x 4 x i32> @llvm.riscv.vfcvt.x.f.v.nxv4i32.nxv4f32(
<vscale x 4 x i32>,
<vscale x 4 x float>,
iXLen);

Expand All @@ -343,6 +360,7 @@ define <vscale x 4 x i32> @intrinsic_vfcvt_x.f.v_nxv4i32_nxv4f32(<vscale x 4 x f
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i32> @llvm.riscv.vfcvt.x.f.v.nxv4i32.nxv4f32(
<vscale x 4 x i32> undef,
<vscale x 4 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -373,6 +391,7 @@ entry:
}

declare <vscale x 8 x i32> @llvm.riscv.vfcvt.x.f.v.nxv8i32.nxv8f32(
<vscale x 8 x i32>,
<vscale x 8 x float>,
iXLen);

Expand All @@ -384,6 +403,7 @@ define <vscale x 8 x i32> @intrinsic_vfcvt_x.f.v_nxv8i32_nxv8f32(<vscale x 8 x f
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i32> @llvm.riscv.vfcvt.x.f.v.nxv8i32.nxv8f32(
<vscale x 8 x i32> undef,
<vscale x 8 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -414,6 +434,7 @@ entry:
}

declare <vscale x 16 x i32> @llvm.riscv.vfcvt.x.f.v.nxv16i32.nxv16f32(
<vscale x 16 x i32>,
<vscale x 16 x float>,
iXLen);

Expand All @@ -425,6 +446,7 @@ define <vscale x 16 x i32> @intrinsic_vfcvt_x.f.v_nxv16i32_nxv16f32(<vscale x 16
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i32> @llvm.riscv.vfcvt.x.f.v.nxv16i32.nxv16f32(
<vscale x 16 x i32> undef,
<vscale x 16 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -455,6 +477,7 @@ entry:
}

declare <vscale x 1 x i64> @llvm.riscv.vfcvt.x.f.v.nxv1i64.nxv1f64(
<vscale x 1 x i64>,
<vscale x 1 x double>,
iXLen);

Expand All @@ -466,6 +489,7 @@ define <vscale x 1 x i64> @intrinsic_vfcvt_x.f.v_nxv1i64_nxv1f64(<vscale x 1 x d
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i64> @llvm.riscv.vfcvt.x.f.v.nxv1i64.nxv1f64(
<vscale x 1 x i64> undef,
<vscale x 1 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -496,6 +520,7 @@ entry:
}

declare <vscale x 2 x i64> @llvm.riscv.vfcvt.x.f.v.nxv2i64.nxv2f64(
<vscale x 2 x i64>,
<vscale x 2 x double>,
iXLen);

Expand All @@ -507,6 +532,7 @@ define <vscale x 2 x i64> @intrinsic_vfcvt_x.f.v_nxv2i64_nxv2f64(<vscale x 2 x d
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i64> @llvm.riscv.vfcvt.x.f.v.nxv2i64.nxv2f64(
<vscale x 2 x i64> undef,
<vscale x 2 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -537,6 +563,7 @@ entry:
}

declare <vscale x 4 x i64> @llvm.riscv.vfcvt.x.f.v.nxv4i64.nxv4f64(
<vscale x 4 x i64>,
<vscale x 4 x double>,
iXLen);

Expand All @@ -548,6 +575,7 @@ define <vscale x 4 x i64> @intrinsic_vfcvt_x.f.v_nxv4i64_nxv4f64(<vscale x 4 x d
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i64> @llvm.riscv.vfcvt.x.f.v.nxv4i64.nxv4f64(
<vscale x 4 x i64> undef,
<vscale x 4 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -578,6 +606,7 @@ entry:
}

declare <vscale x 8 x i64> @llvm.riscv.vfcvt.x.f.v.nxv8i64.nxv8f64(
<vscale x 8 x i64>,
<vscale x 8 x double>,
iXLen);

Expand All @@ -589,6 +618,7 @@ define <vscale x 8 x i64> @intrinsic_vfcvt_x.f.v_nxv8i64_nxv8f64(<vscale x 8 x d
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i64> @llvm.riscv.vfcvt.x.f.v.nxv8i64.nxv8f64(
<vscale x 8 x i64> undef,
<vscale x 8 x double> %0,
iXLen %1)

Expand Down
30 changes: 30 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfcvt-xu-f.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x i16> @llvm.riscv.vfcvt.xu.f.v.nxv1i16.nxv1f16(
<vscale x 1 x i16>,
<vscale x 1 x half>,
iXLen);

Expand All @@ -15,6 +16,7 @@ define <vscale x 1 x i16> @intrinsic_vfcvt_xu.f.v_nxv1i16_nxv1f16(<vscale x 1 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i16> @llvm.riscv.vfcvt.xu.f.v.nxv1i16.nxv1f16(
<vscale x 1 x i16> undef,
<vscale x 1 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -45,6 +47,7 @@ entry:
}

declare <vscale x 2 x i16> @llvm.riscv.vfcvt.xu.f.v.nxv2i16.nxv2f16(
<vscale x 2 x i16>,
<vscale x 2 x half>,
iXLen);

Expand All @@ -56,6 +59,7 @@ define <vscale x 2 x i16> @intrinsic_vfcvt_xu.f.v_nxv2i16_nxv2f16(<vscale x 2 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i16> @llvm.riscv.vfcvt.xu.f.v.nxv2i16.nxv2f16(
<vscale x 2 x i16> undef,
<vscale x 2 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -86,6 +90,7 @@ entry:
}

declare <vscale x 4 x i16> @llvm.riscv.vfcvt.xu.f.v.nxv4i16.nxv4f16(
<vscale x 4 x i16>,
<vscale x 4 x half>,
iXLen);

Expand All @@ -97,6 +102,7 @@ define <vscale x 4 x i16> @intrinsic_vfcvt_xu.f.v_nxv4i16_nxv4f16(<vscale x 4 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i16> @llvm.riscv.vfcvt.xu.f.v.nxv4i16.nxv4f16(
<vscale x 4 x i16> undef,
<vscale x 4 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -127,6 +133,7 @@ entry:
}

declare <vscale x 8 x i16> @llvm.riscv.vfcvt.xu.f.v.nxv8i16.nxv8f16(
<vscale x 8 x i16>,
<vscale x 8 x half>,
iXLen);

Expand All @@ -138,6 +145,7 @@ define <vscale x 8 x i16> @intrinsic_vfcvt_xu.f.v_nxv8i16_nxv8f16(<vscale x 8 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i16> @llvm.riscv.vfcvt.xu.f.v.nxv8i16.nxv8f16(
<vscale x 8 x i16> undef,
<vscale x 8 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -168,6 +176,7 @@ entry:
}

declare <vscale x 16 x i16> @llvm.riscv.vfcvt.xu.f.v.nxv16i16.nxv16f16(
<vscale x 16 x i16>,
<vscale x 16 x half>,
iXLen);

Expand All @@ -179,6 +188,7 @@ define <vscale x 16 x i16> @intrinsic_vfcvt_xu.f.v_nxv16i16_nxv16f16(<vscale x 1
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i16> @llvm.riscv.vfcvt.xu.f.v.nxv16i16.nxv16f16(
<vscale x 16 x i16> undef,
<vscale x 16 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -209,6 +219,7 @@ entry:
}

declare <vscale x 32 x i16> @llvm.riscv.vfcvt.xu.f.v.nxv32i16.nxv32f16(
<vscale x 32 x i16>,
<vscale x 32 x half>,
iXLen);

Expand All @@ -220,6 +231,7 @@ define <vscale x 32 x i16> @intrinsic_vfcvt_xu.f.v_nxv32i16_nxv32f16(<vscale x 3
; CHECK-NEXT: ret
entry:
%a = call <vscale x 32 x i16> @llvm.riscv.vfcvt.xu.f.v.nxv32i16.nxv32f16(
<vscale x 32 x i16> undef,
<vscale x 32 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -250,6 +262,7 @@ entry:
}

declare <vscale x 1 x i32> @llvm.riscv.vfcvt.xu.f.v.nxv1i32.nxv1f32(
<vscale x 1 x i32>,
<vscale x 1 x float>,
iXLen);

Expand All @@ -261,6 +274,7 @@ define <vscale x 1 x i32> @intrinsic_vfcvt_xu.f.v_nxv1i32_nxv1f32(<vscale x 1 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i32> @llvm.riscv.vfcvt.xu.f.v.nxv1i32.nxv1f32(
<vscale x 1 x i32> undef,
<vscale x 1 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -291,6 +305,7 @@ entry:
}

declare <vscale x 2 x i32> @llvm.riscv.vfcvt.xu.f.v.nxv2i32.nxv2f32(
<vscale x 2 x i32>,
<vscale x 2 x float>,
iXLen);

Expand All @@ -302,6 +317,7 @@ define <vscale x 2 x i32> @intrinsic_vfcvt_xu.f.v_nxv2i32_nxv2f32(<vscale x 2 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i32> @llvm.riscv.vfcvt.xu.f.v.nxv2i32.nxv2f32(
<vscale x 2 x i32> undef,
<vscale x 2 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -332,6 +348,7 @@ entry:
}

declare <vscale x 4 x i32> @llvm.riscv.vfcvt.xu.f.v.nxv4i32.nxv4f32(
<vscale x 4 x i32>,
<vscale x 4 x float>,
iXLen);

Expand All @@ -343,6 +360,7 @@ define <vscale x 4 x i32> @intrinsic_vfcvt_xu.f.v_nxv4i32_nxv4f32(<vscale x 4 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i32> @llvm.riscv.vfcvt.xu.f.v.nxv4i32.nxv4f32(
<vscale x 4 x i32> undef,
<vscale x 4 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -373,6 +391,7 @@ entry:
}

declare <vscale x 8 x i32> @llvm.riscv.vfcvt.xu.f.v.nxv8i32.nxv8f32(
<vscale x 8 x i32>,
<vscale x 8 x float>,
iXLen);

Expand All @@ -384,6 +403,7 @@ define <vscale x 8 x i32> @intrinsic_vfcvt_xu.f.v_nxv8i32_nxv8f32(<vscale x 8 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i32> @llvm.riscv.vfcvt.xu.f.v.nxv8i32.nxv8f32(
<vscale x 8 x i32> undef,
<vscale x 8 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -414,6 +434,7 @@ entry:
}

declare <vscale x 16 x i32> @llvm.riscv.vfcvt.xu.f.v.nxv16i32.nxv16f32(
<vscale x 16 x i32>,
<vscale x 16 x float>,
iXLen);

Expand All @@ -425,6 +446,7 @@ define <vscale x 16 x i32> @intrinsic_vfcvt_xu.f.v_nxv16i32_nxv16f32(<vscale x 1
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i32> @llvm.riscv.vfcvt.xu.f.v.nxv16i32.nxv16f32(
<vscale x 16 x i32> undef,
<vscale x 16 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -455,6 +477,7 @@ entry:
}

declare <vscale x 1 x i64> @llvm.riscv.vfcvt.xu.f.v.nxv1i64.nxv1f64(
<vscale x 1 x i64>,
<vscale x 1 x double>,
iXLen);

Expand All @@ -466,6 +489,7 @@ define <vscale x 1 x i64> @intrinsic_vfcvt_xu.f.v_nxv1i64_nxv1f64(<vscale x 1 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i64> @llvm.riscv.vfcvt.xu.f.v.nxv1i64.nxv1f64(
<vscale x 1 x i64> undef,
<vscale x 1 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -496,6 +520,7 @@ entry:
}

declare <vscale x 2 x i64> @llvm.riscv.vfcvt.xu.f.v.nxv2i64.nxv2f64(
<vscale x 2 x i64>,
<vscale x 2 x double>,
iXLen);

Expand All @@ -507,6 +532,7 @@ define <vscale x 2 x i64> @intrinsic_vfcvt_xu.f.v_nxv2i64_nxv2f64(<vscale x 2 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i64> @llvm.riscv.vfcvt.xu.f.v.nxv2i64.nxv2f64(
<vscale x 2 x i64> undef,
<vscale x 2 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -537,6 +563,7 @@ entry:
}

declare <vscale x 4 x i64> @llvm.riscv.vfcvt.xu.f.v.nxv4i64.nxv4f64(
<vscale x 4 x i64>,
<vscale x 4 x double>,
iXLen);

Expand All @@ -548,6 +575,7 @@ define <vscale x 4 x i64> @intrinsic_vfcvt_xu.f.v_nxv4i64_nxv4f64(<vscale x 4 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i64> @llvm.riscv.vfcvt.xu.f.v.nxv4i64.nxv4f64(
<vscale x 4 x i64> undef,
<vscale x 4 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -578,6 +606,7 @@ entry:
}

declare <vscale x 8 x i64> @llvm.riscv.vfcvt.xu.f.v.nxv8i64.nxv8f64(
<vscale x 8 x i64>,
<vscale x 8 x double>,
iXLen);

Expand All @@ -589,6 +618,7 @@ define <vscale x 8 x i64> @intrinsic_vfcvt_xu.f.v_nxv8i64_nxv8f64(<vscale x 8 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i64> @llvm.riscv.vfcvt.xu.f.v.nxv8i64.nxv8f64(
<vscale x 8 x i64> undef,
<vscale x 8 x double> %0,
iXLen %1)

Expand Down
18 changes: 18 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfncvt-f-f.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x half> @llvm.riscv.vfncvt.f.f.w.nxv1f16.nxv1f32(
<vscale x 1 x half>,
<vscale x 1 x float>,
iXLen);

Expand All @@ -16,6 +17,7 @@ define <vscale x 1 x half> @intrinsic_vfncvt_f.f.w_nxv1f16_nxv1f32(<vscale x 1 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x half> @llvm.riscv.vfncvt.f.f.w.nxv1f16.nxv1f32(
<vscale x 1 x half> undef,
<vscale x 1 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -46,6 +48,7 @@ entry:
}

declare <vscale x 2 x half> @llvm.riscv.vfncvt.f.f.w.nxv2f16.nxv2f32(
<vscale x 2 x half>,
<vscale x 2 x float>,
iXLen);

Expand All @@ -58,6 +61,7 @@ define <vscale x 2 x half> @intrinsic_vfncvt_f.f.w_nxv2f16_nxv2f32(<vscale x 2 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x half> @llvm.riscv.vfncvt.f.f.w.nxv2f16.nxv2f32(
<vscale x 2 x half> undef,
<vscale x 2 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -88,6 +92,7 @@ entry:
}

declare <vscale x 4 x half> @llvm.riscv.vfncvt.f.f.w.nxv4f16.nxv4f32(
<vscale x 4 x half>,
<vscale x 4 x float>,
iXLen);

Expand All @@ -100,6 +105,7 @@ define <vscale x 4 x half> @intrinsic_vfncvt_f.f.w_nxv4f16_nxv4f32(<vscale x 4 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x half> @llvm.riscv.vfncvt.f.f.w.nxv4f16.nxv4f32(
<vscale x 4 x half> undef,
<vscale x 4 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -130,6 +136,7 @@ entry:
}

declare <vscale x 8 x half> @llvm.riscv.vfncvt.f.f.w.nxv8f16.nxv8f32(
<vscale x 8 x half>,
<vscale x 8 x float>,
iXLen);

Expand All @@ -142,6 +149,7 @@ define <vscale x 8 x half> @intrinsic_vfncvt_f.f.w_nxv8f16_nxv8f32(<vscale x 8 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x half> @llvm.riscv.vfncvt.f.f.w.nxv8f16.nxv8f32(
<vscale x 8 x half> undef,
<vscale x 8 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -172,6 +180,7 @@ entry:
}

declare <vscale x 16 x half> @llvm.riscv.vfncvt.f.f.w.nxv16f16.nxv16f32(
<vscale x 16 x half>,
<vscale x 16 x float>,
iXLen);

Expand All @@ -184,6 +193,7 @@ define <vscale x 16 x half> @intrinsic_vfncvt_f.f.w_nxv16f16_nxv16f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x half> @llvm.riscv.vfncvt.f.f.w.nxv16f16.nxv16f32(
<vscale x 16 x half> undef,
<vscale x 16 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -214,6 +224,7 @@ entry:
}

declare <vscale x 1 x float> @llvm.riscv.vfncvt.f.f.w.nxv1f32.nxv1f64(
<vscale x 1 x float>,
<vscale x 1 x double>,
iXLen);

Expand All @@ -226,6 +237,7 @@ define <vscale x 1 x float> @intrinsic_vfncvt_f.f.w_nxv1f32_nxv1f64(<vscale x 1
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x float> @llvm.riscv.vfncvt.f.f.w.nxv1f32.nxv1f64(
<vscale x 1 x float> undef,
<vscale x 1 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -256,6 +268,7 @@ entry:
}

declare <vscale x 2 x float> @llvm.riscv.vfncvt.f.f.w.nxv2f32.nxv2f64(
<vscale x 2 x float>,
<vscale x 2 x double>,
iXLen);

Expand All @@ -268,6 +281,7 @@ define <vscale x 2 x float> @intrinsic_vfncvt_f.f.w_nxv2f32_nxv2f64(<vscale x 2
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x float> @llvm.riscv.vfncvt.f.f.w.nxv2f32.nxv2f64(
<vscale x 2 x float> undef,
<vscale x 2 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -298,6 +312,7 @@ entry:
}

declare <vscale x 4 x float> @llvm.riscv.vfncvt.f.f.w.nxv4f32.nxv4f64(
<vscale x 4 x float>,
<vscale x 4 x double>,
iXLen);

Expand All @@ -310,6 +325,7 @@ define <vscale x 4 x float> @intrinsic_vfncvt_f.f.w_nxv4f32_nxv4f64(<vscale x 4
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x float> @llvm.riscv.vfncvt.f.f.w.nxv4f32.nxv4f64(
<vscale x 4 x float> undef,
<vscale x 4 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -340,6 +356,7 @@ entry:
}

declare <vscale x 8 x float> @llvm.riscv.vfncvt.f.f.w.nxv8f32.nxv8f64(
<vscale x 8 x float>,
<vscale x 8 x double>,
iXLen);

Expand All @@ -352,6 +369,7 @@ define <vscale x 8 x float> @intrinsic_vfncvt_f.f.w_nxv8f32_nxv8f64(<vscale x 8
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x float> @llvm.riscv.vfncvt.f.f.w.nxv8f32.nxv8f64(
<vscale x 8 x float> undef,
<vscale x 8 x double> %0,
iXLen %1)

Expand Down
18 changes: 18 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfncvt-f-x.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x half> @llvm.riscv.vfncvt.f.x.w.nxv1f16.nxv1i32(
<vscale x 1 x half>,
<vscale x 1 x i32>,
iXLen);

Expand All @@ -16,6 +17,7 @@ define <vscale x 1 x half> @intrinsic_vfncvt_f.x.w_nxv1f16_nxv1i32(<vscale x 1 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x half> @llvm.riscv.vfncvt.f.x.w.nxv1f16.nxv1i32(
<vscale x 1 x half> undef,
<vscale x 1 x i32> %0,
iXLen %1)

Expand Down Expand Up @@ -46,6 +48,7 @@ entry:
}

declare <vscale x 2 x half> @llvm.riscv.vfncvt.f.x.w.nxv2f16.nxv2i32(
<vscale x 2 x half>,
<vscale x 2 x i32>,
iXLen);

Expand All @@ -58,6 +61,7 @@ define <vscale x 2 x half> @intrinsic_vfncvt_f.x.w_nxv2f16_nxv2i32(<vscale x 2 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x half> @llvm.riscv.vfncvt.f.x.w.nxv2f16.nxv2i32(
<vscale x 2 x half> undef,
<vscale x 2 x i32> %0,
iXLen %1)

Expand Down Expand Up @@ -88,6 +92,7 @@ entry:
}

declare <vscale x 4 x half> @llvm.riscv.vfncvt.f.x.w.nxv4f16.nxv4i32(
<vscale x 4 x half>,
<vscale x 4 x i32>,
iXLen);

Expand All @@ -100,6 +105,7 @@ define <vscale x 4 x half> @intrinsic_vfncvt_f.x.w_nxv4f16_nxv4i32(<vscale x 4 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x half> @llvm.riscv.vfncvt.f.x.w.nxv4f16.nxv4i32(
<vscale x 4 x half> undef,
<vscale x 4 x i32> %0,
iXLen %1)

Expand Down Expand Up @@ -130,6 +136,7 @@ entry:
}

declare <vscale x 8 x half> @llvm.riscv.vfncvt.f.x.w.nxv8f16.nxv8i32(
<vscale x 8 x half>,
<vscale x 8 x i32>,
iXLen);

Expand All @@ -142,6 +149,7 @@ define <vscale x 8 x half> @intrinsic_vfncvt_f.x.w_nxv8f16_nxv8i32(<vscale x 8 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x half> @llvm.riscv.vfncvt.f.x.w.nxv8f16.nxv8i32(
<vscale x 8 x half> undef,
<vscale x 8 x i32> %0,
iXLen %1)

Expand Down Expand Up @@ -172,6 +180,7 @@ entry:
}

declare <vscale x 16 x half> @llvm.riscv.vfncvt.f.x.w.nxv16f16.nxv16i32(
<vscale x 16 x half>,
<vscale x 16 x i32>,
iXLen);

Expand All @@ -184,6 +193,7 @@ define <vscale x 16 x half> @intrinsic_vfncvt_f.x.w_nxv16f16_nxv16i32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x half> @llvm.riscv.vfncvt.f.x.w.nxv16f16.nxv16i32(
<vscale x 16 x half> undef,
<vscale x 16 x i32> %0,
iXLen %1)

Expand Down Expand Up @@ -214,6 +224,7 @@ entry:
}

declare <vscale x 1 x float> @llvm.riscv.vfncvt.f.x.w.nxv1f32.nxv1i64(
<vscale x 1 x float>,
<vscale x 1 x i64>,
iXLen);

Expand All @@ -226,6 +237,7 @@ define <vscale x 1 x float> @intrinsic_vfncvt_f.x.w_nxv1f32_nxv1i64(<vscale x 1
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x float> @llvm.riscv.vfncvt.f.x.w.nxv1f32.nxv1i64(
<vscale x 1 x float> undef,
<vscale x 1 x i64> %0,
iXLen %1)

Expand Down Expand Up @@ -256,6 +268,7 @@ entry:
}

declare <vscale x 2 x float> @llvm.riscv.vfncvt.f.x.w.nxv2f32.nxv2i64(
<vscale x 2 x float>,
<vscale x 2 x i64>,
iXLen);

Expand All @@ -268,6 +281,7 @@ define <vscale x 2 x float> @intrinsic_vfncvt_f.x.w_nxv2f32_nxv2i64(<vscale x 2
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x float> @llvm.riscv.vfncvt.f.x.w.nxv2f32.nxv2i64(
<vscale x 2 x float> undef,
<vscale x 2 x i64> %0,
iXLen %1)

Expand Down Expand Up @@ -298,6 +312,7 @@ entry:
}

declare <vscale x 4 x float> @llvm.riscv.vfncvt.f.x.w.nxv4f32.nxv4i64(
<vscale x 4 x float>,
<vscale x 4 x i64>,
iXLen);

Expand All @@ -310,6 +325,7 @@ define <vscale x 4 x float> @intrinsic_vfncvt_f.x.w_nxv4f32_nxv4i64(<vscale x 4
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x float> @llvm.riscv.vfncvt.f.x.w.nxv4f32.nxv4i64(
<vscale x 4 x float> undef,
<vscale x 4 x i64> %0,
iXLen %1)

Expand Down Expand Up @@ -340,6 +356,7 @@ entry:
}

declare <vscale x 8 x float> @llvm.riscv.vfncvt.f.x.w.nxv8f32.nxv8i64(
<vscale x 8 x float>,
<vscale x 8 x i64>,
iXLen);

Expand All @@ -352,6 +369,7 @@ define <vscale x 8 x float> @intrinsic_vfncvt_f.x.w_nxv8f32_nxv8i64(<vscale x 8
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x float> @llvm.riscv.vfncvt.f.x.w.nxv8f32.nxv8i64(
<vscale x 8 x float> undef,
<vscale x 8 x i64> %0,
iXLen %1)

Expand Down
18 changes: 18 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfncvt-f-xu.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x half> @llvm.riscv.vfncvt.f.xu.w.nxv1f16.nxv1i32(
<vscale x 1 x half>,
<vscale x 1 x i32>,
iXLen);

Expand All @@ -16,6 +17,7 @@ define <vscale x 1 x half> @intrinsic_vfncvt_f.xu.w_nxv1f16_nxv1i32(<vscale x 1
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x half> @llvm.riscv.vfncvt.f.xu.w.nxv1f16.nxv1i32(
<vscale x 1 x half> undef,
<vscale x 1 x i32> %0,
iXLen %1)

Expand Down Expand Up @@ -46,6 +48,7 @@ entry:
}

declare <vscale x 2 x half> @llvm.riscv.vfncvt.f.xu.w.nxv2f16.nxv2i32(
<vscale x 2 x half>,
<vscale x 2 x i32>,
iXLen);

Expand All @@ -58,6 +61,7 @@ define <vscale x 2 x half> @intrinsic_vfncvt_f.xu.w_nxv2f16_nxv2i32(<vscale x 2
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x half> @llvm.riscv.vfncvt.f.xu.w.nxv2f16.nxv2i32(
<vscale x 2 x half> undef,
<vscale x 2 x i32> %0,
iXLen %1)

Expand Down Expand Up @@ -88,6 +92,7 @@ entry:
}

declare <vscale x 4 x half> @llvm.riscv.vfncvt.f.xu.w.nxv4f16.nxv4i32(
<vscale x 4 x half>,
<vscale x 4 x i32>,
iXLen);

Expand All @@ -100,6 +105,7 @@ define <vscale x 4 x half> @intrinsic_vfncvt_f.xu.w_nxv4f16_nxv4i32(<vscale x 4
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x half> @llvm.riscv.vfncvt.f.xu.w.nxv4f16.nxv4i32(
<vscale x 4 x half> undef,
<vscale x 4 x i32> %0,
iXLen %1)

Expand Down Expand Up @@ -130,6 +136,7 @@ entry:
}

declare <vscale x 8 x half> @llvm.riscv.vfncvt.f.xu.w.nxv8f16.nxv8i32(
<vscale x 8 x half>,
<vscale x 8 x i32>,
iXLen);

Expand All @@ -142,6 +149,7 @@ define <vscale x 8 x half> @intrinsic_vfncvt_f.xu.w_nxv8f16_nxv8i32(<vscale x 8
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x half> @llvm.riscv.vfncvt.f.xu.w.nxv8f16.nxv8i32(
<vscale x 8 x half> undef,
<vscale x 8 x i32> %0,
iXLen %1)

Expand Down Expand Up @@ -172,6 +180,7 @@ entry:
}

declare <vscale x 16 x half> @llvm.riscv.vfncvt.f.xu.w.nxv16f16.nxv16i32(
<vscale x 16 x half>,
<vscale x 16 x i32>,
iXLen);

Expand All @@ -184,6 +193,7 @@ define <vscale x 16 x half> @intrinsic_vfncvt_f.xu.w_nxv16f16_nxv16i32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x half> @llvm.riscv.vfncvt.f.xu.w.nxv16f16.nxv16i32(
<vscale x 16 x half> undef,
<vscale x 16 x i32> %0,
iXLen %1)

Expand Down Expand Up @@ -214,6 +224,7 @@ entry:
}

declare <vscale x 1 x float> @llvm.riscv.vfncvt.f.xu.w.nxv1f32.nxv1i64(
<vscale x 1 x float>,
<vscale x 1 x i64>,
iXLen);

Expand All @@ -226,6 +237,7 @@ define <vscale x 1 x float> @intrinsic_vfncvt_f.xu.w_nxv1f32_nxv1i64(<vscale x 1
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x float> @llvm.riscv.vfncvt.f.xu.w.nxv1f32.nxv1i64(
<vscale x 1 x float> undef,
<vscale x 1 x i64> %0,
iXLen %1)

Expand Down Expand Up @@ -256,6 +268,7 @@ entry:
}

declare <vscale x 2 x float> @llvm.riscv.vfncvt.f.xu.w.nxv2f32.nxv2i64(
<vscale x 2 x float>,
<vscale x 2 x i64>,
iXLen);

Expand All @@ -268,6 +281,7 @@ define <vscale x 2 x float> @intrinsic_vfncvt_f.xu.w_nxv2f32_nxv2i64(<vscale x 2
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x float> @llvm.riscv.vfncvt.f.xu.w.nxv2f32.nxv2i64(
<vscale x 2 x float> undef,
<vscale x 2 x i64> %0,
iXLen %1)

Expand Down Expand Up @@ -298,6 +312,7 @@ entry:
}

declare <vscale x 4 x float> @llvm.riscv.vfncvt.f.xu.w.nxv4f32.nxv4i64(
<vscale x 4 x float>,
<vscale x 4 x i64>,
iXLen);

Expand All @@ -310,6 +325,7 @@ define <vscale x 4 x float> @intrinsic_vfncvt_f.xu.w_nxv4f32_nxv4i64(<vscale x 4
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x float> @llvm.riscv.vfncvt.f.xu.w.nxv4f32.nxv4i64(
<vscale x 4 x float> undef,
<vscale x 4 x i64> %0,
iXLen %1)

Expand Down Expand Up @@ -340,6 +356,7 @@ entry:
}

declare <vscale x 8 x float> @llvm.riscv.vfncvt.f.xu.w.nxv8f32.nxv8i64(
<vscale x 8 x float>,
<vscale x 8 x i64>,
iXLen);

Expand All @@ -352,6 +369,7 @@ define <vscale x 8 x float> @intrinsic_vfncvt_f.xu.w_nxv8f32_nxv8i64(<vscale x 8
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x float> @llvm.riscv.vfncvt.f.xu.w.nxv8f32.nxv8i64(
<vscale x 8 x float> undef,
<vscale x 8 x i64> %0,
iXLen %1)

Expand Down
18 changes: 18 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfncvt-rod-f-f.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x half> @llvm.riscv.vfncvt.rod.f.f.w.nxv1f16.nxv1f32(
<vscale x 1 x half>,
<vscale x 1 x float>,
iXLen);

Expand All @@ -16,6 +17,7 @@ define <vscale x 1 x half> @intrinsic_vfncvt_rod.f.f.w_nxv1f16_nxv1f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x half> @llvm.riscv.vfncvt.rod.f.f.w.nxv1f16.nxv1f32(
<vscale x 1 x half> undef,
<vscale x 1 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -46,6 +48,7 @@ entry:
}

declare <vscale x 2 x half> @llvm.riscv.vfncvt.rod.f.f.w.nxv2f16.nxv2f32(
<vscale x 2 x half>,
<vscale x 2 x float>,
iXLen);

Expand All @@ -58,6 +61,7 @@ define <vscale x 2 x half> @intrinsic_vfncvt_rod.f.f.w_nxv2f16_nxv2f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x half> @llvm.riscv.vfncvt.rod.f.f.w.nxv2f16.nxv2f32(
<vscale x 2 x half> undef,
<vscale x 2 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -88,6 +92,7 @@ entry:
}

declare <vscale x 4 x half> @llvm.riscv.vfncvt.rod.f.f.w.nxv4f16.nxv4f32(
<vscale x 4 x half>,
<vscale x 4 x float>,
iXLen);

Expand All @@ -100,6 +105,7 @@ define <vscale x 4 x half> @intrinsic_vfncvt_rod.f.f.w_nxv4f16_nxv4f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x half> @llvm.riscv.vfncvt.rod.f.f.w.nxv4f16.nxv4f32(
<vscale x 4 x half> undef,
<vscale x 4 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -130,6 +136,7 @@ entry:
}

declare <vscale x 8 x half> @llvm.riscv.vfncvt.rod.f.f.w.nxv8f16.nxv8f32(
<vscale x 8 x half>,
<vscale x 8 x float>,
iXLen);

Expand All @@ -142,6 +149,7 @@ define <vscale x 8 x half> @intrinsic_vfncvt_rod.f.f.w_nxv8f16_nxv8f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x half> @llvm.riscv.vfncvt.rod.f.f.w.nxv8f16.nxv8f32(
<vscale x 8 x half> undef,
<vscale x 8 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -172,6 +180,7 @@ entry:
}

declare <vscale x 16 x half> @llvm.riscv.vfncvt.rod.f.f.w.nxv16f16.nxv16f32(
<vscale x 16 x half>,
<vscale x 16 x float>,
iXLen);

Expand All @@ -184,6 +193,7 @@ define <vscale x 16 x half> @intrinsic_vfncvt_rod.f.f.w_nxv16f16_nxv16f32(<vscal
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x half> @llvm.riscv.vfncvt.rod.f.f.w.nxv16f16.nxv16f32(
<vscale x 16 x half> undef,
<vscale x 16 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -214,6 +224,7 @@ entry:
}

declare <vscale x 1 x float> @llvm.riscv.vfncvt.rod.f.f.w.nxv1f32.nxv1f64(
<vscale x 1 x float>,
<vscale x 1 x double>,
iXLen);

Expand All @@ -226,6 +237,7 @@ define <vscale x 1 x float> @intrinsic_vfncvt_rod.f.f.w_nxv1f32_nxv1f64(<vscale
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x float> @llvm.riscv.vfncvt.rod.f.f.w.nxv1f32.nxv1f64(
<vscale x 1 x float> undef,
<vscale x 1 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -256,6 +268,7 @@ entry:
}

declare <vscale x 2 x float> @llvm.riscv.vfncvt.rod.f.f.w.nxv2f32.nxv2f64(
<vscale x 2 x float>,
<vscale x 2 x double>,
iXLen);

Expand All @@ -268,6 +281,7 @@ define <vscale x 2 x float> @intrinsic_vfncvt_rod.f.f.w_nxv2f32_nxv2f64(<vscale
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x float> @llvm.riscv.vfncvt.rod.f.f.w.nxv2f32.nxv2f64(
<vscale x 2 x float> undef,
<vscale x 2 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -298,6 +312,7 @@ entry:
}

declare <vscale x 4 x float> @llvm.riscv.vfncvt.rod.f.f.w.nxv4f32.nxv4f64(
<vscale x 4 x float>,
<vscale x 4 x double>,
iXLen);

Expand All @@ -310,6 +325,7 @@ define <vscale x 4 x float> @intrinsic_vfncvt_rod.f.f.w_nxv4f32_nxv4f64(<vscale
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x float> @llvm.riscv.vfncvt.rod.f.f.w.nxv4f32.nxv4f64(
<vscale x 4 x float> undef,
<vscale x 4 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -340,6 +356,7 @@ entry:
}

declare <vscale x 8 x float> @llvm.riscv.vfncvt.rod.f.f.w.nxv8f32.nxv8f64(
<vscale x 8 x float>,
<vscale x 8 x double>,
iXLen);

Expand All @@ -352,6 +369,7 @@ define <vscale x 8 x float> @intrinsic_vfncvt_rod.f.f.w_nxv8f32_nxv8f64(<vscale
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x float> @llvm.riscv.vfncvt.rod.f.f.w.nxv8f32.nxv8f64(
<vscale x 8 x float> undef,
<vscale x 8 x double> %0,
iXLen %1)

Expand Down
30 changes: 30 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfncvt-rtz-x-f.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x i8> @llvm.riscv.vfncvt.rtz.x.f.w.nxv1i8.nxv1f16(
<vscale x 1 x i8>,
<vscale x 1 x half>,
iXLen);

Expand All @@ -16,6 +17,7 @@ define <vscale x 1 x i8> @intrinsic_vfncvt_rtz.x.f.w_nxv1i8_nxv1f16(<vscale x 1
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i8> @llvm.riscv.vfncvt.rtz.x.f.w.nxv1i8.nxv1f16(
<vscale x 1 x i8> undef,
<vscale x 1 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -46,6 +48,7 @@ entry:
}

declare <vscale x 2 x i8> @llvm.riscv.vfncvt.rtz.x.f.w.nxv2i8.nxv2f16(
<vscale x 2 x i8>,
<vscale x 2 x half>,
iXLen);

Expand All @@ -58,6 +61,7 @@ define <vscale x 2 x i8> @intrinsic_vfncvt_rtz.x.f.w_nxv2i8_nxv2f16(<vscale x 2
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i8> @llvm.riscv.vfncvt.rtz.x.f.w.nxv2i8.nxv2f16(
<vscale x 2 x i8> undef,
<vscale x 2 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -88,6 +92,7 @@ entry:
}

declare <vscale x 4 x i8> @llvm.riscv.vfncvt.rtz.x.f.w.nxv4i8.nxv4f16(
<vscale x 4 x i8>,
<vscale x 4 x half>,
iXLen);

Expand All @@ -100,6 +105,7 @@ define <vscale x 4 x i8> @intrinsic_vfncvt_rtz.x.f.w_nxv4i8_nxv4f16(<vscale x 4
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i8> @llvm.riscv.vfncvt.rtz.x.f.w.nxv4i8.nxv4f16(
<vscale x 4 x i8> undef,
<vscale x 4 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -130,6 +136,7 @@ entry:
}

declare <vscale x 8 x i8> @llvm.riscv.vfncvt.rtz.x.f.w.nxv8i8.nxv8f16(
<vscale x 8 x i8>,
<vscale x 8 x half>,
iXLen);

Expand All @@ -142,6 +149,7 @@ define <vscale x 8 x i8> @intrinsic_vfncvt_rtz.x.f.w_nxv8i8_nxv8f16(<vscale x 8
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i8> @llvm.riscv.vfncvt.rtz.x.f.w.nxv8i8.nxv8f16(
<vscale x 8 x i8> undef,
<vscale x 8 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -172,6 +180,7 @@ entry:
}

declare <vscale x 16 x i8> @llvm.riscv.vfncvt.rtz.x.f.w.nxv16i8.nxv16f16(
<vscale x 16 x i8>,
<vscale x 16 x half>,
iXLen);

Expand All @@ -184,6 +193,7 @@ define <vscale x 16 x i8> @intrinsic_vfncvt_rtz.x.f.w_nxv16i8_nxv16f16(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i8> @llvm.riscv.vfncvt.rtz.x.f.w.nxv16i8.nxv16f16(
<vscale x 16 x i8> undef,
<vscale x 16 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -214,6 +224,7 @@ entry:
}

declare <vscale x 32 x i8> @llvm.riscv.vfncvt.rtz.x.f.w.nxv32i8.nxv32f16(
<vscale x 32 x i8>,
<vscale x 32 x half>,
iXLen);

Expand All @@ -226,6 +237,7 @@ define <vscale x 32 x i8> @intrinsic_vfncvt_rtz.x.f.w_nxv32i8_nxv32f16(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 32 x i8> @llvm.riscv.vfncvt.rtz.x.f.w.nxv32i8.nxv32f16(
<vscale x 32 x i8> undef,
<vscale x 32 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -256,6 +268,7 @@ entry:
}

declare <vscale x 1 x i16> @llvm.riscv.vfncvt.rtz.x.f.w.nxv1i16.nxv1f32(
<vscale x 1 x i16>,
<vscale x 1 x float>,
iXLen);

Expand All @@ -268,6 +281,7 @@ define <vscale x 1 x i16> @intrinsic_vfncvt_rtz.x.f.w_nxv1i16_nxv1f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i16> @llvm.riscv.vfncvt.rtz.x.f.w.nxv1i16.nxv1f32(
<vscale x 1 x i16> undef,
<vscale x 1 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -298,6 +312,7 @@ entry:
}

declare <vscale x 2 x i16> @llvm.riscv.vfncvt.rtz.x.f.w.nxv2i16.nxv2f32(
<vscale x 2 x i16>,
<vscale x 2 x float>,
iXLen);

Expand All @@ -310,6 +325,7 @@ define <vscale x 2 x i16> @intrinsic_vfncvt_rtz.x.f.w_nxv2i16_nxv2f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i16> @llvm.riscv.vfncvt.rtz.x.f.w.nxv2i16.nxv2f32(
<vscale x 2 x i16> undef,
<vscale x 2 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -340,6 +356,7 @@ entry:
}

declare <vscale x 4 x i16> @llvm.riscv.vfncvt.rtz.x.f.w.nxv4i16.nxv4f32(
<vscale x 4 x i16>,
<vscale x 4 x float>,
iXLen);

Expand All @@ -352,6 +369,7 @@ define <vscale x 4 x i16> @intrinsic_vfncvt_rtz.x.f.w_nxv4i16_nxv4f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i16> @llvm.riscv.vfncvt.rtz.x.f.w.nxv4i16.nxv4f32(
<vscale x 4 x i16> undef,
<vscale x 4 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -382,6 +400,7 @@ entry:
}

declare <vscale x 8 x i16> @llvm.riscv.vfncvt.rtz.x.f.w.nxv8i16.nxv8f32(
<vscale x 8 x i16>,
<vscale x 8 x float>,
iXLen);

Expand All @@ -394,6 +413,7 @@ define <vscale x 8 x i16> @intrinsic_vfncvt_rtz.x.f.w_nxv8i16_nxv8f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i16> @llvm.riscv.vfncvt.rtz.x.f.w.nxv8i16.nxv8f32(
<vscale x 8 x i16> undef,
<vscale x 8 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -424,6 +444,7 @@ entry:
}

declare <vscale x 16 x i16> @llvm.riscv.vfncvt.rtz.x.f.w.nxv16i16.nxv16f32(
<vscale x 16 x i16>,
<vscale x 16 x float>,
iXLen);

Expand All @@ -436,6 +457,7 @@ define <vscale x 16 x i16> @intrinsic_vfncvt_rtz.x.f.w_nxv16i16_nxv16f32(<vscale
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i16> @llvm.riscv.vfncvt.rtz.x.f.w.nxv16i16.nxv16f32(
<vscale x 16 x i16> undef,
<vscale x 16 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -466,6 +488,7 @@ entry:
}

declare <vscale x 1 x i32> @llvm.riscv.vfncvt.rtz.x.f.w.nxv1i32.nxv1f64(
<vscale x 1 x i32>,
<vscale x 1 x double>,
iXLen);

Expand All @@ -478,6 +501,7 @@ define <vscale x 1 x i32> @intrinsic_vfncvt_rtz.x.f.w_nxv1i32_nxv1f64(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i32> @llvm.riscv.vfncvt.rtz.x.f.w.nxv1i32.nxv1f64(
<vscale x 1 x i32> undef,
<vscale x 1 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -508,6 +532,7 @@ entry:
}

declare <vscale x 2 x i32> @llvm.riscv.vfncvt.rtz.x.f.w.nxv2i32.nxv2f64(
<vscale x 2 x i32>,
<vscale x 2 x double>,
iXLen);

Expand All @@ -520,6 +545,7 @@ define <vscale x 2 x i32> @intrinsic_vfncvt_rtz.x.f.w_nxv2i32_nxv2f64(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i32> @llvm.riscv.vfncvt.rtz.x.f.w.nxv2i32.nxv2f64(
<vscale x 2 x i32> undef,
<vscale x 2 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -550,6 +576,7 @@ entry:
}

declare <vscale x 4 x i32> @llvm.riscv.vfncvt.rtz.x.f.w.nxv4i32.nxv4f64(
<vscale x 4 x i32>,
<vscale x 4 x double>,
iXLen);

Expand All @@ -562,6 +589,7 @@ define <vscale x 4 x i32> @intrinsic_vfncvt_rtz.x.f.w_nxv4i32_nxv4f64(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i32> @llvm.riscv.vfncvt.rtz.x.f.w.nxv4i32.nxv4f64(
<vscale x 4 x i32> undef,
<vscale x 4 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -592,6 +620,7 @@ entry:
}

declare <vscale x 8 x i32> @llvm.riscv.vfncvt.rtz.x.f.w.nxv8i32.nxv8f64(
<vscale x 8 x i32>,
<vscale x 8 x double>,
iXLen);

Expand All @@ -604,6 +633,7 @@ define <vscale x 8 x i32> @intrinsic_vfncvt_rtz.x.f.w_nxv8i32_nxv8f64(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i32> @llvm.riscv.vfncvt.rtz.x.f.w.nxv8i32.nxv8f64(
<vscale x 8 x i32> undef,
<vscale x 8 x double> %0,
iXLen %1)

Expand Down
30 changes: 30 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfncvt-rtz-xu-f.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x i8> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv1i8.nxv1f16(
<vscale x 1 x i8>,
<vscale x 1 x half>,
iXLen);

Expand All @@ -16,6 +17,7 @@ define <vscale x 1 x i8> @intrinsic_vfncvt_rtz.xu.f.w_nxv1i8_nxv1f16(<vscale x 1
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i8> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv1i8.nxv1f16(
<vscale x 1 x i8> undef,
<vscale x 1 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -46,6 +48,7 @@ entry:
}

declare <vscale x 2 x i8> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv2i8.nxv2f16(
<vscale x 2 x i8>,
<vscale x 2 x half>,
iXLen);

Expand All @@ -58,6 +61,7 @@ define <vscale x 2 x i8> @intrinsic_vfncvt_rtz.xu.f.w_nxv2i8_nxv2f16(<vscale x 2
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i8> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv2i8.nxv2f16(
<vscale x 2 x i8> undef,
<vscale x 2 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -88,6 +92,7 @@ entry:
}

declare <vscale x 4 x i8> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv4i8.nxv4f16(
<vscale x 4 x i8>,
<vscale x 4 x half>,
iXLen);

Expand All @@ -100,6 +105,7 @@ define <vscale x 4 x i8> @intrinsic_vfncvt_rtz.xu.f.w_nxv4i8_nxv4f16(<vscale x 4
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i8> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv4i8.nxv4f16(
<vscale x 4 x i8> undef,
<vscale x 4 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -130,6 +136,7 @@ entry:
}

declare <vscale x 8 x i8> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv8i8.nxv8f16(
<vscale x 8 x i8>,
<vscale x 8 x half>,
iXLen);

Expand All @@ -142,6 +149,7 @@ define <vscale x 8 x i8> @intrinsic_vfncvt_rtz.xu.f.w_nxv8i8_nxv8f16(<vscale x 8
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i8> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv8i8.nxv8f16(
<vscale x 8 x i8> undef,
<vscale x 8 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -172,6 +180,7 @@ entry:
}

declare <vscale x 16 x i8> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv16i8.nxv16f16(
<vscale x 16 x i8>,
<vscale x 16 x half>,
iXLen);

Expand All @@ -184,6 +193,7 @@ define <vscale x 16 x i8> @intrinsic_vfncvt_rtz.xu.f.w_nxv16i8_nxv16f16(<vscale
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i8> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv16i8.nxv16f16(
<vscale x 16 x i8> undef,
<vscale x 16 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -214,6 +224,7 @@ entry:
}

declare <vscale x 32 x i8> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv32i8.nxv32f16(
<vscale x 32 x i8>,
<vscale x 32 x half>,
iXLen);

Expand All @@ -226,6 +237,7 @@ define <vscale x 32 x i8> @intrinsic_vfncvt_rtz.xu.f.w_nxv32i8_nxv32f16(<vscale
; CHECK-NEXT: ret
entry:
%a = call <vscale x 32 x i8> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv32i8.nxv32f16(
<vscale x 32 x i8> undef,
<vscale x 32 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -256,6 +268,7 @@ entry:
}

declare <vscale x 1 x i16> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv1i16.nxv1f32(
<vscale x 1 x i16>,
<vscale x 1 x float>,
iXLen);

Expand All @@ -268,6 +281,7 @@ define <vscale x 1 x i16> @intrinsic_vfncvt_rtz.xu.f.w_nxv1i16_nxv1f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i16> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv1i16.nxv1f32(
<vscale x 1 x i16> undef,
<vscale x 1 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -298,6 +312,7 @@ entry:
}

declare <vscale x 2 x i16> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv2i16.nxv2f32(
<vscale x 2 x i16>,
<vscale x 2 x float>,
iXLen);

Expand All @@ -310,6 +325,7 @@ define <vscale x 2 x i16> @intrinsic_vfncvt_rtz.xu.f.w_nxv2i16_nxv2f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i16> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv2i16.nxv2f32(
<vscale x 2 x i16> undef,
<vscale x 2 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -340,6 +356,7 @@ entry:
}

declare <vscale x 4 x i16> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv4i16.nxv4f32(
<vscale x 4 x i16>,
<vscale x 4 x float>,
iXLen);

Expand All @@ -352,6 +369,7 @@ define <vscale x 4 x i16> @intrinsic_vfncvt_rtz.xu.f.w_nxv4i16_nxv4f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i16> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv4i16.nxv4f32(
<vscale x 4 x i16> undef,
<vscale x 4 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -382,6 +400,7 @@ entry:
}

declare <vscale x 8 x i16> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv8i16.nxv8f32(
<vscale x 8 x i16>,
<vscale x 8 x float>,
iXLen);

Expand All @@ -394,6 +413,7 @@ define <vscale x 8 x i16> @intrinsic_vfncvt_rtz.xu.f.w_nxv8i16_nxv8f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i16> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv8i16.nxv8f32(
<vscale x 8 x i16> undef,
<vscale x 8 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -424,6 +444,7 @@ entry:
}

declare <vscale x 16 x i16> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv16i16.nxv16f32(
<vscale x 16 x i16>,
<vscale x 16 x float>,
iXLen);

Expand All @@ -436,6 +457,7 @@ define <vscale x 16 x i16> @intrinsic_vfncvt_rtz.xu.f.w_nxv16i16_nxv16f32(<vscal
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i16> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv16i16.nxv16f32(
<vscale x 16 x i16> undef,
<vscale x 16 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -466,6 +488,7 @@ entry:
}

declare <vscale x 1 x i32> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv1i32.nxv1f64(
<vscale x 1 x i32>,
<vscale x 1 x double>,
iXLen);

Expand All @@ -478,6 +501,7 @@ define <vscale x 1 x i32> @intrinsic_vfncvt_rtz.xu.f.w_nxv1i32_nxv1f64(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i32> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv1i32.nxv1f64(
<vscale x 1 x i32> undef,
<vscale x 1 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -508,6 +532,7 @@ entry:
}

declare <vscale x 2 x i32> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv2i32.nxv2f64(
<vscale x 2 x i32>,
<vscale x 2 x double>,
iXLen);

Expand All @@ -520,6 +545,7 @@ define <vscale x 2 x i32> @intrinsic_vfncvt_rtz.xu.f.w_nxv2i32_nxv2f64(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i32> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv2i32.nxv2f64(
<vscale x 2 x i32> undef,
<vscale x 2 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -550,6 +576,7 @@ entry:
}

declare <vscale x 4 x i32> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv4i32.nxv4f64(
<vscale x 4 x i32>,
<vscale x 4 x double>,
iXLen);

Expand All @@ -562,6 +589,7 @@ define <vscale x 4 x i32> @intrinsic_vfncvt_rtz.xu.f.w_nxv4i32_nxv4f64(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i32> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv4i32.nxv4f64(
<vscale x 4 x i32> undef,
<vscale x 4 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -592,6 +620,7 @@ entry:
}

declare <vscale x 8 x i32> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv8i32.nxv8f64(
<vscale x 8 x i32>,
<vscale x 8 x double>,
iXLen);

Expand All @@ -604,6 +633,7 @@ define <vscale x 8 x i32> @intrinsic_vfncvt_rtz.xu.f.w_nxv8i32_nxv8f64(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i32> @llvm.riscv.vfncvt.rtz.xu.f.w.nxv8i32.nxv8f64(
<vscale x 8 x i32> undef,
<vscale x 8 x double> %0,
iXLen %1)

Expand Down
30 changes: 30 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfncvt-x-f.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x i8> @llvm.riscv.vfncvt.x.f.w.nxv1i8.nxv1f16(
<vscale x 1 x i8>,
<vscale x 1 x half>,
iXLen);

Expand All @@ -16,6 +17,7 @@ define <vscale x 1 x i8> @intrinsic_vfncvt_x.f.w_nxv1i8_nxv1f16(<vscale x 1 x ha
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i8> @llvm.riscv.vfncvt.x.f.w.nxv1i8.nxv1f16(
<vscale x 1 x i8> undef,
<vscale x 1 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -46,6 +48,7 @@ entry:
}

declare <vscale x 2 x i8> @llvm.riscv.vfncvt.x.f.w.nxv2i8.nxv2f16(
<vscale x 2 x i8>,
<vscale x 2 x half>,
iXLen);

Expand All @@ -58,6 +61,7 @@ define <vscale x 2 x i8> @intrinsic_vfncvt_x.f.w_nxv2i8_nxv2f16(<vscale x 2 x ha
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i8> @llvm.riscv.vfncvt.x.f.w.nxv2i8.nxv2f16(
<vscale x 2 x i8> undef,
<vscale x 2 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -88,6 +92,7 @@ entry:
}

declare <vscale x 4 x i8> @llvm.riscv.vfncvt.x.f.w.nxv4i8.nxv4f16(
<vscale x 4 x i8>,
<vscale x 4 x half>,
iXLen);

Expand All @@ -100,6 +105,7 @@ define <vscale x 4 x i8> @intrinsic_vfncvt_x.f.w_nxv4i8_nxv4f16(<vscale x 4 x ha
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i8> @llvm.riscv.vfncvt.x.f.w.nxv4i8.nxv4f16(
<vscale x 4 x i8> undef,
<vscale x 4 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -130,6 +136,7 @@ entry:
}

declare <vscale x 8 x i8> @llvm.riscv.vfncvt.x.f.w.nxv8i8.nxv8f16(
<vscale x 8 x i8>,
<vscale x 8 x half>,
iXLen);

Expand All @@ -142,6 +149,7 @@ define <vscale x 8 x i8> @intrinsic_vfncvt_x.f.w_nxv8i8_nxv8f16(<vscale x 8 x ha
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i8> @llvm.riscv.vfncvt.x.f.w.nxv8i8.nxv8f16(
<vscale x 8 x i8> undef,
<vscale x 8 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -172,6 +180,7 @@ entry:
}

declare <vscale x 16 x i8> @llvm.riscv.vfncvt.x.f.w.nxv16i8.nxv16f16(
<vscale x 16 x i8>,
<vscale x 16 x half>,
iXLen);

Expand All @@ -184,6 +193,7 @@ define <vscale x 16 x i8> @intrinsic_vfncvt_x.f.w_nxv16i8_nxv16f16(<vscale x 16
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i8> @llvm.riscv.vfncvt.x.f.w.nxv16i8.nxv16f16(
<vscale x 16 x i8> undef,
<vscale x 16 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -214,6 +224,7 @@ entry:
}

declare <vscale x 32 x i8> @llvm.riscv.vfncvt.x.f.w.nxv32i8.nxv32f16(
<vscale x 32 x i8>,
<vscale x 32 x half>,
iXLen);

Expand All @@ -226,6 +237,7 @@ define <vscale x 32 x i8> @intrinsic_vfncvt_x.f.w_nxv32i8_nxv32f16(<vscale x 32
; CHECK-NEXT: ret
entry:
%a = call <vscale x 32 x i8> @llvm.riscv.vfncvt.x.f.w.nxv32i8.nxv32f16(
<vscale x 32 x i8> undef,
<vscale x 32 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -256,6 +268,7 @@ entry:
}

declare <vscale x 1 x i16> @llvm.riscv.vfncvt.x.f.w.nxv1i16.nxv1f32(
<vscale x 1 x i16>,
<vscale x 1 x float>,
iXLen);

Expand All @@ -268,6 +281,7 @@ define <vscale x 1 x i16> @intrinsic_vfncvt_x.f.w_nxv1i16_nxv1f32(<vscale x 1 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i16> @llvm.riscv.vfncvt.x.f.w.nxv1i16.nxv1f32(
<vscale x 1 x i16> undef,
<vscale x 1 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -298,6 +312,7 @@ entry:
}

declare <vscale x 2 x i16> @llvm.riscv.vfncvt.x.f.w.nxv2i16.nxv2f32(
<vscale x 2 x i16>,
<vscale x 2 x float>,
iXLen);

Expand All @@ -310,6 +325,7 @@ define <vscale x 2 x i16> @intrinsic_vfncvt_x.f.w_nxv2i16_nxv2f32(<vscale x 2 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i16> @llvm.riscv.vfncvt.x.f.w.nxv2i16.nxv2f32(
<vscale x 2 x i16> undef,
<vscale x 2 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -340,6 +356,7 @@ entry:
}

declare <vscale x 4 x i16> @llvm.riscv.vfncvt.x.f.w.nxv4i16.nxv4f32(
<vscale x 4 x i16>,
<vscale x 4 x float>,
iXLen);

Expand All @@ -352,6 +369,7 @@ define <vscale x 4 x i16> @intrinsic_vfncvt_x.f.w_nxv4i16_nxv4f32(<vscale x 4 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i16> @llvm.riscv.vfncvt.x.f.w.nxv4i16.nxv4f32(
<vscale x 4 x i16> undef,
<vscale x 4 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -382,6 +400,7 @@ entry:
}

declare <vscale x 8 x i16> @llvm.riscv.vfncvt.x.f.w.nxv8i16.nxv8f32(
<vscale x 8 x i16>,
<vscale x 8 x float>,
iXLen);

Expand All @@ -394,6 +413,7 @@ define <vscale x 8 x i16> @intrinsic_vfncvt_x.f.w_nxv8i16_nxv8f32(<vscale x 8 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i16> @llvm.riscv.vfncvt.x.f.w.nxv8i16.nxv8f32(
<vscale x 8 x i16> undef,
<vscale x 8 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -424,6 +444,7 @@ entry:
}

declare <vscale x 16 x i16> @llvm.riscv.vfncvt.x.f.w.nxv16i16.nxv16f32(
<vscale x 16 x i16>,
<vscale x 16 x float>,
iXLen);

Expand All @@ -436,6 +457,7 @@ define <vscale x 16 x i16> @intrinsic_vfncvt_x.f.w_nxv16i16_nxv16f32(<vscale x 1
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i16> @llvm.riscv.vfncvt.x.f.w.nxv16i16.nxv16f32(
<vscale x 16 x i16> undef,
<vscale x 16 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -466,6 +488,7 @@ entry:
}

declare <vscale x 1 x i32> @llvm.riscv.vfncvt.x.f.w.nxv1i32.nxv1f64(
<vscale x 1 x i32>,
<vscale x 1 x double>,
iXLen);

Expand All @@ -478,6 +501,7 @@ define <vscale x 1 x i32> @intrinsic_vfncvt_x.f.w_nxv1i32_nxv1f64(<vscale x 1 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i32> @llvm.riscv.vfncvt.x.f.w.nxv1i32.nxv1f64(
<vscale x 1 x i32> undef,
<vscale x 1 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -508,6 +532,7 @@ entry:
}

declare <vscale x 2 x i32> @llvm.riscv.vfncvt.x.f.w.nxv2i32.nxv2f64(
<vscale x 2 x i32>,
<vscale x 2 x double>,
iXLen);

Expand All @@ -520,6 +545,7 @@ define <vscale x 2 x i32> @intrinsic_vfncvt_x.f.w_nxv2i32_nxv2f64(<vscale x 2 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i32> @llvm.riscv.vfncvt.x.f.w.nxv2i32.nxv2f64(
<vscale x 2 x i32> undef,
<vscale x 2 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -550,6 +576,7 @@ entry:
}

declare <vscale x 4 x i32> @llvm.riscv.vfncvt.x.f.w.nxv4i32.nxv4f64(
<vscale x 4 x i32>,
<vscale x 4 x double>,
iXLen);

Expand All @@ -562,6 +589,7 @@ define <vscale x 4 x i32> @intrinsic_vfncvt_x.f.w_nxv4i32_nxv4f64(<vscale x 4 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i32> @llvm.riscv.vfncvt.x.f.w.nxv4i32.nxv4f64(
<vscale x 4 x i32> undef,
<vscale x 4 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -592,6 +620,7 @@ entry:
}

declare <vscale x 8 x i32> @llvm.riscv.vfncvt.x.f.w.nxv8i32.nxv8f64(
<vscale x 8 x i32>,
<vscale x 8 x double>,
iXLen);

Expand All @@ -604,6 +633,7 @@ define <vscale x 8 x i32> @intrinsic_vfncvt_x.f.w_nxv8i32_nxv8f64(<vscale x 8 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i32> @llvm.riscv.vfncvt.x.f.w.nxv8i32.nxv8f64(
<vscale x 8 x i32> undef,
<vscale x 8 x double> %0,
iXLen %1)

Expand Down
30 changes: 30 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfncvt-xu-f.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x i8> @llvm.riscv.vfncvt.xu.f.w.nxv1i8.nxv1f16(
<vscale x 1 x i8>,
<vscale x 1 x half>,
iXLen);

Expand All @@ -16,6 +17,7 @@ define <vscale x 1 x i8> @intrinsic_vfncvt_xu.f.w_nxv1i8_nxv1f16(<vscale x 1 x h
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i8> @llvm.riscv.vfncvt.xu.f.w.nxv1i8.nxv1f16(
<vscale x 1 x i8> undef,
<vscale x 1 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -46,6 +48,7 @@ entry:
}

declare <vscale x 2 x i8> @llvm.riscv.vfncvt.xu.f.w.nxv2i8.nxv2f16(
<vscale x 2 x i8>,
<vscale x 2 x half>,
iXLen);

Expand All @@ -58,6 +61,7 @@ define <vscale x 2 x i8> @intrinsic_vfncvt_xu.f.w_nxv2i8_nxv2f16(<vscale x 2 x h
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i8> @llvm.riscv.vfncvt.xu.f.w.nxv2i8.nxv2f16(
<vscale x 2 x i8> undef,
<vscale x 2 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -88,6 +92,7 @@ entry:
}

declare <vscale x 4 x i8> @llvm.riscv.vfncvt.xu.f.w.nxv4i8.nxv4f16(
<vscale x 4 x i8>,
<vscale x 4 x half>,
iXLen);

Expand All @@ -100,6 +105,7 @@ define <vscale x 4 x i8> @intrinsic_vfncvt_xu.f.w_nxv4i8_nxv4f16(<vscale x 4 x h
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i8> @llvm.riscv.vfncvt.xu.f.w.nxv4i8.nxv4f16(
<vscale x 4 x i8> undef,
<vscale x 4 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -130,6 +136,7 @@ entry:
}

declare <vscale x 8 x i8> @llvm.riscv.vfncvt.xu.f.w.nxv8i8.nxv8f16(
<vscale x 8 x i8>,
<vscale x 8 x half>,
iXLen);

Expand All @@ -142,6 +149,7 @@ define <vscale x 8 x i8> @intrinsic_vfncvt_xu.f.w_nxv8i8_nxv8f16(<vscale x 8 x h
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i8> @llvm.riscv.vfncvt.xu.f.w.nxv8i8.nxv8f16(
<vscale x 8 x i8> undef,
<vscale x 8 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -172,6 +180,7 @@ entry:
}

declare <vscale x 16 x i8> @llvm.riscv.vfncvt.xu.f.w.nxv16i8.nxv16f16(
<vscale x 16 x i8>,
<vscale x 16 x half>,
iXLen);

Expand All @@ -184,6 +193,7 @@ define <vscale x 16 x i8> @intrinsic_vfncvt_xu.f.w_nxv16i8_nxv16f16(<vscale x 16
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i8> @llvm.riscv.vfncvt.xu.f.w.nxv16i8.nxv16f16(
<vscale x 16 x i8> undef,
<vscale x 16 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -214,6 +224,7 @@ entry:
}

declare <vscale x 32 x i8> @llvm.riscv.vfncvt.xu.f.w.nxv32i8.nxv32f16(
<vscale x 32 x i8>,
<vscale x 32 x half>,
iXLen);

Expand All @@ -226,6 +237,7 @@ define <vscale x 32 x i8> @intrinsic_vfncvt_xu.f.w_nxv32i8_nxv32f16(<vscale x 32
; CHECK-NEXT: ret
entry:
%a = call <vscale x 32 x i8> @llvm.riscv.vfncvt.xu.f.w.nxv32i8.nxv32f16(
<vscale x 32 x i8> undef,
<vscale x 32 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -256,6 +268,7 @@ entry:
}

declare <vscale x 1 x i16> @llvm.riscv.vfncvt.xu.f.w.nxv1i16.nxv1f32(
<vscale x 1 x i16>,
<vscale x 1 x float>,
iXLen);

Expand All @@ -268,6 +281,7 @@ define <vscale x 1 x i16> @intrinsic_vfncvt_xu.f.w_nxv1i16_nxv1f32(<vscale x 1 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i16> @llvm.riscv.vfncvt.xu.f.w.nxv1i16.nxv1f32(
<vscale x 1 x i16> undef,
<vscale x 1 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -298,6 +312,7 @@ entry:
}

declare <vscale x 2 x i16> @llvm.riscv.vfncvt.xu.f.w.nxv2i16.nxv2f32(
<vscale x 2 x i16>,
<vscale x 2 x float>,
iXLen);

Expand All @@ -310,6 +325,7 @@ define <vscale x 2 x i16> @intrinsic_vfncvt_xu.f.w_nxv2i16_nxv2f32(<vscale x 2 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i16> @llvm.riscv.vfncvt.xu.f.w.nxv2i16.nxv2f32(
<vscale x 2 x i16> undef,
<vscale x 2 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -340,6 +356,7 @@ entry:
}

declare <vscale x 4 x i16> @llvm.riscv.vfncvt.xu.f.w.nxv4i16.nxv4f32(
<vscale x 4 x i16>,
<vscale x 4 x float>,
iXLen);

Expand All @@ -352,6 +369,7 @@ define <vscale x 4 x i16> @intrinsic_vfncvt_xu.f.w_nxv4i16_nxv4f32(<vscale x 4 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i16> @llvm.riscv.vfncvt.xu.f.w.nxv4i16.nxv4f32(
<vscale x 4 x i16> undef,
<vscale x 4 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -382,6 +400,7 @@ entry:
}

declare <vscale x 8 x i16> @llvm.riscv.vfncvt.xu.f.w.nxv8i16.nxv8f32(
<vscale x 8 x i16>,
<vscale x 8 x float>,
iXLen);

Expand All @@ -394,6 +413,7 @@ define <vscale x 8 x i16> @intrinsic_vfncvt_xu.f.w_nxv8i16_nxv8f32(<vscale x 8 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i16> @llvm.riscv.vfncvt.xu.f.w.nxv8i16.nxv8f32(
<vscale x 8 x i16> undef,
<vscale x 8 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -424,6 +444,7 @@ entry:
}

declare <vscale x 16 x i16> @llvm.riscv.vfncvt.xu.f.w.nxv16i16.nxv16f32(
<vscale x 16 x i16>,
<vscale x 16 x float>,
iXLen);

Expand All @@ -436,6 +457,7 @@ define <vscale x 16 x i16> @intrinsic_vfncvt_xu.f.w_nxv16i16_nxv16f32(<vscale x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x i16> @llvm.riscv.vfncvt.xu.f.w.nxv16i16.nxv16f32(
<vscale x 16 x i16> undef,
<vscale x 16 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -466,6 +488,7 @@ entry:
}

declare <vscale x 1 x i32> @llvm.riscv.vfncvt.xu.f.w.nxv1i32.nxv1f64(
<vscale x 1 x i32>,
<vscale x 1 x double>,
iXLen);

Expand All @@ -478,6 +501,7 @@ define <vscale x 1 x i32> @intrinsic_vfncvt_xu.f.w_nxv1i32_nxv1f64(<vscale x 1 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x i32> @llvm.riscv.vfncvt.xu.f.w.nxv1i32.nxv1f64(
<vscale x 1 x i32> undef,
<vscale x 1 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -508,6 +532,7 @@ entry:
}

declare <vscale x 2 x i32> @llvm.riscv.vfncvt.xu.f.w.nxv2i32.nxv2f64(
<vscale x 2 x i32>,
<vscale x 2 x double>,
iXLen);

Expand All @@ -520,6 +545,7 @@ define <vscale x 2 x i32> @intrinsic_vfncvt_xu.f.w_nxv2i32_nxv2f64(<vscale x 2 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x i32> @llvm.riscv.vfncvt.xu.f.w.nxv2i32.nxv2f64(
<vscale x 2 x i32> undef,
<vscale x 2 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -550,6 +576,7 @@ entry:
}

declare <vscale x 4 x i32> @llvm.riscv.vfncvt.xu.f.w.nxv4i32.nxv4f64(
<vscale x 4 x i32>,
<vscale x 4 x double>,
iXLen);

Expand All @@ -562,6 +589,7 @@ define <vscale x 4 x i32> @intrinsic_vfncvt_xu.f.w_nxv4i32_nxv4f64(<vscale x 4 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x i32> @llvm.riscv.vfncvt.xu.f.w.nxv4i32.nxv4f64(
<vscale x 4 x i32> undef,
<vscale x 4 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -592,6 +620,7 @@ entry:
}

declare <vscale x 8 x i32> @llvm.riscv.vfncvt.xu.f.w.nxv8i32.nxv8f64(
<vscale x 8 x i32>,
<vscale x 8 x double>,
iXLen);

Expand All @@ -604,6 +633,7 @@ define <vscale x 8 x i32> @intrinsic_vfncvt_xu.f.w_nxv8i32_nxv8f64(<vscale x 8 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x i32> @llvm.riscv.vfncvt.xu.f.w.nxv8i32.nxv8f64(
<vscale x 8 x i32> undef,
<vscale x 8 x double> %0,
iXLen %1)

Expand Down
30 changes: 30 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfrec7.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x half> @llvm.riscv.vfrec7.nxv1f16(
<vscale x 1 x half>,
<vscale x 1 x half>,
iXLen);

Expand All @@ -15,6 +16,7 @@ define <vscale x 1 x half> @intrinsic_vfrec7_v_nxv1f16_nxv1f16(<vscale x 1 x hal
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x half> @llvm.riscv.vfrec7.nxv1f16(
<vscale x 1 x half> undef,
<vscale x 1 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -45,6 +47,7 @@ entry:
}

declare <vscale x 2 x half> @llvm.riscv.vfrec7.nxv2f16(
<vscale x 2 x half>,
<vscale x 2 x half>,
iXLen);

Expand All @@ -56,6 +59,7 @@ define <vscale x 2 x half> @intrinsic_vfrec7_v_nxv2f16_nxv2f16(<vscale x 2 x hal
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x half> @llvm.riscv.vfrec7.nxv2f16(
<vscale x 2 x half> undef,
<vscale x 2 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -86,6 +90,7 @@ entry:
}

declare <vscale x 4 x half> @llvm.riscv.vfrec7.nxv4f16(
<vscale x 4 x half>,
<vscale x 4 x half>,
iXLen);

Expand All @@ -97,6 +102,7 @@ define <vscale x 4 x half> @intrinsic_vfrec7_v_nxv4f16_nxv4f16(<vscale x 4 x hal
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x half> @llvm.riscv.vfrec7.nxv4f16(
<vscale x 4 x half> undef,
<vscale x 4 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -127,6 +133,7 @@ entry:
}

declare <vscale x 8 x half> @llvm.riscv.vfrec7.nxv8f16(
<vscale x 8 x half>,
<vscale x 8 x half>,
iXLen);

Expand All @@ -138,6 +145,7 @@ define <vscale x 8 x half> @intrinsic_vfrec7_v_nxv8f16_nxv8f16(<vscale x 8 x hal
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x half> @llvm.riscv.vfrec7.nxv8f16(
<vscale x 8 x half> undef,
<vscale x 8 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -168,6 +176,7 @@ entry:
}

declare <vscale x 16 x half> @llvm.riscv.vfrec7.nxv16f16(
<vscale x 16 x half>,
<vscale x 16 x half>,
iXLen);

Expand All @@ -179,6 +188,7 @@ define <vscale x 16 x half> @intrinsic_vfrec7_v_nxv16f16_nxv16f16(<vscale x 16 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x half> @llvm.riscv.vfrec7.nxv16f16(
<vscale x 16 x half> undef,
<vscale x 16 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -209,6 +219,7 @@ entry:
}

declare <vscale x 32 x half> @llvm.riscv.vfrec7.nxv32f16(
<vscale x 32 x half>,
<vscale x 32 x half>,
iXLen);

Expand All @@ -220,6 +231,7 @@ define <vscale x 32 x half> @intrinsic_vfrec7_v_nxv32f16_nxv32f16(<vscale x 32 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 32 x half> @llvm.riscv.vfrec7.nxv32f16(
<vscale x 32 x half> undef,
<vscale x 32 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -250,6 +262,7 @@ entry:
}

declare <vscale x 1 x float> @llvm.riscv.vfrec7.nxv1f32(
<vscale x 1 x float>,
<vscale x 1 x float>,
iXLen);

Expand All @@ -261,6 +274,7 @@ define <vscale x 1 x float> @intrinsic_vfrec7_v_nxv1f32_nxv1f32(<vscale x 1 x fl
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x float> @llvm.riscv.vfrec7.nxv1f32(
<vscale x 1 x float> undef,
<vscale x 1 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -291,6 +305,7 @@ entry:
}

declare <vscale x 2 x float> @llvm.riscv.vfrec7.nxv2f32(
<vscale x 2 x float>,
<vscale x 2 x float>,
iXLen);

Expand All @@ -302,6 +317,7 @@ define <vscale x 2 x float> @intrinsic_vfrec7_v_nxv2f32_nxv2f32(<vscale x 2 x fl
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x float> @llvm.riscv.vfrec7.nxv2f32(
<vscale x 2 x float> undef,
<vscale x 2 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -332,6 +348,7 @@ entry:
}

declare <vscale x 4 x float> @llvm.riscv.vfrec7.nxv4f32(
<vscale x 4 x float>,
<vscale x 4 x float>,
iXLen);

Expand All @@ -343,6 +360,7 @@ define <vscale x 4 x float> @intrinsic_vfrec7_v_nxv4f32_nxv4f32(<vscale x 4 x fl
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x float> @llvm.riscv.vfrec7.nxv4f32(
<vscale x 4 x float> undef,
<vscale x 4 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -373,6 +391,7 @@ entry:
}

declare <vscale x 8 x float> @llvm.riscv.vfrec7.nxv8f32(
<vscale x 8 x float>,
<vscale x 8 x float>,
iXLen);

Expand All @@ -384,6 +403,7 @@ define <vscale x 8 x float> @intrinsic_vfrec7_v_nxv8f32_nxv8f32(<vscale x 8 x fl
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x float> @llvm.riscv.vfrec7.nxv8f32(
<vscale x 8 x float> undef,
<vscale x 8 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -414,6 +434,7 @@ entry:
}

declare <vscale x 16 x float> @llvm.riscv.vfrec7.nxv16f32(
<vscale x 16 x float>,
<vscale x 16 x float>,
iXLen);

Expand All @@ -425,6 +446,7 @@ define <vscale x 16 x float> @intrinsic_vfrec7_v_nxv16f32_nxv16f32(<vscale x 16
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x float> @llvm.riscv.vfrec7.nxv16f32(
<vscale x 16 x float> undef,
<vscale x 16 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -455,6 +477,7 @@ entry:
}

declare <vscale x 1 x double> @llvm.riscv.vfrec7.nxv1f64(
<vscale x 1 x double>,
<vscale x 1 x double>,
iXLen);

Expand All @@ -466,6 +489,7 @@ define <vscale x 1 x double> @intrinsic_vfrec7_v_nxv1f64_nxv1f64(<vscale x 1 x d
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x double> @llvm.riscv.vfrec7.nxv1f64(
<vscale x 1 x double> undef,
<vscale x 1 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -496,6 +520,7 @@ entry:
}

declare <vscale x 2 x double> @llvm.riscv.vfrec7.nxv2f64(
<vscale x 2 x double>,
<vscale x 2 x double>,
iXLen);

Expand All @@ -507,6 +532,7 @@ define <vscale x 2 x double> @intrinsic_vfrec7_v_nxv2f64_nxv2f64(<vscale x 2 x d
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x double> @llvm.riscv.vfrec7.nxv2f64(
<vscale x 2 x double> undef,
<vscale x 2 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -537,6 +563,7 @@ entry:
}

declare <vscale x 4 x double> @llvm.riscv.vfrec7.nxv4f64(
<vscale x 4 x double>,
<vscale x 4 x double>,
iXLen);

Expand All @@ -548,6 +575,7 @@ define <vscale x 4 x double> @intrinsic_vfrec7_v_nxv4f64_nxv4f64(<vscale x 4 x d
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x double> @llvm.riscv.vfrec7.nxv4f64(
<vscale x 4 x double> undef,
<vscale x 4 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -578,6 +606,7 @@ entry:
}

declare <vscale x 8 x double> @llvm.riscv.vfrec7.nxv8f64(
<vscale x 8 x double>,
<vscale x 8 x double>,
iXLen);

Expand All @@ -589,6 +618,7 @@ define <vscale x 8 x double> @intrinsic_vfrec7_v_nxv8f64_nxv8f64(<vscale x 8 x d
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x double> @llvm.riscv.vfrec7.nxv8f64(
<vscale x 8 x double> undef,
<vscale x 8 x double> %0,
iXLen %1)

Expand Down
30 changes: 30 additions & 0 deletions llvm/test/CodeGen/RISCV/rvv/vfrsqrt7.ll
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
; RUN: sed 's/iXLen/i64/g' %s | llc -mtriple=riscv64 -mattr=+v,+zfh \
; RUN: -verify-machineinstrs -target-abi=lp64d | FileCheck %s
declare <vscale x 1 x half> @llvm.riscv.vfrsqrt7.nxv1f16(
<vscale x 1 x half>,
<vscale x 1 x half>,
iXLen);

Expand All @@ -15,6 +16,7 @@ define <vscale x 1 x half> @intrinsic_vfrsqrt7_v_nxv1f16_nxv1f16(<vscale x 1 x h
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x half> @llvm.riscv.vfrsqrt7.nxv1f16(
<vscale x 1 x half> undef,
<vscale x 1 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -45,6 +47,7 @@ entry:
}

declare <vscale x 2 x half> @llvm.riscv.vfrsqrt7.nxv2f16(
<vscale x 2 x half>,
<vscale x 2 x half>,
iXLen);

Expand All @@ -56,6 +59,7 @@ define <vscale x 2 x half> @intrinsic_vfrsqrt7_v_nxv2f16_nxv2f16(<vscale x 2 x h
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x half> @llvm.riscv.vfrsqrt7.nxv2f16(
<vscale x 2 x half> undef,
<vscale x 2 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -86,6 +90,7 @@ entry:
}

declare <vscale x 4 x half> @llvm.riscv.vfrsqrt7.nxv4f16(
<vscale x 4 x half>,
<vscale x 4 x half>,
iXLen);

Expand All @@ -97,6 +102,7 @@ define <vscale x 4 x half> @intrinsic_vfrsqrt7_v_nxv4f16_nxv4f16(<vscale x 4 x h
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x half> @llvm.riscv.vfrsqrt7.nxv4f16(
<vscale x 4 x half> undef,
<vscale x 4 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -127,6 +133,7 @@ entry:
}

declare <vscale x 8 x half> @llvm.riscv.vfrsqrt7.nxv8f16(
<vscale x 8 x half>,
<vscale x 8 x half>,
iXLen);

Expand All @@ -138,6 +145,7 @@ define <vscale x 8 x half> @intrinsic_vfrsqrt7_v_nxv8f16_nxv8f16(<vscale x 8 x h
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x half> @llvm.riscv.vfrsqrt7.nxv8f16(
<vscale x 8 x half> undef,
<vscale x 8 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -168,6 +176,7 @@ entry:
}

declare <vscale x 16 x half> @llvm.riscv.vfrsqrt7.nxv16f16(
<vscale x 16 x half>,
<vscale x 16 x half>,
iXLen);

Expand All @@ -179,6 +188,7 @@ define <vscale x 16 x half> @intrinsic_vfrsqrt7_v_nxv16f16_nxv16f16(<vscale x 16
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x half> @llvm.riscv.vfrsqrt7.nxv16f16(
<vscale x 16 x half> undef,
<vscale x 16 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -209,6 +219,7 @@ entry:
}

declare <vscale x 32 x half> @llvm.riscv.vfrsqrt7.nxv32f16(
<vscale x 32 x half>,
<vscale x 32 x half>,
iXLen);

Expand All @@ -220,6 +231,7 @@ define <vscale x 32 x half> @intrinsic_vfrsqrt7_v_nxv32f16_nxv32f16(<vscale x 32
; CHECK-NEXT: ret
entry:
%a = call <vscale x 32 x half> @llvm.riscv.vfrsqrt7.nxv32f16(
<vscale x 32 x half> undef,
<vscale x 32 x half> %0,
iXLen %1)

Expand Down Expand Up @@ -250,6 +262,7 @@ entry:
}

declare <vscale x 1 x float> @llvm.riscv.vfrsqrt7.nxv1f32(
<vscale x 1 x float>,
<vscale x 1 x float>,
iXLen);

Expand All @@ -261,6 +274,7 @@ define <vscale x 1 x float> @intrinsic_vfrsqrt7_v_nxv1f32_nxv1f32(<vscale x 1 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x float> @llvm.riscv.vfrsqrt7.nxv1f32(
<vscale x 1 x float> undef,
<vscale x 1 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -291,6 +305,7 @@ entry:
}

declare <vscale x 2 x float> @llvm.riscv.vfrsqrt7.nxv2f32(
<vscale x 2 x float>,
<vscale x 2 x float>,
iXLen);

Expand All @@ -302,6 +317,7 @@ define <vscale x 2 x float> @intrinsic_vfrsqrt7_v_nxv2f32_nxv2f32(<vscale x 2 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x float> @llvm.riscv.vfrsqrt7.nxv2f32(
<vscale x 2 x float> undef,
<vscale x 2 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -332,6 +348,7 @@ entry:
}

declare <vscale x 4 x float> @llvm.riscv.vfrsqrt7.nxv4f32(
<vscale x 4 x float>,
<vscale x 4 x float>,
iXLen);

Expand All @@ -343,6 +360,7 @@ define <vscale x 4 x float> @intrinsic_vfrsqrt7_v_nxv4f32_nxv4f32(<vscale x 4 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x float> @llvm.riscv.vfrsqrt7.nxv4f32(
<vscale x 4 x float> undef,
<vscale x 4 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -373,6 +391,7 @@ entry:
}

declare <vscale x 8 x float> @llvm.riscv.vfrsqrt7.nxv8f32(
<vscale x 8 x float>,
<vscale x 8 x float>,
iXLen);

Expand All @@ -384,6 +403,7 @@ define <vscale x 8 x float> @intrinsic_vfrsqrt7_v_nxv8f32_nxv8f32(<vscale x 8 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x float> @llvm.riscv.vfrsqrt7.nxv8f32(
<vscale x 8 x float> undef,
<vscale x 8 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -414,6 +434,7 @@ entry:
}

declare <vscale x 16 x float> @llvm.riscv.vfrsqrt7.nxv16f32(
<vscale x 16 x float>,
<vscale x 16 x float>,
iXLen);

Expand All @@ -425,6 +446,7 @@ define <vscale x 16 x float> @intrinsic_vfrsqrt7_v_nxv16f32_nxv16f32(<vscale x 1
; CHECK-NEXT: ret
entry:
%a = call <vscale x 16 x float> @llvm.riscv.vfrsqrt7.nxv16f32(
<vscale x 16 x float> undef,
<vscale x 16 x float> %0,
iXLen %1)

Expand Down Expand Up @@ -455,6 +477,7 @@ entry:
}

declare <vscale x 1 x double> @llvm.riscv.vfrsqrt7.nxv1f64(
<vscale x 1 x double>,
<vscale x 1 x double>,
iXLen);

Expand All @@ -466,6 +489,7 @@ define <vscale x 1 x double> @intrinsic_vfrsqrt7_v_nxv1f64_nxv1f64(<vscale x 1 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 1 x double> @llvm.riscv.vfrsqrt7.nxv1f64(
<vscale x 1 x double> undef,
<vscale x 1 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -496,6 +520,7 @@ entry:
}

declare <vscale x 2 x double> @llvm.riscv.vfrsqrt7.nxv2f64(
<vscale x 2 x double>,
<vscale x 2 x double>,
iXLen);

Expand All @@ -507,6 +532,7 @@ define <vscale x 2 x double> @intrinsic_vfrsqrt7_v_nxv2f64_nxv2f64(<vscale x 2 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 2 x double> @llvm.riscv.vfrsqrt7.nxv2f64(
<vscale x 2 x double> undef,
<vscale x 2 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -537,6 +563,7 @@ entry:
}

declare <vscale x 4 x double> @llvm.riscv.vfrsqrt7.nxv4f64(
<vscale x 4 x double>,
<vscale x 4 x double>,
iXLen);

Expand All @@ -548,6 +575,7 @@ define <vscale x 4 x double> @intrinsic_vfrsqrt7_v_nxv4f64_nxv4f64(<vscale x 4 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 4 x double> @llvm.riscv.vfrsqrt7.nxv4f64(
<vscale x 4 x double> undef,
<vscale x 4 x double> %0,
iXLen %1)

Expand Down Expand Up @@ -578,6 +606,7 @@ entry:
}

declare <vscale x 8 x double> @llvm.riscv.vfrsqrt7.nxv8f64(
<vscale x 8 x double>,
<vscale x 8 x double>,
iXLen);

Expand All @@ -589,6 +618,7 @@ define <vscale x 8 x double> @intrinsic_vfrsqrt7_v_nxv8f64_nxv8f64(<vscale x 8 x
; CHECK-NEXT: ret
entry:
%a = call <vscale x 8 x double> @llvm.riscv.vfrsqrt7.nxv8f64(
<vscale x 8 x double> undef,
<vscale x 8 x double> %0,
iXLen %1)

Expand Down
Loading