-
Notifications
You must be signed in to change notification settings - Fork 14.9k
[DAGCombiner] Lower dynamic insertelt chain more efficiently #162368
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
@llvm/pr-subscribers-backend-powerpc @llvm/pr-subscribers-backend-nvptx Author: Princeton Ferro (Prince781) ChangesA chain of dynamic insertelts (that is: insertelt (insertelt (...)) with dynamic indices) can be spilled at once. This avoids each insertelt being spilled in DAGTypeLegalizer which reduces code size and compile time. Full diff: https://github.com/llvm/llvm-project/pull/162368.diff 2 Files Affected:
diff --git a/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp b/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
index 309f1bea8b77c..3071aac3a511a 100644
--- a/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
+++ b/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
@@ -23445,6 +23445,57 @@ SDValue DAGCombiner::visitINSERT_VECTOR_ELT(SDNode *N) {
// inselt undef, InVal, EltNo --> build_vector < InVal, InVal, ... >
if (InVec.isUndef() && TLI.shouldSplatInsEltVarIndex(VT))
return DAG.getSplat(VT, DL, InVal);
+
+ // For dynamic insertelts, the type legalizer will spill the entire vector.
+ // For a chain of dynamic insertelts, this can be really inefficient and
+ // bad for compile time. If each insertelt is only fed into the next, the
+ // vector is write-only across this chain, and we can just spill once.
+ SmallVector<SDNode *> Seq{N};
+ while (true) {
+ SDValue InVec = Seq.back()->getOperand(0);
+ SDValue EltNo = InVec.getOperand(2);
+ if (!(InVec.getOpcode() == ISD::INSERT_VECTOR_ELT &&
+ !isa<ConstantSDNode>(EltNo)))
+ break;
+ Seq.push_back(InVec.getNode());
+ }
+
+ // Only care about chains, otherwise this instruction can be handled by
+ // the type legalizer just fine.
+ if (Seq.size() > 1) {
+ // In cases where the vector is illegal it will be broken down into parts
+ // and stored in parts - we should use the alignment for the smallest part.
+ Align SmallestAlign = DAG.getReducedAlign(VT, /*UseABI=*/false);
+ SDValue StackPtr = DAG.CreateStackTemporary(VT.getStoreSize(), SmallestAlign);
+ auto &MF = DAG.getMachineFunction();
+ auto FrameIndex = cast<FrameIndexSDNode>(StackPtr.getNode())->getIndex();
+ auto PtrInfo = MachinePointerInfo::getFixedStack(MF, FrameIndex);
+
+ // Begin spilling
+ SDValue InVec = Seq.back()->getOperand(0);
+ SDValue Store = DAG.getStore(DAG.getEntryNode(), DL, InVec, StackPtr, PtrInfo,
+ SmallestAlign);
+
+ // Lower each dynamic insertelt to a store
+ for (SDNode *N : reverse(Seq)) {
+ SDValue Elmnt = N->getOperand(1);
+ SDValue Index = N->getOperand(2);
+
+ // Store the new element. This may be larger than the vector element type,
+ // so use a truncating store.
+ SDValue EltPtr = TLI.getVectorElementPointer(DAG, StackPtr, VT, Index);
+ EVT EltVT = Elmnt.getValueType();
+ Store = DAG.getTruncStore(
+ Store, DL, Elmnt, EltPtr, MachinePointerInfo::getUnknownStack(MF),
+ EltVT, commonAlignment(SmallestAlign,
+ EltVT.getFixedSizeInBits() / 8));
+ }
+
+ // Load the spilled vector
+ SDValue Load = DAG.getLoad(VT, DL, Store, StackPtr, PtrInfo, SmallestAlign);
+ return Load.getValue(0);
+ }
+
return SDValue();
}
diff --git a/llvm/test/CodeGen/NVPTX/vector-spill.ll b/llvm/test/CodeGen/NVPTX/vector-spill.ll
new file mode 100644
index 0000000000000..0a26069b02cc2
--- /dev/null
+++ b/llvm/test/CodeGen/NVPTX/vector-spill.ll
@@ -0,0 +1,215 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 6
+; RUN: llc < %s -mcpu=sm_20 | FileCheck %s
+; RUN: %if ptxas %{ llc < %s -mcpu=sm_20 | %ptxas-verify %}
+target datalayout = "e-i64:64-v16:16-v32:32-n16:32:64"
+target triple = "nvptx64-nvidia-cuda"
+
+; COM: Spill the vector once.
+define ptx_kernel void @spill_once(ptr addrspace(3) %shared.mem, <8 x double> %vector, i32 %idx0, i32 %idx1, i32 %idx2, i32 %idx3) local_unnamed_addr {
+; CHECK-LABEL: spill_once(
+; CHECK: {
+; CHECK-NEXT: .local .align 8 .b8 __local_depot0[64];
+; CHECK-NEXT: .reg .b64 %SP;
+; CHECK-NEXT: .reg .b64 %SPL;
+; CHECK-NEXT: .reg .b64 %rd<39>;
+; CHECK-EMPTY:
+; CHECK-NEXT: // %bb.0: // %entry
+; CHECK-NEXT: mov.b64 %SPL, __local_depot0;
+; CHECK-NEXT: cvta.local.u64 %SP, %SPL;
+; CHECK-NEXT: ld.param.b64 %rd1, [spill_once_param_0];
+; CHECK-NEXT: ld.param.v2.b64 {%rd2, %rd3}, [spill_once_param_1];
+; CHECK-NEXT: ld.param.v2.b64 {%rd4, %rd5}, [spill_once_param_1+16];
+; CHECK-NEXT: ld.param.v2.b64 {%rd6, %rd7}, [spill_once_param_1+32];
+; CHECK-NEXT: ld.param.v2.b64 {%rd8, %rd9}, [spill_once_param_1+48];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd10, %rd11}, [%rd1];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd12, %rd13}, [%rd1+16];
+; CHECK-NEXT: ld.param.b32 %rd14, [spill_once_param_2];
+; CHECK-NEXT: shl.b64 %rd15, %rd14, 3;
+; CHECK-NEXT: and.b64 %rd16, %rd15, 56;
+; CHECK-NEXT: add.u64 %rd17, %SP, 0;
+; CHECK-NEXT: add.s64 %rd18, %rd17, %rd16;
+; CHECK-NEXT: ld.param.b32 %rd19, [spill_once_param_3];
+; CHECK-NEXT: shl.b64 %rd20, %rd19, 3;
+; CHECK-NEXT: and.b64 %rd21, %rd20, 56;
+; CHECK-NEXT: add.s64 %rd22, %rd17, %rd21;
+; CHECK-NEXT: ld.param.b32 %rd23, [spill_once_param_4];
+; CHECK-NEXT: shl.b64 %rd24, %rd23, 3;
+; CHECK-NEXT: and.b64 %rd25, %rd24, 56;
+; CHECK-NEXT: add.s64 %rd26, %rd17, %rd25;
+; CHECK-NEXT: st.b64 [%SP+56], %rd9;
+; CHECK-NEXT: st.b64 [%SP+48], %rd8;
+; CHECK-NEXT: st.b64 [%SP+40], %rd7;
+; CHECK-NEXT: st.b64 [%SP+32], %rd6;
+; CHECK-NEXT: st.b64 [%SP+24], %rd5;
+; CHECK-NEXT: st.b64 [%SP+16], %rd4;
+; CHECK-NEXT: st.b64 [%SP+8], %rd3;
+; CHECK-NEXT: st.b64 [%SP], %rd2;
+; CHECK-NEXT: st.b64 [%rd18], %rd10;
+; CHECK-NEXT: st.b64 [%rd22], %rd11;
+; CHECK-NEXT: st.b64 [%rd26], %rd12;
+; CHECK-NEXT: ld.param.b32 %rd27, [spill_once_param_5];
+; CHECK-NEXT: shl.b64 %rd28, %rd27, 3;
+; CHECK-NEXT: and.b64 %rd29, %rd28, 56;
+; CHECK-NEXT: add.s64 %rd30, %rd17, %rd29;
+; CHECK-NEXT: st.b64 [%rd30], %rd13;
+; CHECK-NEXT: ld.b64 %rd31, [%SP+8];
+; CHECK-NEXT: ld.b64 %rd32, [%SP];
+; CHECK-NEXT: ld.b64 %rd33, [%SP+24];
+; CHECK-NEXT: ld.b64 %rd34, [%SP+16];
+; CHECK-NEXT: ld.b64 %rd35, [%SP+40];
+; CHECK-NEXT: ld.b64 %rd36, [%SP+32];
+; CHECK-NEXT: ld.b64 %rd37, [%SP+56];
+; CHECK-NEXT: ld.b64 %rd38, [%SP+48];
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1072], {%rd38, %rd37};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1056], {%rd36, %rd35};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1040], {%rd34, %rd33};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1024], {%rd32, %rd31};
+; CHECK-NEXT: ret;
+entry:
+ %offset.0 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 0
+ %element.0 = load double, ptr addrspace(3) %offset.0, align 64
+ %offset.1 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 8
+ %element.1 = load double, ptr addrspace(3) %offset.1, align 8
+ %offset.2 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 16
+ %element.2 = load double, ptr addrspace(3) %offset.2, align 8
+ %offset.3 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 24
+ %element.3 = load double, ptr addrspace(3) %offset.3, align 8
+ %vector.build0 = insertelement <8 x double> %vector, double %element.0, i32 %idx0
+ %vector.build1 = insertelement <8 x double> %vector.build0, double %element.1, i32 %idx1
+ %vector.build2 = insertelement <8 x double> %vector.build1, double %element.2, i32 %idx2
+ %vector.build3 = insertelement <8 x double> %vector.build2, double %element.3, i32 %idx3
+ %location = getelementptr i8, ptr addrspace(3) %shared.mem, i32 1024
+ store <8 x double> %vector.build3, ptr addrspace(3) %location, align 64
+ ret void
+}
+
+; COM: Spill the vector twice. Because these are in two different slots, the
+; resulting spill codes may be non-overlapping even though the insertelt
+; sequences overlap.
+define ptx_kernel void @spill_twice(ptr addrspace(3) %shared.mem, <8 x double> %vector, i32 %idx0, i32 %idx1, i32 %idx2, i32 %idx3) local_unnamed_addr {
+; CHECK-LABEL: spill_twice(
+; CHECK: {
+; CHECK-NEXT: .local .align 8 .b8 __local_depot1[128];
+; CHECK-NEXT: .reg .b64 %SP;
+; CHECK-NEXT: .reg .b64 %SPL;
+; CHECK-NEXT: .reg .b32 %r<7>;
+; CHECK-NEXT: .reg .b64 %rd<54>;
+; CHECK-EMPTY:
+; CHECK-NEXT: // %bb.0: // %entry
+; CHECK-NEXT: mov.b64 %SPL, __local_depot1;
+; CHECK-NEXT: cvta.local.u64 %SP, %SPL;
+; CHECK-NEXT: ld.param.b64 %rd1, [spill_twice_param_0];
+; CHECK-NEXT: ld.param.v2.b64 {%rd2, %rd3}, [spill_twice_param_1];
+; CHECK-NEXT: ld.param.v2.b64 {%rd4, %rd5}, [spill_twice_param_1+16];
+; CHECK-NEXT: ld.param.v2.b64 {%rd6, %rd7}, [spill_twice_param_1+32];
+; CHECK-NEXT: ld.param.v2.b64 {%rd8, %rd9}, [spill_twice_param_1+48];
+; CHECK-NEXT: ld.param.b32 %r1, [spill_twice_param_2];
+; CHECK-NEXT: ld.param.b32 %r2, [spill_twice_param_3];
+; CHECK-NEXT: ld.param.b32 %r3, [spill_twice_param_4];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd10, %rd11}, [%rd1];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd12, %rd13}, [%rd1+16];
+; CHECK-NEXT: mul.wide.u32 %rd14, %r1, 8;
+; CHECK-NEXT: and.b64 %rd15, %rd14, 56;
+; CHECK-NEXT: add.u64 %rd16, %SP, 0;
+; CHECK-NEXT: add.s64 %rd17, %rd16, %rd15;
+; CHECK-NEXT: shl.b32 %r4, %r1, 3;
+; CHECK-NEXT: cvt.u64.u32 %rd18, %r4;
+; CHECK-NEXT: and.b64 %rd19, %rd18, 56;
+; CHECK-NEXT: add.u64 %rd20, %SP, 64;
+; CHECK-NEXT: add.s64 %rd21, %rd20, %rd19;
+; CHECK-NEXT: mul.wide.u32 %rd22, %r2, 8;
+; CHECK-NEXT: and.b64 %rd23, %rd22, 56;
+; CHECK-NEXT: add.s64 %rd24, %rd16, %rd23;
+; CHECK-NEXT: shl.b32 %r5, %r2, 3;
+; CHECK-NEXT: cvt.u64.u32 %rd25, %r5;
+; CHECK-NEXT: and.b64 %rd26, %rd25, 56;
+; CHECK-NEXT: add.s64 %rd27, %rd20, %rd26;
+; CHECK-NEXT: st.b64 [%SP+120], %rd9;
+; CHECK-NEXT: st.b64 [%SP+112], %rd8;
+; CHECK-NEXT: st.b64 [%SP+104], %rd7;
+; CHECK-NEXT: st.b64 [%SP+96], %rd6;
+; CHECK-NEXT: st.b64 [%SP+88], %rd5;
+; CHECK-NEXT: st.b64 [%SP+80], %rd4;
+; CHECK-NEXT: st.b64 [%SP+72], %rd3;
+; CHECK-NEXT: st.b64 [%SP+64], %rd2;
+; CHECK-NEXT: st.b64 [%rd21], %rd10;
+; CHECK-NEXT: st.b64 [%rd27], %rd11;
+; CHECK-NEXT: shl.b32 %r6, %r3, 3;
+; CHECK-NEXT: cvt.u64.u32 %rd28, %r6;
+; CHECK-NEXT: and.b64 %rd29, %rd28, 56;
+; CHECK-NEXT: add.s64 %rd30, %rd20, %rd29;
+; CHECK-NEXT: st.b64 [%rd30], %rd12;
+; CHECK-NEXT: mul.wide.u32 %rd31, %r3, 8;
+; CHECK-NEXT: and.b64 %rd32, %rd31, 56;
+; CHECK-NEXT: add.s64 %rd33, %rd16, %rd32;
+; CHECK-NEXT: ld.b64 %rd34, [%SP+72];
+; CHECK-NEXT: ld.b64 %rd35, [%SP+64];
+; CHECK-NEXT: ld.b64 %rd36, [%SP+88];
+; CHECK-NEXT: ld.b64 %rd37, [%SP+80];
+; CHECK-NEXT: ld.b64 %rd38, [%SP+104];
+; CHECK-NEXT: ld.b64 %rd39, [%SP+96];
+; CHECK-NEXT: ld.b64 %rd40, [%SP+120];
+; CHECK-NEXT: ld.b64 %rd41, [%SP+112];
+; CHECK-NEXT: st.b64 [%SP+56], %rd9;
+; CHECK-NEXT: st.b64 [%SP+48], %rd8;
+; CHECK-NEXT: st.b64 [%SP+40], %rd7;
+; CHECK-NEXT: st.b64 [%SP+32], %rd6;
+; CHECK-NEXT: st.b64 [%SP+24], %rd5;
+; CHECK-NEXT: st.b64 [%SP+16], %rd4;
+; CHECK-NEXT: st.b64 [%SP+8], %rd3;
+; CHECK-NEXT: st.b64 [%SP], %rd2;
+; CHECK-NEXT: st.b64 [%rd17], %rd10;
+; CHECK-NEXT: st.b64 [%rd24], %rd11;
+; CHECK-NEXT: st.b64 [%rd33], %rd12;
+; CHECK-NEXT: ld.param.b32 %rd42, [spill_twice_param_5];
+; CHECK-NEXT: shl.b64 %rd43, %rd42, 3;
+; CHECK-NEXT: and.b64 %rd44, %rd43, 56;
+; CHECK-NEXT: add.s64 %rd45, %rd16, %rd44;
+; CHECK-NEXT: st.b64 [%rd45], %rd13;
+; CHECK-NEXT: ld.b64 %rd46, [%SP+8];
+; CHECK-NEXT: ld.b64 %rd47, [%SP];
+; CHECK-NEXT: ld.b64 %rd48, [%SP+24];
+; CHECK-NEXT: ld.b64 %rd49, [%SP+16];
+; CHECK-NEXT: ld.b64 %rd50, [%SP+40];
+; CHECK-NEXT: ld.b64 %rd51, [%SP+32];
+; CHECK-NEXT: ld.b64 %rd52, [%SP+56];
+; CHECK-NEXT: ld.b64 %rd53, [%SP+48];
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1072], {%rd53, %rd52};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1056], {%rd51, %rd50};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1040], {%rd49, %rd48};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1024], {%rd47, %rd46};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1144], {%rd41, %rd40};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1128], {%rd39, %rd38};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1112], {%rd37, %rd36};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1096], {%rd35, %rd34};
+; CHECK-NEXT: ret;
+entry:
+ %offset.0 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 0
+ %element.0 = load double, ptr addrspace(3) %offset.0, align 64
+ %offset.1 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 8
+ %element.1 = load double, ptr addrspace(3) %offset.1, align 8
+ %offset.2 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 16
+ %element.2 = load double, ptr addrspace(3) %offset.2, align 8
+ %offset.3 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 24
+ %element.3 = load double, ptr addrspace(3) %offset.3, align 8
+
+; COM: begin chain 1
+ %vector.build0 = insertelement <8 x double> %vector, double %element.0, i32 %idx0
+ %vector.build1 = insertelement <8 x double> %vector.build0, double %element.1, i32 %idx1
+
+; COM: interleave a second chain of insertelements
+ %vector.build1-2 = insertelement <8 x double> %vector.build1, double %element.2, i32 %idx2
+
+; COM: continue chain 1
+ %vector.build2 = insertelement <8 x double> %vector.build1, double %element.2, i32 %idx2
+ %vector.build3 = insertelement <8 x double> %vector.build2, double %element.3, i32 %idx3
+
+; COM: save chain 1
+ %location = getelementptr i8, ptr addrspace(3) %shared.mem, i32 1024
+ store <8 x double> %vector.build3, ptr addrspace(3) %location, align 64
+
+; COM: save chain 2
+ %location-2 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 1096
+ store <8 x double> %vector.build1-2, ptr addrspace(3) %location-2, align 64
+ ret void
+}
|
@llvm/pr-subscribers-llvm-selectiondag Author: Princeton Ferro (Prince781) ChangesA chain of dynamic insertelts (that is: insertelt (insertelt (...)) with dynamic indices) can be spilled at once. This avoids each insertelt being spilled in DAGTypeLegalizer which reduces code size and compile time. Full diff: https://github.com/llvm/llvm-project/pull/162368.diff 2 Files Affected:
diff --git a/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp b/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
index 309f1bea8b77c..3071aac3a511a 100644
--- a/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
+++ b/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
@@ -23445,6 +23445,57 @@ SDValue DAGCombiner::visitINSERT_VECTOR_ELT(SDNode *N) {
// inselt undef, InVal, EltNo --> build_vector < InVal, InVal, ... >
if (InVec.isUndef() && TLI.shouldSplatInsEltVarIndex(VT))
return DAG.getSplat(VT, DL, InVal);
+
+ // For dynamic insertelts, the type legalizer will spill the entire vector.
+ // For a chain of dynamic insertelts, this can be really inefficient and
+ // bad for compile time. If each insertelt is only fed into the next, the
+ // vector is write-only across this chain, and we can just spill once.
+ SmallVector<SDNode *> Seq{N};
+ while (true) {
+ SDValue InVec = Seq.back()->getOperand(0);
+ SDValue EltNo = InVec.getOperand(2);
+ if (!(InVec.getOpcode() == ISD::INSERT_VECTOR_ELT &&
+ !isa<ConstantSDNode>(EltNo)))
+ break;
+ Seq.push_back(InVec.getNode());
+ }
+
+ // Only care about chains, otherwise this instruction can be handled by
+ // the type legalizer just fine.
+ if (Seq.size() > 1) {
+ // In cases where the vector is illegal it will be broken down into parts
+ // and stored in parts - we should use the alignment for the smallest part.
+ Align SmallestAlign = DAG.getReducedAlign(VT, /*UseABI=*/false);
+ SDValue StackPtr = DAG.CreateStackTemporary(VT.getStoreSize(), SmallestAlign);
+ auto &MF = DAG.getMachineFunction();
+ auto FrameIndex = cast<FrameIndexSDNode>(StackPtr.getNode())->getIndex();
+ auto PtrInfo = MachinePointerInfo::getFixedStack(MF, FrameIndex);
+
+ // Begin spilling
+ SDValue InVec = Seq.back()->getOperand(0);
+ SDValue Store = DAG.getStore(DAG.getEntryNode(), DL, InVec, StackPtr, PtrInfo,
+ SmallestAlign);
+
+ // Lower each dynamic insertelt to a store
+ for (SDNode *N : reverse(Seq)) {
+ SDValue Elmnt = N->getOperand(1);
+ SDValue Index = N->getOperand(2);
+
+ // Store the new element. This may be larger than the vector element type,
+ // so use a truncating store.
+ SDValue EltPtr = TLI.getVectorElementPointer(DAG, StackPtr, VT, Index);
+ EVT EltVT = Elmnt.getValueType();
+ Store = DAG.getTruncStore(
+ Store, DL, Elmnt, EltPtr, MachinePointerInfo::getUnknownStack(MF),
+ EltVT, commonAlignment(SmallestAlign,
+ EltVT.getFixedSizeInBits() / 8));
+ }
+
+ // Load the spilled vector
+ SDValue Load = DAG.getLoad(VT, DL, Store, StackPtr, PtrInfo, SmallestAlign);
+ return Load.getValue(0);
+ }
+
return SDValue();
}
diff --git a/llvm/test/CodeGen/NVPTX/vector-spill.ll b/llvm/test/CodeGen/NVPTX/vector-spill.ll
new file mode 100644
index 0000000000000..0a26069b02cc2
--- /dev/null
+++ b/llvm/test/CodeGen/NVPTX/vector-spill.ll
@@ -0,0 +1,215 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 6
+; RUN: llc < %s -mcpu=sm_20 | FileCheck %s
+; RUN: %if ptxas %{ llc < %s -mcpu=sm_20 | %ptxas-verify %}
+target datalayout = "e-i64:64-v16:16-v32:32-n16:32:64"
+target triple = "nvptx64-nvidia-cuda"
+
+; COM: Spill the vector once.
+define ptx_kernel void @spill_once(ptr addrspace(3) %shared.mem, <8 x double> %vector, i32 %idx0, i32 %idx1, i32 %idx2, i32 %idx3) local_unnamed_addr {
+; CHECK-LABEL: spill_once(
+; CHECK: {
+; CHECK-NEXT: .local .align 8 .b8 __local_depot0[64];
+; CHECK-NEXT: .reg .b64 %SP;
+; CHECK-NEXT: .reg .b64 %SPL;
+; CHECK-NEXT: .reg .b64 %rd<39>;
+; CHECK-EMPTY:
+; CHECK-NEXT: // %bb.0: // %entry
+; CHECK-NEXT: mov.b64 %SPL, __local_depot0;
+; CHECK-NEXT: cvta.local.u64 %SP, %SPL;
+; CHECK-NEXT: ld.param.b64 %rd1, [spill_once_param_0];
+; CHECK-NEXT: ld.param.v2.b64 {%rd2, %rd3}, [spill_once_param_1];
+; CHECK-NEXT: ld.param.v2.b64 {%rd4, %rd5}, [spill_once_param_1+16];
+; CHECK-NEXT: ld.param.v2.b64 {%rd6, %rd7}, [spill_once_param_1+32];
+; CHECK-NEXT: ld.param.v2.b64 {%rd8, %rd9}, [spill_once_param_1+48];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd10, %rd11}, [%rd1];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd12, %rd13}, [%rd1+16];
+; CHECK-NEXT: ld.param.b32 %rd14, [spill_once_param_2];
+; CHECK-NEXT: shl.b64 %rd15, %rd14, 3;
+; CHECK-NEXT: and.b64 %rd16, %rd15, 56;
+; CHECK-NEXT: add.u64 %rd17, %SP, 0;
+; CHECK-NEXT: add.s64 %rd18, %rd17, %rd16;
+; CHECK-NEXT: ld.param.b32 %rd19, [spill_once_param_3];
+; CHECK-NEXT: shl.b64 %rd20, %rd19, 3;
+; CHECK-NEXT: and.b64 %rd21, %rd20, 56;
+; CHECK-NEXT: add.s64 %rd22, %rd17, %rd21;
+; CHECK-NEXT: ld.param.b32 %rd23, [spill_once_param_4];
+; CHECK-NEXT: shl.b64 %rd24, %rd23, 3;
+; CHECK-NEXT: and.b64 %rd25, %rd24, 56;
+; CHECK-NEXT: add.s64 %rd26, %rd17, %rd25;
+; CHECK-NEXT: st.b64 [%SP+56], %rd9;
+; CHECK-NEXT: st.b64 [%SP+48], %rd8;
+; CHECK-NEXT: st.b64 [%SP+40], %rd7;
+; CHECK-NEXT: st.b64 [%SP+32], %rd6;
+; CHECK-NEXT: st.b64 [%SP+24], %rd5;
+; CHECK-NEXT: st.b64 [%SP+16], %rd4;
+; CHECK-NEXT: st.b64 [%SP+8], %rd3;
+; CHECK-NEXT: st.b64 [%SP], %rd2;
+; CHECK-NEXT: st.b64 [%rd18], %rd10;
+; CHECK-NEXT: st.b64 [%rd22], %rd11;
+; CHECK-NEXT: st.b64 [%rd26], %rd12;
+; CHECK-NEXT: ld.param.b32 %rd27, [spill_once_param_5];
+; CHECK-NEXT: shl.b64 %rd28, %rd27, 3;
+; CHECK-NEXT: and.b64 %rd29, %rd28, 56;
+; CHECK-NEXT: add.s64 %rd30, %rd17, %rd29;
+; CHECK-NEXT: st.b64 [%rd30], %rd13;
+; CHECK-NEXT: ld.b64 %rd31, [%SP+8];
+; CHECK-NEXT: ld.b64 %rd32, [%SP];
+; CHECK-NEXT: ld.b64 %rd33, [%SP+24];
+; CHECK-NEXT: ld.b64 %rd34, [%SP+16];
+; CHECK-NEXT: ld.b64 %rd35, [%SP+40];
+; CHECK-NEXT: ld.b64 %rd36, [%SP+32];
+; CHECK-NEXT: ld.b64 %rd37, [%SP+56];
+; CHECK-NEXT: ld.b64 %rd38, [%SP+48];
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1072], {%rd38, %rd37};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1056], {%rd36, %rd35};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1040], {%rd34, %rd33};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1024], {%rd32, %rd31};
+; CHECK-NEXT: ret;
+entry:
+ %offset.0 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 0
+ %element.0 = load double, ptr addrspace(3) %offset.0, align 64
+ %offset.1 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 8
+ %element.1 = load double, ptr addrspace(3) %offset.1, align 8
+ %offset.2 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 16
+ %element.2 = load double, ptr addrspace(3) %offset.2, align 8
+ %offset.3 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 24
+ %element.3 = load double, ptr addrspace(3) %offset.3, align 8
+ %vector.build0 = insertelement <8 x double> %vector, double %element.0, i32 %idx0
+ %vector.build1 = insertelement <8 x double> %vector.build0, double %element.1, i32 %idx1
+ %vector.build2 = insertelement <8 x double> %vector.build1, double %element.2, i32 %idx2
+ %vector.build3 = insertelement <8 x double> %vector.build2, double %element.3, i32 %idx3
+ %location = getelementptr i8, ptr addrspace(3) %shared.mem, i32 1024
+ store <8 x double> %vector.build3, ptr addrspace(3) %location, align 64
+ ret void
+}
+
+; COM: Spill the vector twice. Because these are in two different slots, the
+; resulting spill codes may be non-overlapping even though the insertelt
+; sequences overlap.
+define ptx_kernel void @spill_twice(ptr addrspace(3) %shared.mem, <8 x double> %vector, i32 %idx0, i32 %idx1, i32 %idx2, i32 %idx3) local_unnamed_addr {
+; CHECK-LABEL: spill_twice(
+; CHECK: {
+; CHECK-NEXT: .local .align 8 .b8 __local_depot1[128];
+; CHECK-NEXT: .reg .b64 %SP;
+; CHECK-NEXT: .reg .b64 %SPL;
+; CHECK-NEXT: .reg .b32 %r<7>;
+; CHECK-NEXT: .reg .b64 %rd<54>;
+; CHECK-EMPTY:
+; CHECK-NEXT: // %bb.0: // %entry
+; CHECK-NEXT: mov.b64 %SPL, __local_depot1;
+; CHECK-NEXT: cvta.local.u64 %SP, %SPL;
+; CHECK-NEXT: ld.param.b64 %rd1, [spill_twice_param_0];
+; CHECK-NEXT: ld.param.v2.b64 {%rd2, %rd3}, [spill_twice_param_1];
+; CHECK-NEXT: ld.param.v2.b64 {%rd4, %rd5}, [spill_twice_param_1+16];
+; CHECK-NEXT: ld.param.v2.b64 {%rd6, %rd7}, [spill_twice_param_1+32];
+; CHECK-NEXT: ld.param.v2.b64 {%rd8, %rd9}, [spill_twice_param_1+48];
+; CHECK-NEXT: ld.param.b32 %r1, [spill_twice_param_2];
+; CHECK-NEXT: ld.param.b32 %r2, [spill_twice_param_3];
+; CHECK-NEXT: ld.param.b32 %r3, [spill_twice_param_4];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd10, %rd11}, [%rd1];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd12, %rd13}, [%rd1+16];
+; CHECK-NEXT: mul.wide.u32 %rd14, %r1, 8;
+; CHECK-NEXT: and.b64 %rd15, %rd14, 56;
+; CHECK-NEXT: add.u64 %rd16, %SP, 0;
+; CHECK-NEXT: add.s64 %rd17, %rd16, %rd15;
+; CHECK-NEXT: shl.b32 %r4, %r1, 3;
+; CHECK-NEXT: cvt.u64.u32 %rd18, %r4;
+; CHECK-NEXT: and.b64 %rd19, %rd18, 56;
+; CHECK-NEXT: add.u64 %rd20, %SP, 64;
+; CHECK-NEXT: add.s64 %rd21, %rd20, %rd19;
+; CHECK-NEXT: mul.wide.u32 %rd22, %r2, 8;
+; CHECK-NEXT: and.b64 %rd23, %rd22, 56;
+; CHECK-NEXT: add.s64 %rd24, %rd16, %rd23;
+; CHECK-NEXT: shl.b32 %r5, %r2, 3;
+; CHECK-NEXT: cvt.u64.u32 %rd25, %r5;
+; CHECK-NEXT: and.b64 %rd26, %rd25, 56;
+; CHECK-NEXT: add.s64 %rd27, %rd20, %rd26;
+; CHECK-NEXT: st.b64 [%SP+120], %rd9;
+; CHECK-NEXT: st.b64 [%SP+112], %rd8;
+; CHECK-NEXT: st.b64 [%SP+104], %rd7;
+; CHECK-NEXT: st.b64 [%SP+96], %rd6;
+; CHECK-NEXT: st.b64 [%SP+88], %rd5;
+; CHECK-NEXT: st.b64 [%SP+80], %rd4;
+; CHECK-NEXT: st.b64 [%SP+72], %rd3;
+; CHECK-NEXT: st.b64 [%SP+64], %rd2;
+; CHECK-NEXT: st.b64 [%rd21], %rd10;
+; CHECK-NEXT: st.b64 [%rd27], %rd11;
+; CHECK-NEXT: shl.b32 %r6, %r3, 3;
+; CHECK-NEXT: cvt.u64.u32 %rd28, %r6;
+; CHECK-NEXT: and.b64 %rd29, %rd28, 56;
+; CHECK-NEXT: add.s64 %rd30, %rd20, %rd29;
+; CHECK-NEXT: st.b64 [%rd30], %rd12;
+; CHECK-NEXT: mul.wide.u32 %rd31, %r3, 8;
+; CHECK-NEXT: and.b64 %rd32, %rd31, 56;
+; CHECK-NEXT: add.s64 %rd33, %rd16, %rd32;
+; CHECK-NEXT: ld.b64 %rd34, [%SP+72];
+; CHECK-NEXT: ld.b64 %rd35, [%SP+64];
+; CHECK-NEXT: ld.b64 %rd36, [%SP+88];
+; CHECK-NEXT: ld.b64 %rd37, [%SP+80];
+; CHECK-NEXT: ld.b64 %rd38, [%SP+104];
+; CHECK-NEXT: ld.b64 %rd39, [%SP+96];
+; CHECK-NEXT: ld.b64 %rd40, [%SP+120];
+; CHECK-NEXT: ld.b64 %rd41, [%SP+112];
+; CHECK-NEXT: st.b64 [%SP+56], %rd9;
+; CHECK-NEXT: st.b64 [%SP+48], %rd8;
+; CHECK-NEXT: st.b64 [%SP+40], %rd7;
+; CHECK-NEXT: st.b64 [%SP+32], %rd6;
+; CHECK-NEXT: st.b64 [%SP+24], %rd5;
+; CHECK-NEXT: st.b64 [%SP+16], %rd4;
+; CHECK-NEXT: st.b64 [%SP+8], %rd3;
+; CHECK-NEXT: st.b64 [%SP], %rd2;
+; CHECK-NEXT: st.b64 [%rd17], %rd10;
+; CHECK-NEXT: st.b64 [%rd24], %rd11;
+; CHECK-NEXT: st.b64 [%rd33], %rd12;
+; CHECK-NEXT: ld.param.b32 %rd42, [spill_twice_param_5];
+; CHECK-NEXT: shl.b64 %rd43, %rd42, 3;
+; CHECK-NEXT: and.b64 %rd44, %rd43, 56;
+; CHECK-NEXT: add.s64 %rd45, %rd16, %rd44;
+; CHECK-NEXT: st.b64 [%rd45], %rd13;
+; CHECK-NEXT: ld.b64 %rd46, [%SP+8];
+; CHECK-NEXT: ld.b64 %rd47, [%SP];
+; CHECK-NEXT: ld.b64 %rd48, [%SP+24];
+; CHECK-NEXT: ld.b64 %rd49, [%SP+16];
+; CHECK-NEXT: ld.b64 %rd50, [%SP+40];
+; CHECK-NEXT: ld.b64 %rd51, [%SP+32];
+; CHECK-NEXT: ld.b64 %rd52, [%SP+56];
+; CHECK-NEXT: ld.b64 %rd53, [%SP+48];
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1072], {%rd53, %rd52};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1056], {%rd51, %rd50};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1040], {%rd49, %rd48};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1024], {%rd47, %rd46};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1144], {%rd41, %rd40};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1128], {%rd39, %rd38};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1112], {%rd37, %rd36};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1096], {%rd35, %rd34};
+; CHECK-NEXT: ret;
+entry:
+ %offset.0 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 0
+ %element.0 = load double, ptr addrspace(3) %offset.0, align 64
+ %offset.1 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 8
+ %element.1 = load double, ptr addrspace(3) %offset.1, align 8
+ %offset.2 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 16
+ %element.2 = load double, ptr addrspace(3) %offset.2, align 8
+ %offset.3 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 24
+ %element.3 = load double, ptr addrspace(3) %offset.3, align 8
+
+; COM: begin chain 1
+ %vector.build0 = insertelement <8 x double> %vector, double %element.0, i32 %idx0
+ %vector.build1 = insertelement <8 x double> %vector.build0, double %element.1, i32 %idx1
+
+; COM: interleave a second chain of insertelements
+ %vector.build1-2 = insertelement <8 x double> %vector.build1, double %element.2, i32 %idx2
+
+; COM: continue chain 1
+ %vector.build2 = insertelement <8 x double> %vector.build1, double %element.2, i32 %idx2
+ %vector.build3 = insertelement <8 x double> %vector.build2, double %element.3, i32 %idx3
+
+; COM: save chain 1
+ %location = getelementptr i8, ptr addrspace(3) %shared.mem, i32 1024
+ store <8 x double> %vector.build3, ptr addrspace(3) %location, align 64
+
+; COM: save chain 2
+ %location-2 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 1096
+ store <8 x double> %vector.build1-2, ptr addrspace(3) %location-2, align 64
+ ret void
+}
|
✅ With the latest revision this PR passed the C/C++ code formatter. |
This change improves ISel compile time on a vector-heavy kernel from 1 hour to 200s. I've included a reduced test case in NVPTX, although this is a target-independent change. |
028c586
to
b67c59f
Compare
b67c59f
to
74eb28a
Compare
74eb28a
to
26e9079
Compare
For an insertelt with dynamic indices, the default handling in DAGTypeLegalizer and LegalizeDAG will reserve a stack slot for the vector, lower the insertelt to a store, then load the modified vector back into temporaries. The vector store and load may be legalized into a sequence of smaller operations depending on the target. Let V = the vector size and L = the length of a chain of insertelts with dynamic indices. In the worse case, this chain will lower to O(VL) operations, which can increase code size dramatically. Instead, identify such chains, reserve one stack slot for the vector, and lower all of the insertelts to stores at once. This requires only O(V + L) operations. This change only affects the default lowering behavior and still leaves targets to do their own thing.
26e9079
to
60adde5
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
target triple = "nvptx64-nvidia-cuda" | ||
|
||
; COM: Save the vector to the stack once. | ||
define ptx_kernel void @lower_once(ptr addrspace(3) %shared.mem, <8 x double> %vector, i32 %idx0, i32 %idx1, i32 %idx2, i32 %idx3) local_unnamed_addr { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Those are fairly large test cases. Can they be further reduced?
- Do we need to operate on double? Using smaller type may reduce the number of loads/stores we may need to do.
- Can we use literal element values? Then we would not need to generate load instructions.
- In some cases we can reduce vector size. We do not need 8-element vector for something that modifies just one element.
%vector.build0 = insertelement <8 x double> %vector, double %element.0, i32 0 | ||
%vector.build1 = insertelement <8 x double> %vector.build0, double %element.1, i32 1 | ||
%vector.build2 = insertelement <8 x double> %vector.build1, double %element.2, i32 2 | ||
%vector.build3 = insertelement <8 x double> %vector.build2, double %element.3, i32 %idx0 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What's the general strategy for testing these changes? I see some complex cases tested, but the basic ones appear to be absent. While making sure that we can handle nontrivial use cases is great, having small very specific test cases is also important -- when things break, the less moving parts we have in the failing test, the better.
For the new functionality like this I would expect to see a somewhat more systemic approach covering individual corner cases, before testing the combination of them. It makes debugging/troubleshooting easier.
E.g.:
- individual tests for dynamic insertion:
- at the beginning of the chain
- at the end of the chain
- in the middle of the chain
- repeated insertion with the same dynamic index
Then build more complicated tests with multiple dynamic indices, conditional insertion, etc.
; RUN: %if ptxas %{ llc < %s -mcpu=sm_20 | %ptxas-verify %} | ||
target triple = "nvptx64-nvidia-cuda" | ||
|
||
; COM: Save the vector to the stack once. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
You don't need to write COM
on a regular comment. COM is just for commenting out what would otherwise be FileCheck directive.
For an insertelt with a dynamic index, the default handling in DAGTypeLegalizer and LegalizeDAG will reserve a stack slot for the vector, lower the insertelt to a store, then load the modified vector back into temporaries. The vector store and load may be legalized into a sequence of smaller operations depending on the target.
Let V = the vector size and L = the length of a chain of insertelts with dynamic indices. In the worse case, this chain will lower to O(VL) operations, which can increase code size dramatically.
Instead, identify such chains, reserve one stack slot for the vector, and lower all of the insertelts to stores at once. This requires only O(V + L) operations. This change only affects the default lowering behavior and still leaves targets to do their own thing.