diff --git a/llvm/lib/Transforms/Scalar/LowerMatrixIntrinsics.cpp b/llvm/lib/Transforms/Scalar/LowerMatrixIntrinsics.cpp index 08c6406f0a30e..17594b98c5bcb 100644 --- a/llvm/lib/Transforms/Scalar/LowerMatrixIntrinsics.cpp +++ b/llvm/lib/Transforms/Scalar/LowerMatrixIntrinsics.cpp @@ -93,6 +93,19 @@ static DISubprogram *getSubprogram(DIScope *Scope) { return cast(Scope)->getSubprogram(); } +/// Erase \p V from \p BB and move \II forward to avoid invalidating +/// iterators. +static void eraseFromParentAndMove(Value *V, BasicBlock::reverse_iterator &II, + BasicBlock &BB) { + auto *Inst = cast(V); + // Still used, don't erase. + if (!Inst->use_empty()) + return; + if (II != BB.rend() && Inst == &*II) + ++II; + Inst->eraseFromParent(); +} + /// Return true if V is a splat of a value (which is used when multiplying a /// matrix with a scalar). static bool isSplat(Value *V) { @@ -107,6 +120,12 @@ auto m_AnyMul(const LTy &L, const RTy &R) { return m_CombineOr(m_Mul(L, R), m_FMul(L, R)); } +/// Match any add operation (fp or integer). +template +auto m_AnyAdd(const LTy &L, const RTy &R) { + return m_CombineOr(m_Add(L, R), m_FAdd(L, R)); +} + namespace { // Given an element pointer \p BasePtr to the start of a (sub) matrix, compute @@ -725,135 +744,179 @@ class LowerMatrixIntrinsics { return Operation(T0, Shape0.t(), T1, Shape1.t()); } - /// Try moving transposes in order to fold them away or into multiplies. - void optimizeTransposes() { - auto ReplaceAllUsesWith = [this](Instruction &Old, Value *New) { - // We need to remove Old from the ShapeMap otherwise RAUW will replace it - // with New. We should only add New it it supportsShapeInfo so we insert - // it conditionally instead. - auto S = ShapeMap.find(&Old); - if (S != ShapeMap.end()) { - ShapeMap.erase(S); - if (supportsShapeInfo(New)) - ShapeMap.insert({New, S->second}); - } - Old.replaceAllUsesWith(New); + void updateShapeAndReplaceAllUsesWith(Instruction &Old, Value *New) { + // We need to remove Old from the ShapeMap otherwise RAUW will replace it + // with New. We should only add New it it supportsShapeInfo so we insert + // it conditionally instead. + auto S = ShapeMap.find(&Old); + if (S != ShapeMap.end()) { + ShapeMap.erase(S); + if (supportsShapeInfo(New)) + ShapeMap.insert({New, S->second}); + } + Old.replaceAllUsesWith(New); + } + + /// Sink a top-level transpose inside matmuls and adds. + /// This creates and erases instructions as needed, and returns the newly + /// created instruction while updating the iterator to avoid invalidation. If + /// this returns nullptr, no new instruction was created. + Instruction *sinkTranspose(Instruction &I, BasicBlock::reverse_iterator &II) { + BasicBlock &BB = *I.getParent(); + IRBuilder<> IB(&I); + MatrixBuilder Builder(IB); + + Value *TA, *TAMA, *TAMB; + ConstantInt *R, *K, *C; + if (!match(&I, m_Intrinsic( + m_Value(TA), m_ConstantInt(R), m_ConstantInt(C)))) + return nullptr; + + // Transpose of a transpose is a nop + Value *TATA; + if (match(TA, m_Intrinsic(m_Value(TATA)))) { + updateShapeAndReplaceAllUsesWith(I, TATA); + eraseFromParentAndMove(&I, II, BB); + eraseFromParentAndMove(TA, II, BB); + return nullptr; + } + + // k^T -> k + if (isSplat(TA)) { + updateShapeAndReplaceAllUsesWith(I, TA); + eraseFromParentAndMove(&I, II, BB); + return nullptr; + } + + // (A * B)^t -> B^t * A^t + // RxK KxC CxK KxR + if (match(TA, m_Intrinsic( + m_Value(TAMA), m_Value(TAMB), m_ConstantInt(R), + m_ConstantInt(K), m_ConstantInt(C)))) { + auto NewInst = distributeTransposes( + TAMB, {K, C}, TAMA, {R, K}, Builder, + [&](Value *T0, ShapeInfo Shape0, Value *T1, ShapeInfo Shape1) { + return Builder.CreateMatrixMultiply(T0, T1, Shape0.NumRows, + Shape0.NumColumns, + Shape1.NumColumns, "mmul"); + }); + updateShapeAndReplaceAllUsesWith(I, NewInst); + eraseFromParentAndMove(&I, II, BB); + eraseFromParentAndMove(TA, II, BB); + return NewInst; + } + + // Same as above, but with a mul, which occurs when multiplied + // with a scalar. + // (A * k)^t -> A^t * k + // R x C RxC + if (match(TA, m_AnyMul(m_Value(TAMA), m_Value(TAMB))) && + (isSplat(TAMA) || isSplat(TAMB))) { + IRBuilder<> LocalBuilder(&I); + // We know that the transposed operand is of shape RxC. + // An when multiplied with a scalar, the shape is preserved. + auto NewInst = distributeTransposes( + TAMA, {R, C}, TAMB, {R, C}, Builder, + [&](Value *T0, ShapeInfo Shape0, Value *T1, ShapeInfo Shape1) { + bool IsFP = I.getType()->isFPOrFPVectorTy(); + auto *Mul = IsFP ? LocalBuilder.CreateFMul(T0, T1, "mmul") + : LocalBuilder.CreateMul(T0, T1, "mmul"); + auto *Result = cast(Mul); + setShapeInfo(Result, Shape0); + return Result; + }); + updateShapeAndReplaceAllUsesWith(I, NewInst); + eraseFromParentAndMove(&I, II, BB); + eraseFromParentAndMove(TA, II, BB); + return NewInst; + } + + // (A + B)^t -> A^t + B^t + // RxC RxC CxR CxR + if (match(TA, m_AnyAdd(m_Value(TAMA), m_Value(TAMB)))) { + IRBuilder<> LocalBuilder(&I); + auto NewInst = distributeTransposes( + TAMA, {R, C}, TAMB, {R, C}, Builder, + [&](Value *T0, ShapeInfo Shape0, Value *T1, ShapeInfo Shape1) { + auto *FAdd = + cast(LocalBuilder.CreateFAdd(T0, T1, "mfadd")); + setShapeInfo(FAdd, Shape0); + return FAdd; + }); + updateShapeAndReplaceAllUsesWith(I, NewInst); + eraseFromParentAndMove(&I, II, BB); + eraseFromParentAndMove(TA, II, BB); + return NewInst; + } + + return nullptr; + } + + void liftTranspose(Instruction &I) { + // Erase dead Instructions after lifting transposes from binops. + auto CleanupBinOp = [](Instruction &T, Value *A, Value *B) { + if (T.use_empty()) + T.eraseFromParent(); + if (A->use_empty()) + cast(A)->eraseFromParent(); + if (A != B && B->use_empty()) + cast(B)->eraseFromParent(); }; - // First sink all transposes inside matmuls, hoping that we end up with NN, - // NT or TN variants. + Value *A, *B, *AT, *BT; + ConstantInt *R, *K, *C; + // A^t * B ^t -> (B * A)^t + if (match(&I, m_Intrinsic( + m_Value(A), m_Value(B), m_ConstantInt(R), + m_ConstantInt(K), m_ConstantInt(C))) && + match(A, m_Intrinsic(m_Value(AT))) && + match(B, m_Intrinsic(m_Value((BT))))) { + IRBuilder<> IB(&I); + MatrixBuilder Builder(IB); + Value *M = Builder.CreateMatrixMultiply( + BT, AT, C->getZExtValue(), K->getZExtValue(), R->getZExtValue()); + setShapeInfo(M, {C, R}); + Instruction *NewInst = Builder.CreateMatrixTranspose(M, C->getZExtValue(), + R->getZExtValue()); + updateShapeAndReplaceAllUsesWith(I, NewInst); + CleanupBinOp(I, A, B); + } + // A^t + B ^t -> (A + B)^t + else if (match(&I, m_FAdd(m_Value(A), m_Value(B))) && + match(A, m_Intrinsic( + m_Value(AT), m_ConstantInt(R), m_ConstantInt(C))) && + match(B, m_Intrinsic( + m_Value(BT), m_ConstantInt(R), m_ConstantInt(C)))) { + IRBuilder<> Builder(&I); + Value *Add = cast(Builder.CreateFAdd(AT, BT, "mfadd")); + setShapeInfo(Add, {C, R}); + MatrixBuilder MBuilder(Builder); + Instruction *NewInst = MBuilder.CreateMatrixTranspose( + Add, C->getZExtValue(), R->getZExtValue(), "mfadd_t"); + updateShapeAndReplaceAllUsesWith(I, NewInst); + CleanupBinOp(I, A, B); + } + } + + /// Try moving transposes in order to fold them away or into multiplies. + void optimizeTransposes() { + // First sink all transposes inside matmuls and adds, hoping that we end up + // with NN, NT or TN variants. for (BasicBlock &BB : reverse(Func)) { for (auto II = BB.rbegin(); II != BB.rend();) { Instruction &I = *II; // We may remove II. By default continue on the next/prev instruction. ++II; - // If we were to erase II, move again. - auto EraseFromParent = [&II, &BB](Value *V) { - auto *Inst = cast(V); - if (Inst->use_empty()) { - if (II != BB.rend() && Inst == &*II) { - ++II; - } - Inst->eraseFromParent(); - } - }; - - // If we're creating a new instruction, continue from there. - Instruction *NewInst = nullptr; - - IRBuilder<> IB(&I); - MatrixBuilder Builder(IB); - - Value *TA, *TAMA, *TAMB; - ConstantInt *R, *K, *C; - if (match(&I, m_Intrinsic( - m_Value(TA), m_ConstantInt(R), m_ConstantInt(C)))) { - // Transpose of a transpose is a nop - Value *TATA; - if (match(TA, - m_Intrinsic(m_Value(TATA)))) { - ReplaceAllUsesWith(I, TATA); - EraseFromParent(&I); - EraseFromParent(TA); - } - // k^T -> k - else if (isSplat(TA)) { - ReplaceAllUsesWith(I, TA); - EraseFromParent(&I); - } - // (A * B)^t -> B^t * A^t - // RxK KxC CxK KxR - else if (match(TA, m_Intrinsic( - m_Value(TAMA), m_Value(TAMB), m_ConstantInt(R), - m_ConstantInt(K), m_ConstantInt(C)))) { - NewInst = distributeTransposes( - TAMB, {K, C}, TAMA, {R, K}, Builder, - [&](Value *T0, ShapeInfo Shape0, Value *T1, ShapeInfo Shape1) { - return Builder.CreateMatrixMultiply( - T0, T1, Shape0.NumRows, Shape0.NumColumns, - Shape1.NumColumns, "mmul"); - }); - ReplaceAllUsesWith(I, NewInst); - EraseFromParent(&I); - EraseFromParent(TA); - // Same as above, but with a mul, which occurs when multiplied - // with a scalar. - // (A * k)^t -> A^t * k - // R x C RxC - } else if (match(TA, m_AnyMul(m_Value(TAMA), m_Value(TAMB))) && - (isSplat(TAMA) || isSplat(TAMB))) { - IRBuilder<> LocalBuilder(&I); - // We know that the transposed operand is of shape RxC. - // An when multiplied with a scalar, the shape is preserved. - NewInst = distributeTransposes( - TAMA, {R, C}, TAMB, {R, C}, Builder, - [&](Value *T0, ShapeInfo Shape0, Value *T1, ShapeInfo Shape1) { - bool IsFP = I.getType()->isFPOrFPVectorTy(); - auto *Mul = IsFP ? LocalBuilder.CreateFMul(T0, T1, "mmul") - : LocalBuilder.CreateMul(T0, T1, "mmul"); - auto *Result = cast(Mul); - setShapeInfo(Result, Shape0); - return Result; - }); - ReplaceAllUsesWith(I, NewInst); - EraseFromParent(&I); - EraseFromParent(TA); - } - } - - // If we replaced I with a new instruction, continue from there. - if (NewInst) + if (Instruction *NewInst = sinkTranspose(I, II)) II = std::next(BasicBlock::reverse_iterator(NewInst)); } } - // If we have a TT matmul, lift the transpose. We may be able to fold into - // consuming multiply. + // If we have a TT matmul or a TT add, lift the transpose. We may be able + // to fold into consuming multiply or add. for (BasicBlock &BB : Func) { for (Instruction &I : llvm::make_early_inc_range(BB)) { - Value *A, *B, *AT, *BT; - ConstantInt *R, *K, *C; - // A^t * B ^t -> (B * A)^t - if (match(&I, m_Intrinsic( - m_Value(A), m_Value(B), m_ConstantInt(R), - m_ConstantInt(K), m_ConstantInt(C))) && - match(A, m_Intrinsic(m_Value(AT))) && - match(B, m_Intrinsic(m_Value((BT))))) { - IRBuilder<> IB(&I); - MatrixBuilder Builder(IB); - Value *M = Builder.CreateMatrixMultiply( - BT, AT, C->getZExtValue(), K->getZExtValue(), R->getZExtValue()); - setShapeInfo(M, {C, R}); - Instruction *NewInst = Builder.CreateMatrixTranspose( - M, C->getZExtValue(), R->getZExtValue()); - ReplaceAllUsesWith(I, NewInst); - if (I.use_empty()) - I.eraseFromParent(); - if (A->use_empty()) - cast(A)->eraseFromParent(); - if (A != B && B->use_empty()) - cast(B)->eraseFromParent(); - } + liftTranspose(I); } } } diff --git a/llvm/test/Transforms/LowerMatrixIntrinsics/after-transpose-opts.ll b/llvm/test/Transforms/LowerMatrixIntrinsics/after-transpose-opts.ll index 9f1670360191e..7510f43d31986 100644 --- a/llvm/test/Transforms/LowerMatrixIntrinsics/after-transpose-opts.ll +++ b/llvm/test/Transforms/LowerMatrixIntrinsics/after-transpose-opts.ll @@ -100,10 +100,9 @@ define void @at_plus_bt(ptr %Aptr, ptr %Bptr, ptr %C) { ; CHECK-NEXT: entry: ; CHECK-NEXT: [[A:%.*]] = load <9 x double>, ptr [[APTR:%.*]], align 128 ; CHECK-NEXT: [[B:%.*]] = load <9 x double>, ptr [[BPTR:%.*]], align 128 -; CHECK-NEXT: [[AT:%.*]] = call <9 x double> @llvm.matrix.transpose.v9f64(<9 x double> [[A]], i32 3, i32 3) -; CHECK-NEXT: [[BT:%.*]] = call <9 x double> @llvm.matrix.transpose.v9f64(<9 x double> [[B]], i32 3, i32 3) -; CHECK-NEXT: [[FADD:%.*]] = fadd <9 x double> [[AT]], [[BT]] -; CHECK-NEXT: store <9 x double> [[FADD]], ptr [[C:%.*]], align 128 +; CHECK-NEXT: [[MFADD:%.*]] = fadd <9 x double> [[A]], [[B]] +; CHECK-NEXT: [[MFADD_T:%.*]] = call <9 x double> @llvm.matrix.transpose.v9f64(<9 x double> [[MFADD]], i32 3, i32 3) +; CHECK-NEXT: store <9 x double> [[MFADD_T]], ptr [[C:%.*]], align 128 ; CHECK-NEXT: ret void ; entry: @@ -122,9 +121,9 @@ define void @a_plus_b_t(ptr %Aptr, ptr %Bptr, ptr %C) { ; CHECK-NEXT: entry: ; CHECK-NEXT: [[A:%.*]] = load <9 x double>, ptr [[APTR:%.*]], align 128 ; CHECK-NEXT: [[B:%.*]] = load <9 x double>, ptr [[BPTR:%.*]], align 128 -; CHECK-NEXT: [[FADD:%.*]] = fadd <9 x double> [[A]], [[B]] -; CHECK-NEXT: [[T:%.*]] = call <9 x double> @llvm.matrix.transpose.v9f64(<9 x double> [[FADD]], i32 3, i32 3) -; CHECK-NEXT: store <9 x double> [[T]], ptr [[C:%.*]], align 128 +; CHECK-NEXT: [[MFADD1:%.*]] = fadd <9 x double> [[A]], [[B]] +; CHECK-NEXT: [[MFADD_T:%.*]] = call <9 x double> @llvm.matrix.transpose.v9f64(<9 x double> [[MFADD1]], i32 3, i32 3) +; CHECK-NEXT: store <9 x double> [[MFADD_T]], ptr [[C:%.*]], align 128 ; CHECK-NEXT: ret void ; entry: @@ -145,11 +144,10 @@ define void @atbt_plus_ctdt(ptr %Aptr, ptr %Bptr, ptr %Cptr, ptr %Dptr, ptr %E) ; CHECK-NEXT: [[C:%.*]] = load <9 x double>, ptr [[CPTR:%.*]], align 128 ; CHECK-NEXT: [[D:%.*]] = load <9 x double>, ptr [[DPTR:%.*]], align 128 ; CHECK-NEXT: [[TMP0:%.*]] = call <9 x double> @llvm.matrix.multiply.v9f64.v9f64.v9f64(<9 x double> [[B]], <9 x double> [[A]], i32 3, i32 3, i32 3) -; CHECK-NEXT: [[TMP1:%.*]] = call <9 x double> @llvm.matrix.transpose.v9f64(<9 x double> [[TMP0]], i32 3, i32 3) -; CHECK-NEXT: [[TMP2:%.*]] = call <9 x double> @llvm.matrix.multiply.v9f64.v9f64.v9f64(<9 x double> [[D]], <9 x double> [[C]], i32 3, i32 3, i32 3) -; CHECK-NEXT: [[TMP3:%.*]] = call <9 x double> @llvm.matrix.transpose.v9f64(<9 x double> [[TMP2]], i32 3, i32 3) -; CHECK-NEXT: [[FADD:%.*]] = fadd <9 x double> [[TMP1]], [[TMP3]] -; CHECK-NEXT: store <9 x double> [[FADD]], ptr [[E:%.*]], align 128 +; CHECK-NEXT: [[TMP1:%.*]] = call <9 x double> @llvm.matrix.multiply.v9f64.v9f64.v9f64(<9 x double> [[D]], <9 x double> [[C]], i32 3, i32 3, i32 3) +; CHECK-NEXT: [[MFADD:%.*]] = fadd <9 x double> [[TMP0]], [[TMP1]] +; CHECK-NEXT: [[MFADD_T:%.*]] = call <9 x double> @llvm.matrix.transpose.v9f64(<9 x double> [[MFADD]], i32 3, i32 3) +; CHECK-NEXT: store <9 x double> [[MFADD_T]], ptr [[E:%.*]], align 128 ; CHECK-NEXT: ret void ; entry: @@ -200,17 +198,13 @@ define void @atbt_plus_kctdt_t(ptr %Aptr, ptr %Bptr, ptr %Cptr, ptr %Dptr, doubl ; CHECK-NEXT: [[B:%.*]] = load <9 x double>, ptr [[BPTR:%.*]], align 128 ; CHECK-NEXT: [[C:%.*]] = load <9 x double>, ptr [[CPTR:%.*]], align 128 ; CHECK-NEXT: [[D:%.*]] = load <9 x double>, ptr [[DPTR:%.*]], align 128 -; CHECK-NEXT: [[CT:%.*]] = call <9 x double> @llvm.matrix.transpose.v9f64(<9 x double> [[C]], i32 3, i32 3) -; CHECK-NEXT: [[DT:%.*]] = call <9 x double> @llvm.matrix.transpose.v9f64(<9 x double> [[D]], i32 3, i32 3) -; CHECK-NEXT: [[TMP0:%.*]] = call <9 x double> @llvm.matrix.multiply.v9f64.v9f64.v9f64(<9 x double> [[B]], <9 x double> [[A]], i32 3, i32 3, i32 3) -; CHECK-NEXT: [[TMP1:%.*]] = call <9 x double> @llvm.matrix.transpose.v9f64(<9 x double> [[TMP0]], i32 3, i32 3) ; CHECK-NEXT: [[VECK:%.*]] = insertelement <9 x double> poison, double [[K:%.*]], i64 0 ; CHECK-NEXT: [[SPLAT:%.*]] = shufflevector <9 x double> [[VECK]], <9 x double> poison, <9 x i32> zeroinitializer -; CHECK-NEXT: [[KCT:%.*]] = call <9 x double> @llvm.matrix.multiply.v9f64.v9f64.v9f64(<9 x double> [[SPLAT]], <9 x double> [[CT]], i32 3, i32 3, i32 3) -; CHECK-NEXT: [[KCTDT:%.*]] = call <9 x double> @llvm.matrix.multiply.v9f64.v9f64.v9f64(<9 x double> [[KCT]], <9 x double> [[DT]], i32 3, i32 3, i32 3) -; CHECK-NEXT: [[FADD:%.*]] = fadd <9 x double> [[TMP1]], [[KCTDT]] -; CHECK-NEXT: [[T:%.*]] = call <9 x double> @llvm.matrix.transpose.v9f64(<9 x double> [[FADD]], i32 3, i32 3) -; CHECK-NEXT: store <9 x double> [[T]], ptr [[E:%.*]], align 128 +; CHECK-NEXT: [[MMUL2:%.*]] = call <9 x double> @llvm.matrix.multiply.v9f64.v9f64.v9f64(<9 x double> [[B]], <9 x double> [[A]], i32 3, i32 3, i32 3) +; CHECK-NEXT: [[MMUL1:%.*]] = call <9 x double> @llvm.matrix.multiply.v9f64.v9f64.v9f64(<9 x double> [[C]], <9 x double> [[SPLAT]], i32 3, i32 3, i32 3) +; CHECK-NEXT: [[MMUL:%.*]] = call <9 x double> @llvm.matrix.multiply.v9f64.v9f64.v9f64(<9 x double> [[D]], <9 x double> [[MMUL1]], i32 3, i32 3, i32 3) +; CHECK-NEXT: [[MFADD:%.*]] = fadd <9 x double> [[MMUL2]], [[MMUL]] +; CHECK-NEXT: store <9 x double> [[MFADD]], ptr [[E:%.*]], align 128 ; CHECK-NEXT: ret void ; entry: diff --git a/llvm/test/Transforms/LowerMatrixIntrinsics/propagate-backward.ll b/llvm/test/Transforms/LowerMatrixIntrinsics/propagate-backward.ll index 33a338dbc4ea0..82ae93b31035d 100644 --- a/llvm/test/Transforms/LowerMatrixIntrinsics/propagate-backward.ll +++ b/llvm/test/Transforms/LowerMatrixIntrinsics/propagate-backward.ll @@ -4,35 +4,31 @@ define <8 x double> @fadd_transpose(<8 x double> %a, <8 x double> %b) { ; CHECK-LABEL: @fadd_transpose( ; CHECK-NEXT: entry: -; CHECK-NEXT: [[SPLIT:%.*]] = shufflevector <8 x double> [[A:%.*]], <8 x double> poison, <2 x i32> -; CHECK-NEXT: [[SPLIT1:%.*]] = shufflevector <8 x double> [[A]], <8 x double> poison, <2 x i32> -; CHECK-NEXT: [[SPLIT2:%.*]] = shufflevector <8 x double> [[A]], <8 x double> poison, <2 x i32> -; CHECK-NEXT: [[SPLIT3:%.*]] = shufflevector <8 x double> [[A]], <8 x double> poison, <2 x i32> -; CHECK-NEXT: [[SPLIT4:%.*]] = shufflevector <8 x double> [[B:%.*]], <8 x double> poison, <2 x i32> -; CHECK-NEXT: [[SPLIT5:%.*]] = shufflevector <8 x double> [[B]], <8 x double> poison, <2 x i32> -; CHECK-NEXT: [[SPLIT6:%.*]] = shufflevector <8 x double> [[B]], <8 x double> poison, <2 x i32> -; CHECK-NEXT: [[SPLIT7:%.*]] = shufflevector <8 x double> [[B]], <8 x double> poison, <2 x i32> -; CHECK-NEXT: [[TMP0:%.*]] = fadd <2 x double> [[SPLIT]], [[SPLIT4]] -; CHECK-NEXT: [[TMP1:%.*]] = fadd <2 x double> [[SPLIT1]], [[SPLIT5]] -; CHECK-NEXT: [[TMP2:%.*]] = fadd <2 x double> [[SPLIT2]], [[SPLIT6]] -; CHECK-NEXT: [[TMP3:%.*]] = fadd <2 x double> [[SPLIT3]], [[SPLIT7]] -; CHECK-NEXT: [[TMP4:%.*]] = extractelement <2 x double> [[TMP0]], i64 0 -; CHECK-NEXT: [[TMP5:%.*]] = insertelement <4 x double> poison, double [[TMP4]], i64 0 -; CHECK-NEXT: [[TMP6:%.*]] = extractelement <2 x double> [[TMP1]], i64 0 -; CHECK-NEXT: [[TMP7:%.*]] = insertelement <4 x double> [[TMP5]], double [[TMP6]], i64 1 -; CHECK-NEXT: [[TMP8:%.*]] = extractelement <2 x double> [[TMP2]], i64 0 -; CHECK-NEXT: [[TMP9:%.*]] = insertelement <4 x double> [[TMP7]], double [[TMP8]], i64 2 -; CHECK-NEXT: [[TMP10:%.*]] = extractelement <2 x double> [[TMP3]], i64 0 -; CHECK-NEXT: [[TMP11:%.*]] = insertelement <4 x double> [[TMP9]], double [[TMP10]], i64 3 -; CHECK-NEXT: [[TMP12:%.*]] = extractelement <2 x double> [[TMP0]], i64 1 -; CHECK-NEXT: [[TMP13:%.*]] = insertelement <4 x double> poison, double [[TMP12]], i64 0 -; CHECK-NEXT: [[TMP14:%.*]] = extractelement <2 x double> [[TMP1]], i64 1 -; CHECK-NEXT: [[TMP15:%.*]] = insertelement <4 x double> [[TMP13]], double [[TMP14]], i64 1 -; CHECK-NEXT: [[TMP16:%.*]] = extractelement <2 x double> [[TMP2]], i64 1 -; CHECK-NEXT: [[TMP17:%.*]] = insertelement <4 x double> [[TMP15]], double [[TMP16]], i64 2 -; CHECK-NEXT: [[TMP18:%.*]] = extractelement <2 x double> [[TMP3]], i64 1 -; CHECK-NEXT: [[TMP19:%.*]] = insertelement <4 x double> [[TMP17]], double [[TMP18]], i64 3 -; CHECK-NEXT: [[TMP20:%.*]] = shufflevector <4 x double> [[TMP11]], <4 x double> [[TMP19]], <8 x i32> +; CHECK-NEXT: [[SPLIT:%.*]] = shufflevector <8 x double> [[A:%.*]], <8 x double> poison, <4 x i32> +; CHECK-NEXT: [[SPLIT2:%.*]] = shufflevector <8 x double> [[A]], <8 x double> poison, <4 x i32> +; CHECK-NEXT: [[SPLIT3:%.*]] = shufflevector <8 x double> [[B:%.*]], <8 x double> poison, <4 x i32> +; CHECK-NEXT: [[SPLIT4:%.*]] = shufflevector <8 x double> [[B]], <8 x double> poison, <4 x i32> +; CHECK-NEXT: [[TMP0:%.*]] = fadd <4 x double> [[SPLIT]], [[SPLIT3]] +; CHECK-NEXT: [[TMP1:%.*]] = fadd <4 x double> [[SPLIT2]], [[SPLIT4]] +; CHECK-NEXT: [[TMP2:%.*]] = extractelement <4 x double> [[TMP0]], i64 0 +; CHECK-NEXT: [[TMP3:%.*]] = insertelement <2 x double> poison, double [[TMP2]], i64 0 +; CHECK-NEXT: [[TMP4:%.*]] = extractelement <4 x double> [[TMP1]], i64 0 +; CHECK-NEXT: [[TMP5:%.*]] = insertelement <2 x double> [[TMP3]], double [[TMP4]], i64 1 +; CHECK-NEXT: [[TMP6:%.*]] = extractelement <4 x double> [[TMP0]], i64 1 +; CHECK-NEXT: [[TMP7:%.*]] = insertelement <2 x double> poison, double [[TMP6]], i64 0 +; CHECK-NEXT: [[TMP8:%.*]] = extractelement <4 x double> [[TMP1]], i64 1 +; CHECK-NEXT: [[TMP9:%.*]] = insertelement <2 x double> [[TMP7]], double [[TMP8]], i64 1 +; CHECK-NEXT: [[TMP10:%.*]] = extractelement <4 x double> [[TMP0]], i64 2 +; CHECK-NEXT: [[TMP11:%.*]] = insertelement <2 x double> poison, double [[TMP10]], i64 0 +; CHECK-NEXT: [[TMP12:%.*]] = extractelement <4 x double> [[TMP1]], i64 2 +; CHECK-NEXT: [[TMP13:%.*]] = insertelement <2 x double> [[TMP11]], double [[TMP12]], i64 1 +; CHECK-NEXT: [[TMP14:%.*]] = extractelement <4 x double> [[TMP0]], i64 3 +; CHECK-NEXT: [[TMP15:%.*]] = insertelement <2 x double> poison, double [[TMP14]], i64 0 +; CHECK-NEXT: [[TMP16:%.*]] = extractelement <4 x double> [[TMP1]], i64 3 +; CHECK-NEXT: [[TMP17:%.*]] = insertelement <2 x double> [[TMP15]], double [[TMP16]], i64 1 +; CHECK-NEXT: [[TMP18:%.*]] = shufflevector <2 x double> [[TMP5]], <2 x double> [[TMP9]], <4 x i32> +; CHECK-NEXT: [[TMP19:%.*]] = shufflevector <2 x double> [[TMP13]], <2 x double> [[TMP17]], <4 x i32> +; CHECK-NEXT: [[TMP20:%.*]] = shufflevector <4 x double> [[TMP18]], <4 x double> [[TMP19]], <8 x i32> ; CHECK-NEXT: ret <8 x double> [[TMP20]] ; entry: @@ -46,37 +42,40 @@ define <8 x double> @load_fadd_transpose(ptr %A.Ptr, <8 x double> %b) { ; CHECK-NEXT: entry: ; CHECK-NEXT: [[COL_LOAD:%.*]] = load <2 x double>, ptr [[A_PTR:%.*]], align 8 ; CHECK-NEXT: [[VEC_GEP:%.*]] = getelementptr double, ptr [[A_PTR]], i64 2 -; CHECK-NEXT: [[COL_LOAD1:%.*]] = load <2 x double>, ptr [[VEC_GEP]], align 8 -; CHECK-NEXT: [[VEC_GEP2:%.*]] = getelementptr double, ptr [[A_PTR]], i64 4 -; CHECK-NEXT: [[COL_LOAD3:%.*]] = load <2 x double>, ptr [[VEC_GEP2]], align 8 -; CHECK-NEXT: [[VEC_GEP4:%.*]] = getelementptr double, ptr [[A_PTR]], i64 6 -; CHECK-NEXT: [[COL_LOAD5:%.*]] = load <2 x double>, ptr [[VEC_GEP4]], align 8 -; CHECK-NEXT: [[SPLIT:%.*]] = shufflevector <8 x double> [[B:%.*]], <8 x double> poison, <2 x i32> -; CHECK-NEXT: [[SPLIT6:%.*]] = shufflevector <8 x double> [[B]], <8 x double> poison, <2 x i32> -; CHECK-NEXT: [[SPLIT7:%.*]] = shufflevector <8 x double> [[B]], <8 x double> poison, <2 x i32> -; CHECK-NEXT: [[SPLIT8:%.*]] = shufflevector <8 x double> [[B]], <8 x double> poison, <2 x i32> -; CHECK-NEXT: [[TMP0:%.*]] = fadd <2 x double> [[COL_LOAD]], [[SPLIT]] -; CHECK-NEXT: [[TMP1:%.*]] = fadd <2 x double> [[COL_LOAD1]], [[SPLIT6]] -; CHECK-NEXT: [[TMP2:%.*]] = fadd <2 x double> [[COL_LOAD3]], [[SPLIT7]] -; CHECK-NEXT: [[TMP3:%.*]] = fadd <2 x double> [[COL_LOAD5]], [[SPLIT8]] -; CHECK-NEXT: [[TMP4:%.*]] = extractelement <2 x double> [[TMP0]], i64 0 -; CHECK-NEXT: [[TMP5:%.*]] = insertelement <4 x double> poison, double [[TMP4]], i64 0 -; CHECK-NEXT: [[TMP6:%.*]] = extractelement <2 x double> [[TMP1]], i64 0 -; CHECK-NEXT: [[TMP7:%.*]] = insertelement <4 x double> [[TMP5]], double [[TMP6]], i64 1 -; CHECK-NEXT: [[TMP8:%.*]] = extractelement <2 x double> [[TMP2]], i64 0 -; CHECK-NEXT: [[TMP9:%.*]] = insertelement <4 x double> [[TMP7]], double [[TMP8]], i64 2 -; CHECK-NEXT: [[TMP10:%.*]] = extractelement <2 x double> [[TMP3]], i64 0 -; CHECK-NEXT: [[TMP11:%.*]] = insertelement <4 x double> [[TMP9]], double [[TMP10]], i64 3 -; CHECK-NEXT: [[TMP12:%.*]] = extractelement <2 x double> [[TMP0]], i64 1 -; CHECK-NEXT: [[TMP13:%.*]] = insertelement <4 x double> poison, double [[TMP12]], i64 0 -; CHECK-NEXT: [[TMP14:%.*]] = extractelement <2 x double> [[TMP1]], i64 1 -; CHECK-NEXT: [[TMP15:%.*]] = insertelement <4 x double> [[TMP13]], double [[TMP14]], i64 1 -; CHECK-NEXT: [[TMP16:%.*]] = extractelement <2 x double> [[TMP2]], i64 1 -; CHECK-NEXT: [[TMP17:%.*]] = insertelement <4 x double> [[TMP15]], double [[TMP16]], i64 2 -; CHECK-NEXT: [[TMP18:%.*]] = extractelement <2 x double> [[TMP3]], i64 1 -; CHECK-NEXT: [[TMP19:%.*]] = insertelement <4 x double> [[TMP17]], double [[TMP18]], i64 3 -; CHECK-NEXT: [[TMP20:%.*]] = shufflevector <4 x double> [[TMP11]], <4 x double> [[TMP19]], <8 x i32> -; CHECK-NEXT: ret <8 x double> [[TMP20]] +; CHECK-NEXT: [[COL_LOAD2:%.*]] = load <2 x double>, ptr [[VEC_GEP]], align 8 +; CHECK-NEXT: [[VEC_GEP3:%.*]] = getelementptr double, ptr [[A_PTR]], i64 4 +; CHECK-NEXT: [[COL_LOAD4:%.*]] = load <2 x double>, ptr [[VEC_GEP3]], align 8 +; CHECK-NEXT: [[VEC_GEP5:%.*]] = getelementptr double, ptr [[A_PTR]], i64 6 +; CHECK-NEXT: [[COL_LOAD6:%.*]] = load <2 x double>, ptr [[VEC_GEP5]], align 8 +; CHECK-NEXT: [[TMP0:%.*]] = shufflevector <2 x double> [[COL_LOAD]], <2 x double> [[COL_LOAD2]], <4 x i32> +; CHECK-NEXT: [[TMP1:%.*]] = shufflevector <2 x double> [[COL_LOAD4]], <2 x double> [[COL_LOAD6]], <4 x i32> +; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <4 x double> [[TMP0]], <4 x double> [[TMP1]], <8 x i32> +; CHECK-NEXT: [[SPLIT:%.*]] = shufflevector <8 x double> [[TMP2]], <8 x double> poison, <4 x i32> +; CHECK-NEXT: [[SPLIT7:%.*]] = shufflevector <8 x double> [[TMP2]], <8 x double> poison, <4 x i32> +; CHECK-NEXT: [[SPLIT8:%.*]] = shufflevector <8 x double> [[B:%.*]], <8 x double> poison, <4 x i32> +; CHECK-NEXT: [[SPLIT9:%.*]] = shufflevector <8 x double> [[B]], <8 x double> poison, <4 x i32> +; CHECK-NEXT: [[TMP3:%.*]] = fadd <4 x double> [[SPLIT]], [[SPLIT8]] +; CHECK-NEXT: [[TMP4:%.*]] = fadd <4 x double> [[SPLIT7]], [[SPLIT9]] +; CHECK-NEXT: [[TMP5:%.*]] = extractelement <4 x double> [[TMP3]], i64 0 +; CHECK-NEXT: [[TMP6:%.*]] = insertelement <2 x double> poison, double [[TMP5]], i64 0 +; CHECK-NEXT: [[TMP7:%.*]] = extractelement <4 x double> [[TMP4]], i64 0 +; CHECK-NEXT: [[TMP8:%.*]] = insertelement <2 x double> [[TMP6]], double [[TMP7]], i64 1 +; CHECK-NEXT: [[TMP9:%.*]] = extractelement <4 x double> [[TMP3]], i64 1 +; CHECK-NEXT: [[TMP10:%.*]] = insertelement <2 x double> poison, double [[TMP9]], i64 0 +; CHECK-NEXT: [[TMP11:%.*]] = extractelement <4 x double> [[TMP4]], i64 1 +; CHECK-NEXT: [[TMP12:%.*]] = insertelement <2 x double> [[TMP10]], double [[TMP11]], i64 1 +; CHECK-NEXT: [[TMP13:%.*]] = extractelement <4 x double> [[TMP3]], i64 2 +; CHECK-NEXT: [[TMP14:%.*]] = insertelement <2 x double> poison, double [[TMP13]], i64 0 +; CHECK-NEXT: [[TMP15:%.*]] = extractelement <4 x double> [[TMP4]], i64 2 +; CHECK-NEXT: [[TMP16:%.*]] = insertelement <2 x double> [[TMP14]], double [[TMP15]], i64 1 +; CHECK-NEXT: [[TMP17:%.*]] = extractelement <4 x double> [[TMP3]], i64 3 +; CHECK-NEXT: [[TMP18:%.*]] = insertelement <2 x double> poison, double [[TMP17]], i64 0 +; CHECK-NEXT: [[TMP19:%.*]] = extractelement <4 x double> [[TMP4]], i64 3 +; CHECK-NEXT: [[TMP20:%.*]] = insertelement <2 x double> [[TMP18]], double [[TMP19]], i64 1 +; CHECK-NEXT: [[TMP21:%.*]] = shufflevector <2 x double> [[TMP8]], <2 x double> [[TMP12]], <4 x i32> +; CHECK-NEXT: [[TMP22:%.*]] = shufflevector <2 x double> [[TMP16]], <2 x double> [[TMP20]], <4 x i32> +; CHECK-NEXT: [[TMP23:%.*]] = shufflevector <4 x double> [[TMP21]], <4 x double> [[TMP22]], <8 x i32> +; CHECK-NEXT: ret <8 x double> [[TMP23]] ;