-
Notifications
You must be signed in to change notification settings - Fork 14.8k
[SLPVectorizer] Remove align 16
in a test.
#161251
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
It is not necessary.
@llvm/pr-subscribers-backend-risc-v Author: Mikhail Gudim (mgudim) ChangesIt is not necessary. Patch is 36.55 KiB, truncated to 20.00 KiB below, full version: https://github.com/llvm/llvm-project/pull/161251.diff 1 Files Affected:
diff --git a/llvm/test/Transforms/SLPVectorizer/RISCV/basic-strided-loads.ll b/llvm/test/Transforms/SLPVectorizer/RISCV/basic-strided-loads.ll
index 645dbc49269f0..a4b2b9b8b6123 100644
--- a/llvm/test/Transforms/SLPVectorizer/RISCV/basic-strided-loads.ll
+++ b/llvm/test/Transforms/SLPVectorizer/RISCV/basic-strided-loads.ll
@@ -7,8 +7,8 @@ define void @const_stride_1_no_reordering(ptr %pl, ptr %ps) {
; CHECK-SAME: ptr [[PL:%.*]], ptr [[PS:%.*]]) #[[ATTR0:[0-9]+]] {
; CHECK-NEXT: [[GEP_L0:%.*]] = getelementptr inbounds i8, ptr [[PL]], i64 0
; CHECK-NEXT: [[GEP_S0:%.*]] = getelementptr inbounds i8, ptr [[PS]], i64 0
-; CHECK-NEXT: [[TMP1:%.*]] = load <16 x i8>, ptr [[GEP_L0]], align 16
-; CHECK-NEXT: store <16 x i8> [[TMP1]], ptr [[GEP_S0]], align 16
+; CHECK-NEXT: [[TMP1:%.*]] = load <16 x i8>, ptr [[GEP_L0]], align 1
+; CHECK-NEXT: store <16 x i8> [[TMP1]], ptr [[GEP_S0]], align 1
; CHECK-NEXT: ret void
;
%gep_l0 = getelementptr inbounds i8, ptr %pl, i64 0
@@ -28,22 +28,22 @@ define void @const_stride_1_no_reordering(ptr %pl, ptr %ps) {
%gep_l14 = getelementptr inbounds i8, ptr %pl, i64 14
%gep_l15 = getelementptr inbounds i8, ptr %pl, i64 15
- %load0 = load i8, ptr %gep_l0 , align 16
- %load1 = load i8, ptr %gep_l1 , align 16
- %load2 = load i8, ptr %gep_l2 , align 16
- %load3 = load i8, ptr %gep_l3 , align 16
- %load4 = load i8, ptr %gep_l4 , align 16
- %load5 = load i8, ptr %gep_l5 , align 16
- %load6 = load i8, ptr %gep_l6 , align 16
- %load7 = load i8, ptr %gep_l7 , align 16
- %load8 = load i8, ptr %gep_l8 , align 16
- %load9 = load i8, ptr %gep_l9 , align 16
- %load10 = load i8, ptr %gep_l10, align 16
- %load11 = load i8, ptr %gep_l11, align 16
- %load12 = load i8, ptr %gep_l12, align 16
- %load13 = load i8, ptr %gep_l13, align 16
- %load14 = load i8, ptr %gep_l14, align 16
- %load15 = load i8, ptr %gep_l15, align 16
+ %load0 = load i8, ptr %gep_l0
+ %load1 = load i8, ptr %gep_l1
+ %load2 = load i8, ptr %gep_l2
+ %load3 = load i8, ptr %gep_l3
+ %load4 = load i8, ptr %gep_l4
+ %load5 = load i8, ptr %gep_l5
+ %load6 = load i8, ptr %gep_l6
+ %load7 = load i8, ptr %gep_l7
+ %load8 = load i8, ptr %gep_l8
+ %load9 = load i8, ptr %gep_l9
+ %load10 = load i8, ptr %gep_l10
+ %load11 = load i8, ptr %gep_l11
+ %load12 = load i8, ptr %gep_l12
+ %load13 = load i8, ptr %gep_l13
+ %load14 = load i8, ptr %gep_l14
+ %load15 = load i8, ptr %gep_l15
%gep_s0 = getelementptr inbounds i8, ptr %ps, i64 0
%gep_s1 = getelementptr inbounds i8, ptr %ps, i64 1
@@ -62,22 +62,22 @@ define void @const_stride_1_no_reordering(ptr %pl, ptr %ps) {
%gep_s14 = getelementptr inbounds i8, ptr %ps, i64 14
%gep_s15 = getelementptr inbounds i8, ptr %ps, i64 15
- store i8 %load0, ptr %gep_s0, align 16
- store i8 %load1, ptr %gep_s1, align 16
- store i8 %load2, ptr %gep_s2, align 16
- store i8 %load3, ptr %gep_s3, align 16
- store i8 %load4, ptr %gep_s4, align 16
- store i8 %load5, ptr %gep_s5, align 16
- store i8 %load6, ptr %gep_s6, align 16
- store i8 %load7, ptr %gep_s7, align 16
- store i8 %load8, ptr %gep_s8, align 16
- store i8 %load9, ptr %gep_s9, align 16
- store i8 %load10, ptr %gep_s10, align 16
- store i8 %load11, ptr %gep_s11, align 16
- store i8 %load12, ptr %gep_s12, align 16
- store i8 %load13, ptr %gep_s13, align 16
- store i8 %load14, ptr %gep_s14, align 16
- store i8 %load15, ptr %gep_s15, align 16
+ store i8 %load0, ptr %gep_s0
+ store i8 %load1, ptr %gep_s1
+ store i8 %load2, ptr %gep_s2
+ store i8 %load3, ptr %gep_s3
+ store i8 %load4, ptr %gep_s4
+ store i8 %load5, ptr %gep_s5
+ store i8 %load6, ptr %gep_s6
+ store i8 %load7, ptr %gep_s7
+ store i8 %load8, ptr %gep_s8
+ store i8 %load9, ptr %gep_s9
+ store i8 %load10, ptr %gep_s10
+ store i8 %load11, ptr %gep_s11
+ store i8 %load12, ptr %gep_s12
+ store i8 %load13, ptr %gep_s13
+ store i8 %load14, ptr %gep_s14
+ store i8 %load15, ptr %gep_s15
ret void
}
@@ -87,9 +87,9 @@ define void @const_stride_1_with_reordering(ptr %pl, ptr %ps) {
; CHECK-SAME: ptr [[PL:%.*]], ptr [[PS:%.*]]) #[[ATTR0]] {
; CHECK-NEXT: [[GEP_L0:%.*]] = getelementptr inbounds i8, ptr [[PL]], i64 0
; CHECK-NEXT: [[GEP_S0:%.*]] = getelementptr inbounds i8, ptr [[PS]], i64 0
-; CHECK-NEXT: [[TMP1:%.*]] = load <16 x i8>, ptr [[GEP_L0]], align 16
+; CHECK-NEXT: [[TMP1:%.*]] = load <16 x i8>, ptr [[GEP_L0]], align 1
; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <16 x i8> [[TMP1]], <16 x i8> poison, <16 x i32> <i32 1, i32 0, i32 2, i32 3, i32 4, i32 5, i32 6, i32 7, i32 8, i32 9, i32 10, i32 11, i32 12, i32 13, i32 14, i32 15>
-; CHECK-NEXT: store <16 x i8> [[TMP2]], ptr [[GEP_S0]], align 16
+; CHECK-NEXT: store <16 x i8> [[TMP2]], ptr [[GEP_S0]], align 1
; CHECK-NEXT: ret void
;
%gep_l0 = getelementptr inbounds i8, ptr %pl, i64 0
@@ -109,22 +109,22 @@ define void @const_stride_1_with_reordering(ptr %pl, ptr %ps) {
%gep_l14 = getelementptr inbounds i8, ptr %pl, i64 14
%gep_l15 = getelementptr inbounds i8, ptr %pl, i64 15
- %load0 = load i8, ptr %gep_l0 , align 16
- %load1 = load i8, ptr %gep_l1 , align 16
- %load2 = load i8, ptr %gep_l2 , align 16
- %load3 = load i8, ptr %gep_l3 , align 16
- %load4 = load i8, ptr %gep_l4 , align 16
- %load5 = load i8, ptr %gep_l5 , align 16
- %load6 = load i8, ptr %gep_l6 , align 16
- %load7 = load i8, ptr %gep_l7 , align 16
- %load8 = load i8, ptr %gep_l8 , align 16
- %load9 = load i8, ptr %gep_l9 , align 16
- %load10 = load i8, ptr %gep_l10, align 16
- %load11 = load i8, ptr %gep_l11, align 16
- %load12 = load i8, ptr %gep_l12, align 16
- %load13 = load i8, ptr %gep_l13, align 16
- %load14 = load i8, ptr %gep_l14, align 16
- %load15 = load i8, ptr %gep_l15, align 16
+ %load0 = load i8, ptr %gep_l0
+ %load1 = load i8, ptr %gep_l1
+ %load2 = load i8, ptr %gep_l2
+ %load3 = load i8, ptr %gep_l3
+ %load4 = load i8, ptr %gep_l4
+ %load5 = load i8, ptr %gep_l5
+ %load6 = load i8, ptr %gep_l6
+ %load7 = load i8, ptr %gep_l7
+ %load8 = load i8, ptr %gep_l8
+ %load9 = load i8, ptr %gep_l9
+ %load10 = load i8, ptr %gep_l10
+ %load11 = load i8, ptr %gep_l11
+ %load12 = load i8, ptr %gep_l12
+ %load13 = load i8, ptr %gep_l13
+ %load14 = load i8, ptr %gep_l14
+ %load15 = load i8, ptr %gep_l15
%gep_s0 = getelementptr inbounds i8, ptr %ps, i64 0
%gep_s1 = getelementptr inbounds i8, ptr %ps, i64 1
@@ -144,22 +144,22 @@ define void @const_stride_1_with_reordering(ptr %pl, ptr %ps) {
%gep_s15 = getelementptr inbounds i8, ptr %ps, i64 15
; NOTE: value from %load1 in stored in %gep_s0
- store i8 %load1, ptr %gep_s0, align 16
- store i8 %load0, ptr %gep_s1, align 16
- store i8 %load2, ptr %gep_s2, align 16
- store i8 %load3, ptr %gep_s3, align 16
- store i8 %load4, ptr %gep_s4, align 16
- store i8 %load5, ptr %gep_s5, align 16
- store i8 %load6, ptr %gep_s6, align 16
- store i8 %load7, ptr %gep_s7, align 16
- store i8 %load8, ptr %gep_s8, align 16
- store i8 %load9, ptr %gep_s9, align 16
- store i8 %load10, ptr %gep_s10, align 16
- store i8 %load11, ptr %gep_s11, align 16
- store i8 %load12, ptr %gep_s12, align 16
- store i8 %load13, ptr %gep_s13, align 16
- store i8 %load14, ptr %gep_s14, align 16
- store i8 %load15, ptr %gep_s15, align 16
+ store i8 %load1, ptr %gep_s0
+ store i8 %load0, ptr %gep_s1
+ store i8 %load2, ptr %gep_s2
+ store i8 %load3, ptr %gep_s3
+ store i8 %load4, ptr %gep_s4
+ store i8 %load5, ptr %gep_s5
+ store i8 %load6, ptr %gep_s6
+ store i8 %load7, ptr %gep_s7
+ store i8 %load8, ptr %gep_s8
+ store i8 %load9, ptr %gep_s9
+ store i8 %load10, ptr %gep_s10
+ store i8 %load11, ptr %gep_s11
+ store i8 %load12, ptr %gep_s12
+ store i8 %load13, ptr %gep_s13
+ store i8 %load14, ptr %gep_s14
+ store i8 %load15, ptr %gep_s15
ret void
}
@@ -170,9 +170,9 @@ define void @const_stride_2_no_reordering(ptr %pl, ptr %ps) {
; CHECK-SAME: ptr [[PL:%.*]], ptr [[PS:%.*]]) #[[ATTR0]] {
; CHECK-NEXT: [[GEP_L0:%.*]] = getelementptr inbounds i8, ptr [[PL]], i64 0
; CHECK-NEXT: [[GEP_S0:%.*]] = getelementptr inbounds i8, ptr [[PS]], i64 0
-; CHECK-NEXT: [[TMP2:%.*]] = call <31 x i8> @llvm.masked.load.v31i8.p0(ptr [[GEP_L0]], i32 16, <31 x i1> <i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true>, <31 x i8> poison)
+; CHECK-NEXT: [[TMP2:%.*]] = call <31 x i8> @llvm.masked.load.v31i8.p0(ptr [[GEP_L0]], i32 1, <31 x i1> <i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true>, <31 x i8> poison)
; CHECK-NEXT: [[TMP1:%.*]] = shufflevector <31 x i8> [[TMP2]], <31 x i8> poison, <16 x i32> <i32 0, i32 2, i32 4, i32 6, i32 8, i32 10, i32 12, i32 14, i32 16, i32 18, i32 20, i32 22, i32 24, i32 26, i32 28, i32 30>
-; CHECK-NEXT: store <16 x i8> [[TMP1]], ptr [[GEP_S0]], align 16
+; CHECK-NEXT: store <16 x i8> [[TMP1]], ptr [[GEP_S0]], align 1
; CHECK-NEXT: ret void
;
%gep_l0 = getelementptr inbounds i8, ptr %pl, i64 0
@@ -192,22 +192,22 @@ define void @const_stride_2_no_reordering(ptr %pl, ptr %ps) {
%gep_l14 = getelementptr inbounds i8, ptr %pl, i64 28
%gep_l15 = getelementptr inbounds i8, ptr %pl, i64 30
- %load0 = load i8, ptr %gep_l0 , align 16
- %load1 = load i8, ptr %gep_l1 , align 16
- %load2 = load i8, ptr %gep_l2 , align 16
- %load3 = load i8, ptr %gep_l3 , align 16
- %load4 = load i8, ptr %gep_l4 , align 16
- %load5 = load i8, ptr %gep_l5 , align 16
- %load6 = load i8, ptr %gep_l6 , align 16
- %load7 = load i8, ptr %gep_l7 , align 16
- %load8 = load i8, ptr %gep_l8 , align 16
- %load9 = load i8, ptr %gep_l9 , align 16
- %load10 = load i8, ptr %gep_l10, align 16
- %load11 = load i8, ptr %gep_l11, align 16
- %load12 = load i8, ptr %gep_l12, align 16
- %load13 = load i8, ptr %gep_l13, align 16
- %load14 = load i8, ptr %gep_l14, align 16
- %load15 = load i8, ptr %gep_l15, align 16
+ %load0 = load i8, ptr %gep_l0
+ %load1 = load i8, ptr %gep_l1
+ %load2 = load i8, ptr %gep_l2
+ %load3 = load i8, ptr %gep_l3
+ %load4 = load i8, ptr %gep_l4
+ %load5 = load i8, ptr %gep_l5
+ %load6 = load i8, ptr %gep_l6
+ %load7 = load i8, ptr %gep_l7
+ %load8 = load i8, ptr %gep_l8
+ %load9 = load i8, ptr %gep_l9
+ %load10 = load i8, ptr %gep_l10
+ %load11 = load i8, ptr %gep_l11
+ %load12 = load i8, ptr %gep_l12
+ %load13 = load i8, ptr %gep_l13
+ %load14 = load i8, ptr %gep_l14
+ %load15 = load i8, ptr %gep_l15
%gep_s0 = getelementptr inbounds i8, ptr %ps, i64 0
%gep_s1 = getelementptr inbounds i8, ptr %ps, i64 1
@@ -226,22 +226,22 @@ define void @const_stride_2_no_reordering(ptr %pl, ptr %ps) {
%gep_s14 = getelementptr inbounds i8, ptr %ps, i64 14
%gep_s15 = getelementptr inbounds i8, ptr %ps, i64 15
- store i8 %load0, ptr %gep_s0, align 16
- store i8 %load1, ptr %gep_s1, align 16
- store i8 %load2, ptr %gep_s2, align 16
- store i8 %load3, ptr %gep_s3, align 16
- store i8 %load4, ptr %gep_s4, align 16
- store i8 %load5, ptr %gep_s5, align 16
- store i8 %load6, ptr %gep_s6, align 16
- store i8 %load7, ptr %gep_s7, align 16
- store i8 %load8, ptr %gep_s8, align 16
- store i8 %load9, ptr %gep_s9, align 16
- store i8 %load10, ptr %gep_s10, align 16
- store i8 %load11, ptr %gep_s11, align 16
- store i8 %load12, ptr %gep_s12, align 16
- store i8 %load13, ptr %gep_s13, align 16
- store i8 %load14, ptr %gep_s14, align 16
- store i8 %load15, ptr %gep_s15, align 16
+ store i8 %load0, ptr %gep_s0
+ store i8 %load1, ptr %gep_s1
+ store i8 %load2, ptr %gep_s2
+ store i8 %load3, ptr %gep_s3
+ store i8 %load4, ptr %gep_s4
+ store i8 %load5, ptr %gep_s5
+ store i8 %load6, ptr %gep_s6
+ store i8 %load7, ptr %gep_s7
+ store i8 %load8, ptr %gep_s8
+ store i8 %load9, ptr %gep_s9
+ store i8 %load10, ptr %gep_s10
+ store i8 %load11, ptr %gep_s11
+ store i8 %load12, ptr %gep_s12
+ store i8 %load13, ptr %gep_s13
+ store i8 %load14, ptr %gep_s14
+ store i8 %load15, ptr %gep_s15
ret void
}
@@ -251,10 +251,10 @@ define void @const_stride_2_with_reordering(ptr %pl, ptr %ps) {
; CHECK-SAME: ptr [[PL:%.*]], ptr [[PS:%.*]]) #[[ATTR0]] {
; CHECK-NEXT: [[GEP_L0:%.*]] = getelementptr inbounds i8, ptr [[PL]], i64 0
; CHECK-NEXT: [[GEP_S0:%.*]] = getelementptr inbounds i8, ptr [[PS]], i64 0
-; CHECK-NEXT: [[TMP1:%.*]] = call <31 x i8> @llvm.masked.load.v31i8.p0(ptr [[GEP_L0]], i32 16, <31 x i1> <i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true>, <31 x i8> poison)
+; CHECK-NEXT: [[TMP1:%.*]] = call <31 x i8> @llvm.masked.load.v31i8.p0(ptr [[GEP_L0]], i32 1, <31 x i1> <i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true>, <31 x i8> poison)
; CHECK-NEXT: [[TMP3:%.*]] = shufflevector <31 x i8> [[TMP1]], <31 x i8> poison, <16 x i32> <i32 0, i32 2, i32 4, i32 6, i32 8, i32 10, i32 12, i32 14, i32 16, i32 18, i32 20, i32 22, i32 24, i32 26, i32 28, i32 30>
; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <31 x i8> [[TMP1]], <31 x i8> poison, <16 x i32> <i32 2, i32 0, i32 4, i32 6, i32 8, i32 10, i32 12, i32 14, i32 16, i32 18, i32 20, i32 22, i32 24, i32 26, i32 28, i32 30>
-; CHECK-NEXT: store <16 x i8> [[TMP2]], ptr [[GEP_S0]], align 16
+; CHECK-NEXT: store <16 x i8> [[TMP2]], ptr [[GEP_S0]], align 1
; CHECK-NEXT: ret void
;
%gep_l0 = getelementptr inbounds i8, ptr %pl, i64 0
@@ -274,22 +274,22 @@ define void @const_stride_2_with_reordering(ptr %pl, ptr %ps) {
%gep_l14 = getelementptr inbounds i8, ptr %pl, i64 28
%gep_l15 = getelementptr inbounds i8, ptr %pl, i64 30
- %load0 = load i8, ptr %gep_l0 , align 16
- %load1 = load i8, ptr %gep_l1 , align 16
- %load2 = load i8, ptr %gep_l2 , align 16
- %load3 = load i8, ptr %gep_l3 , align 16
- %load4 = load i8, ptr %gep_l4 , align 16
- %load5 = load i8, ptr %gep_l5 , align 16
- %load6 = load i8, ptr %gep_l6 , align 16
- %load7 = load i8, ptr %gep_l7 , align 16
- %load8 = load i8, ptr %gep_l8 , align 16
- %load9 = load i8, ptr %gep_l9 , align 16
- %load10 = load i8, ptr %gep_l10, align 16
- %load11 = load i8, ptr %gep_l11, align 16
- %load12 = load i8, ptr %gep_l12, align 16
- %load13 = load i8, ptr %gep_l13, align 16
- %load14 = load i8, ptr %gep_l14, align 16
- %load15 = load i8, ptr %gep_l15, align 16
+ %load0 = load i8, ptr %gep_l0
+ %load1 = load i8, ptr %gep_l1
+ %load2 = load i8, ptr %gep_l2
+ %load3 = load i8, ptr %gep_l3
+ %load4 = load i8, ptr %gep_l4
+ %load5 = load i8, ptr %gep_l5
+ %load6 = load i8, ptr %gep_l6
+ %load7 = load i8, ptr %gep_l7
+ %load8 = load i8, ptr %gep_l8
+ %load9 = load i8, ptr %gep_l9
+ %load10 = load i8, ptr %gep_l10
+ %load11 = load i8, ptr %gep_l11
+ %load12 = load i8, ptr %gep_l12
+ %load13 = load i8, ptr %gep_l13
+ %load14 = load i8, ptr %gep_l14
+ %load15 = load i8, ptr %gep_l15
%gep_s0 = getelementptr inbounds i8, ptr %ps, i64 0
%gep_s1 = getelementptr inbounds i8, ptr %ps, i64 1
@@ -308,22 +308,22 @@ define void @const_stride_2_with_reordering(ptr %pl, ptr %ps) {
%gep_s14 = getelementptr inbounds i8, ptr %ps, i64 14
%gep_s15 = getelementptr inbounds i8, ptr %ps, i64 15
- store i8 %load1, ptr %gep_s0, align 16
- store i8 %load0, ptr %gep_s1, align 16
- store i8 %load2, ptr %gep_s2, align 16
- store i8 %load3, ptr %gep_s3, align 16
- store i8 %load4, ptr %gep_s4, align 16
- store i8 %load5, ptr %gep_s5, align 16
- store i8 %load6, ptr %gep_s6, align 16
- store i8 %load7, ptr %gep_s7, align 16
- store i8 %load8, ptr %gep_s8, align 16
- store i8 %load9, ptr %gep_s9, align 16
- store i8 %load10, ptr %gep_s10, align 16
- store i8 %load11, ptr %gep_s11, align 16
- store i8 %load12, ptr %gep_s12, align 16
- store i8 %load13, ptr %gep_s13, align 16
- store i8 %load14, ptr %gep_s14, align 16
- store i8 %load15, ptr %gep_s15, align 16
+ store i8 %load1, ptr %gep_s0
+ store i8 %load0, ptr %gep_s1
+ store i8 %load2, ptr %gep_s2
+ store i8 %load3, ptr %gep_s3
+ store i8 %load4, ptr %gep_s4
+ store i8 %load5, ptr %gep_s5
+ store i8 %load6, ptr %gep_s6
+ store i8 %load7, ptr %gep_s7
+ store i8 %load8, ptr %gep_s8
+ store i8 %load9, ptr %gep_s9
+ store i8 %load10, ptr %gep_s10
+ store i8 %load11, ptr %gep_s11
+ store i8 %load12, ptr %gep_s12
+ store i8 %load13, ptr %gep_s13
+ store i8 %load14, ptr %gep_s14
+ store i8 %load15, ptr %gep_s15
ret void
}
@@ -335,8 +335,8 @@ define void @rt_stride_1_no_reordering(ptr %pl, i64 %stride, ptr %ps) {
; CHECK-NEXT: [[GEP_L0:%.*]] = getelementptr inbounds i8, ptr [[PL]], i64 [[STRIDE0]]
; CHECK-NEXT: [[GEP_S0:%.*]] = getelementptr inbounds i8, ptr [[PS]], i64 0
; CHECK-NEXT: [[TMP1:%.*]] = mul i64 [[STRIDE]], 1
-; CHECK-NEXT: [[TMP2:%.*]] = call <16 x i8> @llvm.experimental.vp.strided.load.v16i8.p0.i64(ptr align 16 [[GEP_L0]], i64 [[TMP1]], <16 x i1> splat (i1 true), i32 16)
-; CHECK-NEXT: store <16 x i8> [[TMP2]], ptr [[GEP_S0]], align 16
+; CHECK-NEXT: [[TMP2:%.*]] = call <16 x i8> @llvm.experimental.vp.strided.load.v16i8.p0.i64(ptr align 1 [[GEP_L0]], i64 [[TMP1]], <16 x i1> splat (i1 true), i32 16)
+; CHECK-NEXT: store <16 x i8> [[TMP2]], ptr [[GEP_S0]], align 1
; CHECK-NEXT: ret void
;
%stride0 = mul nsw i64 %stride, 0
@@ -373,22 +373,22 @@ define void @rt_stride_1_no_reordering(ptr %pl, i64 %stride, ptr %ps) {
%gep_l14 = getelementptr inbounds i8, ptr %pl, i64 %stride14
%gep_l15 = getelementptr inbounds i8, ptr %pl, i64 %stride15
- %load0 = load i8, ptr %gep_l0 , align 16
- %load1 = load i8, ptr %gep_l1 , align 16
- %load2 = load i8, ptr %gep_l2 , align 16
- %load3 = load i8, ptr %gep_l3 , align 16
- %load4 = load i8, ptr %gep_l4 , align 16
- %load5 = load i8, ptr %gep_l5 , align 16
- %load6 = load i8, ptr %gep_l6 , align 16
- %load7 = load i8, ptr %gep_l7 , align 16
- %load8 = load i8, ptr %gep_l8 , align 16
- %load9 = load i8, ptr %gep_l9 , align 16
- %load10 = load i8, ptr %gep_l10, align 16
- %load11 = load i8, ptr %gep_l11, align 16
- %load12 = load i8, ptr %gep_l12, align 16
- %load13 = load i8, ptr %gep_l13, align 16
- %load14 = load i8, ptr %gep_l14, align 16
- %load15 = load i8, ptr %gep_l15, align 16
+ %load0 = load i8, ptr %gep_l0
+ %load1 = load i8, ptr %gep_l1
+ %load2 = load i8, ptr %gep_l2
+ %load3 = load i8, ptr %gep_l3
+ %load4 = load i8, ptr %gep_l4
+ %load5 = load i8, ptr %gep_l5
+ %load6 = load i8, ptr %gep_l6
+ %load7 = load i8, ptr %gep_l7
+ %load8 = load i8, ptr %gep_l8
+ %load9 = load i8, ptr %gep_l9
+ %load10 = load i8, ptr %gep_l10
+ %load11 = load i8, ptr %gep_l11
+ %load12 = load i8, ptr %gep_l12
+ %load13 = load i8, ptr %gep_l13
+ %load14 = load i8, ptr %gep_l14
+ %load15 = load i8, ptr %gep_l15
%gep_s0 = getelementptr inbounds i8, ptr %ps, i64 0
%gep_s1 = getelementptr inbounds i8, ptr %ps, i64 1
@@ -407,22 +407,22 @@ define void @rt_stride_1_no_reordering(ptr %pl, i64 %stride, ptr %ps) {
%gep_s14 = getelementptr inbounds i8, ptr %ps, i64 14
%gep_s15 = getelementptr inbounds i8, ptr %ps, i64 15
- store i8 %load0, ptr %gep_s0, align 16
- store i8 %load1, ptr %gep_s1, align 16
- store i8 %load2, ptr %gep_s2, align 16
- store i8 %load3, ptr %gep_s3, align 16
- store i8 %load4, ptr %gep_s4, align 16
- store i8 %load5, ptr %gep_s5, align 16
- store i8 %load6, ptr %gep_s6, align 16
- store i8 %load7, ptr %gep_s7, align 16
- store i8 %load8, ptr %gep_s8, align 16
- store i8 %load9, ptr %gep_s9, align 16
- store i8 %load10, ptr %gep_s10, align 16
- store i8 %load11, ptr %...
[truncated]
|
@llvm/pr-subscribers-llvm-transforms Author: Mikhail Gudim (mgudim) ChangesIt is not necessary. Patch is 36.55 KiB, truncated to 20.00 KiB below, full version: https://github.com/llvm/llvm-project/pull/161251.diff 1 Files Affected:
diff --git a/llvm/test/Transforms/SLPVectorizer/RISCV/basic-strided-loads.ll b/llvm/test/Transforms/SLPVectorizer/RISCV/basic-strided-loads.ll
index 645dbc49269f0..a4b2b9b8b6123 100644
--- a/llvm/test/Transforms/SLPVectorizer/RISCV/basic-strided-loads.ll
+++ b/llvm/test/Transforms/SLPVectorizer/RISCV/basic-strided-loads.ll
@@ -7,8 +7,8 @@ define void @const_stride_1_no_reordering(ptr %pl, ptr %ps) {
; CHECK-SAME: ptr [[PL:%.*]], ptr [[PS:%.*]]) #[[ATTR0:[0-9]+]] {
; CHECK-NEXT: [[GEP_L0:%.*]] = getelementptr inbounds i8, ptr [[PL]], i64 0
; CHECK-NEXT: [[GEP_S0:%.*]] = getelementptr inbounds i8, ptr [[PS]], i64 0
-; CHECK-NEXT: [[TMP1:%.*]] = load <16 x i8>, ptr [[GEP_L0]], align 16
-; CHECK-NEXT: store <16 x i8> [[TMP1]], ptr [[GEP_S0]], align 16
+; CHECK-NEXT: [[TMP1:%.*]] = load <16 x i8>, ptr [[GEP_L0]], align 1
+; CHECK-NEXT: store <16 x i8> [[TMP1]], ptr [[GEP_S0]], align 1
; CHECK-NEXT: ret void
;
%gep_l0 = getelementptr inbounds i8, ptr %pl, i64 0
@@ -28,22 +28,22 @@ define void @const_stride_1_no_reordering(ptr %pl, ptr %ps) {
%gep_l14 = getelementptr inbounds i8, ptr %pl, i64 14
%gep_l15 = getelementptr inbounds i8, ptr %pl, i64 15
- %load0 = load i8, ptr %gep_l0 , align 16
- %load1 = load i8, ptr %gep_l1 , align 16
- %load2 = load i8, ptr %gep_l2 , align 16
- %load3 = load i8, ptr %gep_l3 , align 16
- %load4 = load i8, ptr %gep_l4 , align 16
- %load5 = load i8, ptr %gep_l5 , align 16
- %load6 = load i8, ptr %gep_l6 , align 16
- %load7 = load i8, ptr %gep_l7 , align 16
- %load8 = load i8, ptr %gep_l8 , align 16
- %load9 = load i8, ptr %gep_l9 , align 16
- %load10 = load i8, ptr %gep_l10, align 16
- %load11 = load i8, ptr %gep_l11, align 16
- %load12 = load i8, ptr %gep_l12, align 16
- %load13 = load i8, ptr %gep_l13, align 16
- %load14 = load i8, ptr %gep_l14, align 16
- %load15 = load i8, ptr %gep_l15, align 16
+ %load0 = load i8, ptr %gep_l0
+ %load1 = load i8, ptr %gep_l1
+ %load2 = load i8, ptr %gep_l2
+ %load3 = load i8, ptr %gep_l3
+ %load4 = load i8, ptr %gep_l4
+ %load5 = load i8, ptr %gep_l5
+ %load6 = load i8, ptr %gep_l6
+ %load7 = load i8, ptr %gep_l7
+ %load8 = load i8, ptr %gep_l8
+ %load9 = load i8, ptr %gep_l9
+ %load10 = load i8, ptr %gep_l10
+ %load11 = load i8, ptr %gep_l11
+ %load12 = load i8, ptr %gep_l12
+ %load13 = load i8, ptr %gep_l13
+ %load14 = load i8, ptr %gep_l14
+ %load15 = load i8, ptr %gep_l15
%gep_s0 = getelementptr inbounds i8, ptr %ps, i64 0
%gep_s1 = getelementptr inbounds i8, ptr %ps, i64 1
@@ -62,22 +62,22 @@ define void @const_stride_1_no_reordering(ptr %pl, ptr %ps) {
%gep_s14 = getelementptr inbounds i8, ptr %ps, i64 14
%gep_s15 = getelementptr inbounds i8, ptr %ps, i64 15
- store i8 %load0, ptr %gep_s0, align 16
- store i8 %load1, ptr %gep_s1, align 16
- store i8 %load2, ptr %gep_s2, align 16
- store i8 %load3, ptr %gep_s3, align 16
- store i8 %load4, ptr %gep_s4, align 16
- store i8 %load5, ptr %gep_s5, align 16
- store i8 %load6, ptr %gep_s6, align 16
- store i8 %load7, ptr %gep_s7, align 16
- store i8 %load8, ptr %gep_s8, align 16
- store i8 %load9, ptr %gep_s9, align 16
- store i8 %load10, ptr %gep_s10, align 16
- store i8 %load11, ptr %gep_s11, align 16
- store i8 %load12, ptr %gep_s12, align 16
- store i8 %load13, ptr %gep_s13, align 16
- store i8 %load14, ptr %gep_s14, align 16
- store i8 %load15, ptr %gep_s15, align 16
+ store i8 %load0, ptr %gep_s0
+ store i8 %load1, ptr %gep_s1
+ store i8 %load2, ptr %gep_s2
+ store i8 %load3, ptr %gep_s3
+ store i8 %load4, ptr %gep_s4
+ store i8 %load5, ptr %gep_s5
+ store i8 %load6, ptr %gep_s6
+ store i8 %load7, ptr %gep_s7
+ store i8 %load8, ptr %gep_s8
+ store i8 %load9, ptr %gep_s9
+ store i8 %load10, ptr %gep_s10
+ store i8 %load11, ptr %gep_s11
+ store i8 %load12, ptr %gep_s12
+ store i8 %load13, ptr %gep_s13
+ store i8 %load14, ptr %gep_s14
+ store i8 %load15, ptr %gep_s15
ret void
}
@@ -87,9 +87,9 @@ define void @const_stride_1_with_reordering(ptr %pl, ptr %ps) {
; CHECK-SAME: ptr [[PL:%.*]], ptr [[PS:%.*]]) #[[ATTR0]] {
; CHECK-NEXT: [[GEP_L0:%.*]] = getelementptr inbounds i8, ptr [[PL]], i64 0
; CHECK-NEXT: [[GEP_S0:%.*]] = getelementptr inbounds i8, ptr [[PS]], i64 0
-; CHECK-NEXT: [[TMP1:%.*]] = load <16 x i8>, ptr [[GEP_L0]], align 16
+; CHECK-NEXT: [[TMP1:%.*]] = load <16 x i8>, ptr [[GEP_L0]], align 1
; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <16 x i8> [[TMP1]], <16 x i8> poison, <16 x i32> <i32 1, i32 0, i32 2, i32 3, i32 4, i32 5, i32 6, i32 7, i32 8, i32 9, i32 10, i32 11, i32 12, i32 13, i32 14, i32 15>
-; CHECK-NEXT: store <16 x i8> [[TMP2]], ptr [[GEP_S0]], align 16
+; CHECK-NEXT: store <16 x i8> [[TMP2]], ptr [[GEP_S0]], align 1
; CHECK-NEXT: ret void
;
%gep_l0 = getelementptr inbounds i8, ptr %pl, i64 0
@@ -109,22 +109,22 @@ define void @const_stride_1_with_reordering(ptr %pl, ptr %ps) {
%gep_l14 = getelementptr inbounds i8, ptr %pl, i64 14
%gep_l15 = getelementptr inbounds i8, ptr %pl, i64 15
- %load0 = load i8, ptr %gep_l0 , align 16
- %load1 = load i8, ptr %gep_l1 , align 16
- %load2 = load i8, ptr %gep_l2 , align 16
- %load3 = load i8, ptr %gep_l3 , align 16
- %load4 = load i8, ptr %gep_l4 , align 16
- %load5 = load i8, ptr %gep_l5 , align 16
- %load6 = load i8, ptr %gep_l6 , align 16
- %load7 = load i8, ptr %gep_l7 , align 16
- %load8 = load i8, ptr %gep_l8 , align 16
- %load9 = load i8, ptr %gep_l9 , align 16
- %load10 = load i8, ptr %gep_l10, align 16
- %load11 = load i8, ptr %gep_l11, align 16
- %load12 = load i8, ptr %gep_l12, align 16
- %load13 = load i8, ptr %gep_l13, align 16
- %load14 = load i8, ptr %gep_l14, align 16
- %load15 = load i8, ptr %gep_l15, align 16
+ %load0 = load i8, ptr %gep_l0
+ %load1 = load i8, ptr %gep_l1
+ %load2 = load i8, ptr %gep_l2
+ %load3 = load i8, ptr %gep_l3
+ %load4 = load i8, ptr %gep_l4
+ %load5 = load i8, ptr %gep_l5
+ %load6 = load i8, ptr %gep_l6
+ %load7 = load i8, ptr %gep_l7
+ %load8 = load i8, ptr %gep_l8
+ %load9 = load i8, ptr %gep_l9
+ %load10 = load i8, ptr %gep_l10
+ %load11 = load i8, ptr %gep_l11
+ %load12 = load i8, ptr %gep_l12
+ %load13 = load i8, ptr %gep_l13
+ %load14 = load i8, ptr %gep_l14
+ %load15 = load i8, ptr %gep_l15
%gep_s0 = getelementptr inbounds i8, ptr %ps, i64 0
%gep_s1 = getelementptr inbounds i8, ptr %ps, i64 1
@@ -144,22 +144,22 @@ define void @const_stride_1_with_reordering(ptr %pl, ptr %ps) {
%gep_s15 = getelementptr inbounds i8, ptr %ps, i64 15
; NOTE: value from %load1 in stored in %gep_s0
- store i8 %load1, ptr %gep_s0, align 16
- store i8 %load0, ptr %gep_s1, align 16
- store i8 %load2, ptr %gep_s2, align 16
- store i8 %load3, ptr %gep_s3, align 16
- store i8 %load4, ptr %gep_s4, align 16
- store i8 %load5, ptr %gep_s5, align 16
- store i8 %load6, ptr %gep_s6, align 16
- store i8 %load7, ptr %gep_s7, align 16
- store i8 %load8, ptr %gep_s8, align 16
- store i8 %load9, ptr %gep_s9, align 16
- store i8 %load10, ptr %gep_s10, align 16
- store i8 %load11, ptr %gep_s11, align 16
- store i8 %load12, ptr %gep_s12, align 16
- store i8 %load13, ptr %gep_s13, align 16
- store i8 %load14, ptr %gep_s14, align 16
- store i8 %load15, ptr %gep_s15, align 16
+ store i8 %load1, ptr %gep_s0
+ store i8 %load0, ptr %gep_s1
+ store i8 %load2, ptr %gep_s2
+ store i8 %load3, ptr %gep_s3
+ store i8 %load4, ptr %gep_s4
+ store i8 %load5, ptr %gep_s5
+ store i8 %load6, ptr %gep_s6
+ store i8 %load7, ptr %gep_s7
+ store i8 %load8, ptr %gep_s8
+ store i8 %load9, ptr %gep_s9
+ store i8 %load10, ptr %gep_s10
+ store i8 %load11, ptr %gep_s11
+ store i8 %load12, ptr %gep_s12
+ store i8 %load13, ptr %gep_s13
+ store i8 %load14, ptr %gep_s14
+ store i8 %load15, ptr %gep_s15
ret void
}
@@ -170,9 +170,9 @@ define void @const_stride_2_no_reordering(ptr %pl, ptr %ps) {
; CHECK-SAME: ptr [[PL:%.*]], ptr [[PS:%.*]]) #[[ATTR0]] {
; CHECK-NEXT: [[GEP_L0:%.*]] = getelementptr inbounds i8, ptr [[PL]], i64 0
; CHECK-NEXT: [[GEP_S0:%.*]] = getelementptr inbounds i8, ptr [[PS]], i64 0
-; CHECK-NEXT: [[TMP2:%.*]] = call <31 x i8> @llvm.masked.load.v31i8.p0(ptr [[GEP_L0]], i32 16, <31 x i1> <i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true>, <31 x i8> poison)
+; CHECK-NEXT: [[TMP2:%.*]] = call <31 x i8> @llvm.masked.load.v31i8.p0(ptr [[GEP_L0]], i32 1, <31 x i1> <i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true>, <31 x i8> poison)
; CHECK-NEXT: [[TMP1:%.*]] = shufflevector <31 x i8> [[TMP2]], <31 x i8> poison, <16 x i32> <i32 0, i32 2, i32 4, i32 6, i32 8, i32 10, i32 12, i32 14, i32 16, i32 18, i32 20, i32 22, i32 24, i32 26, i32 28, i32 30>
-; CHECK-NEXT: store <16 x i8> [[TMP1]], ptr [[GEP_S0]], align 16
+; CHECK-NEXT: store <16 x i8> [[TMP1]], ptr [[GEP_S0]], align 1
; CHECK-NEXT: ret void
;
%gep_l0 = getelementptr inbounds i8, ptr %pl, i64 0
@@ -192,22 +192,22 @@ define void @const_stride_2_no_reordering(ptr %pl, ptr %ps) {
%gep_l14 = getelementptr inbounds i8, ptr %pl, i64 28
%gep_l15 = getelementptr inbounds i8, ptr %pl, i64 30
- %load0 = load i8, ptr %gep_l0 , align 16
- %load1 = load i8, ptr %gep_l1 , align 16
- %load2 = load i8, ptr %gep_l2 , align 16
- %load3 = load i8, ptr %gep_l3 , align 16
- %load4 = load i8, ptr %gep_l4 , align 16
- %load5 = load i8, ptr %gep_l5 , align 16
- %load6 = load i8, ptr %gep_l6 , align 16
- %load7 = load i8, ptr %gep_l7 , align 16
- %load8 = load i8, ptr %gep_l8 , align 16
- %load9 = load i8, ptr %gep_l9 , align 16
- %load10 = load i8, ptr %gep_l10, align 16
- %load11 = load i8, ptr %gep_l11, align 16
- %load12 = load i8, ptr %gep_l12, align 16
- %load13 = load i8, ptr %gep_l13, align 16
- %load14 = load i8, ptr %gep_l14, align 16
- %load15 = load i8, ptr %gep_l15, align 16
+ %load0 = load i8, ptr %gep_l0
+ %load1 = load i8, ptr %gep_l1
+ %load2 = load i8, ptr %gep_l2
+ %load3 = load i8, ptr %gep_l3
+ %load4 = load i8, ptr %gep_l4
+ %load5 = load i8, ptr %gep_l5
+ %load6 = load i8, ptr %gep_l6
+ %load7 = load i8, ptr %gep_l7
+ %load8 = load i8, ptr %gep_l8
+ %load9 = load i8, ptr %gep_l9
+ %load10 = load i8, ptr %gep_l10
+ %load11 = load i8, ptr %gep_l11
+ %load12 = load i8, ptr %gep_l12
+ %load13 = load i8, ptr %gep_l13
+ %load14 = load i8, ptr %gep_l14
+ %load15 = load i8, ptr %gep_l15
%gep_s0 = getelementptr inbounds i8, ptr %ps, i64 0
%gep_s1 = getelementptr inbounds i8, ptr %ps, i64 1
@@ -226,22 +226,22 @@ define void @const_stride_2_no_reordering(ptr %pl, ptr %ps) {
%gep_s14 = getelementptr inbounds i8, ptr %ps, i64 14
%gep_s15 = getelementptr inbounds i8, ptr %ps, i64 15
- store i8 %load0, ptr %gep_s0, align 16
- store i8 %load1, ptr %gep_s1, align 16
- store i8 %load2, ptr %gep_s2, align 16
- store i8 %load3, ptr %gep_s3, align 16
- store i8 %load4, ptr %gep_s4, align 16
- store i8 %load5, ptr %gep_s5, align 16
- store i8 %load6, ptr %gep_s6, align 16
- store i8 %load7, ptr %gep_s7, align 16
- store i8 %load8, ptr %gep_s8, align 16
- store i8 %load9, ptr %gep_s9, align 16
- store i8 %load10, ptr %gep_s10, align 16
- store i8 %load11, ptr %gep_s11, align 16
- store i8 %load12, ptr %gep_s12, align 16
- store i8 %load13, ptr %gep_s13, align 16
- store i8 %load14, ptr %gep_s14, align 16
- store i8 %load15, ptr %gep_s15, align 16
+ store i8 %load0, ptr %gep_s0
+ store i8 %load1, ptr %gep_s1
+ store i8 %load2, ptr %gep_s2
+ store i8 %load3, ptr %gep_s3
+ store i8 %load4, ptr %gep_s4
+ store i8 %load5, ptr %gep_s5
+ store i8 %load6, ptr %gep_s6
+ store i8 %load7, ptr %gep_s7
+ store i8 %load8, ptr %gep_s8
+ store i8 %load9, ptr %gep_s9
+ store i8 %load10, ptr %gep_s10
+ store i8 %load11, ptr %gep_s11
+ store i8 %load12, ptr %gep_s12
+ store i8 %load13, ptr %gep_s13
+ store i8 %load14, ptr %gep_s14
+ store i8 %load15, ptr %gep_s15
ret void
}
@@ -251,10 +251,10 @@ define void @const_stride_2_with_reordering(ptr %pl, ptr %ps) {
; CHECK-SAME: ptr [[PL:%.*]], ptr [[PS:%.*]]) #[[ATTR0]] {
; CHECK-NEXT: [[GEP_L0:%.*]] = getelementptr inbounds i8, ptr [[PL]], i64 0
; CHECK-NEXT: [[GEP_S0:%.*]] = getelementptr inbounds i8, ptr [[PS]], i64 0
-; CHECK-NEXT: [[TMP1:%.*]] = call <31 x i8> @llvm.masked.load.v31i8.p0(ptr [[GEP_L0]], i32 16, <31 x i1> <i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true>, <31 x i8> poison)
+; CHECK-NEXT: [[TMP1:%.*]] = call <31 x i8> @llvm.masked.load.v31i8.p0(ptr [[GEP_L0]], i32 1, <31 x i1> <i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true>, <31 x i8> poison)
; CHECK-NEXT: [[TMP3:%.*]] = shufflevector <31 x i8> [[TMP1]], <31 x i8> poison, <16 x i32> <i32 0, i32 2, i32 4, i32 6, i32 8, i32 10, i32 12, i32 14, i32 16, i32 18, i32 20, i32 22, i32 24, i32 26, i32 28, i32 30>
; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <31 x i8> [[TMP1]], <31 x i8> poison, <16 x i32> <i32 2, i32 0, i32 4, i32 6, i32 8, i32 10, i32 12, i32 14, i32 16, i32 18, i32 20, i32 22, i32 24, i32 26, i32 28, i32 30>
-; CHECK-NEXT: store <16 x i8> [[TMP2]], ptr [[GEP_S0]], align 16
+; CHECK-NEXT: store <16 x i8> [[TMP2]], ptr [[GEP_S0]], align 1
; CHECK-NEXT: ret void
;
%gep_l0 = getelementptr inbounds i8, ptr %pl, i64 0
@@ -274,22 +274,22 @@ define void @const_stride_2_with_reordering(ptr %pl, ptr %ps) {
%gep_l14 = getelementptr inbounds i8, ptr %pl, i64 28
%gep_l15 = getelementptr inbounds i8, ptr %pl, i64 30
- %load0 = load i8, ptr %gep_l0 , align 16
- %load1 = load i8, ptr %gep_l1 , align 16
- %load2 = load i8, ptr %gep_l2 , align 16
- %load3 = load i8, ptr %gep_l3 , align 16
- %load4 = load i8, ptr %gep_l4 , align 16
- %load5 = load i8, ptr %gep_l5 , align 16
- %load6 = load i8, ptr %gep_l6 , align 16
- %load7 = load i8, ptr %gep_l7 , align 16
- %load8 = load i8, ptr %gep_l8 , align 16
- %load9 = load i8, ptr %gep_l9 , align 16
- %load10 = load i8, ptr %gep_l10, align 16
- %load11 = load i8, ptr %gep_l11, align 16
- %load12 = load i8, ptr %gep_l12, align 16
- %load13 = load i8, ptr %gep_l13, align 16
- %load14 = load i8, ptr %gep_l14, align 16
- %load15 = load i8, ptr %gep_l15, align 16
+ %load0 = load i8, ptr %gep_l0
+ %load1 = load i8, ptr %gep_l1
+ %load2 = load i8, ptr %gep_l2
+ %load3 = load i8, ptr %gep_l3
+ %load4 = load i8, ptr %gep_l4
+ %load5 = load i8, ptr %gep_l5
+ %load6 = load i8, ptr %gep_l6
+ %load7 = load i8, ptr %gep_l7
+ %load8 = load i8, ptr %gep_l8
+ %load9 = load i8, ptr %gep_l9
+ %load10 = load i8, ptr %gep_l10
+ %load11 = load i8, ptr %gep_l11
+ %load12 = load i8, ptr %gep_l12
+ %load13 = load i8, ptr %gep_l13
+ %load14 = load i8, ptr %gep_l14
+ %load15 = load i8, ptr %gep_l15
%gep_s0 = getelementptr inbounds i8, ptr %ps, i64 0
%gep_s1 = getelementptr inbounds i8, ptr %ps, i64 1
@@ -308,22 +308,22 @@ define void @const_stride_2_with_reordering(ptr %pl, ptr %ps) {
%gep_s14 = getelementptr inbounds i8, ptr %ps, i64 14
%gep_s15 = getelementptr inbounds i8, ptr %ps, i64 15
- store i8 %load1, ptr %gep_s0, align 16
- store i8 %load0, ptr %gep_s1, align 16
- store i8 %load2, ptr %gep_s2, align 16
- store i8 %load3, ptr %gep_s3, align 16
- store i8 %load4, ptr %gep_s4, align 16
- store i8 %load5, ptr %gep_s5, align 16
- store i8 %load6, ptr %gep_s6, align 16
- store i8 %load7, ptr %gep_s7, align 16
- store i8 %load8, ptr %gep_s8, align 16
- store i8 %load9, ptr %gep_s9, align 16
- store i8 %load10, ptr %gep_s10, align 16
- store i8 %load11, ptr %gep_s11, align 16
- store i8 %load12, ptr %gep_s12, align 16
- store i8 %load13, ptr %gep_s13, align 16
- store i8 %load14, ptr %gep_s14, align 16
- store i8 %load15, ptr %gep_s15, align 16
+ store i8 %load1, ptr %gep_s0
+ store i8 %load0, ptr %gep_s1
+ store i8 %load2, ptr %gep_s2
+ store i8 %load3, ptr %gep_s3
+ store i8 %load4, ptr %gep_s4
+ store i8 %load5, ptr %gep_s5
+ store i8 %load6, ptr %gep_s6
+ store i8 %load7, ptr %gep_s7
+ store i8 %load8, ptr %gep_s8
+ store i8 %load9, ptr %gep_s9
+ store i8 %load10, ptr %gep_s10
+ store i8 %load11, ptr %gep_s11
+ store i8 %load12, ptr %gep_s12
+ store i8 %load13, ptr %gep_s13
+ store i8 %load14, ptr %gep_s14
+ store i8 %load15, ptr %gep_s15
ret void
}
@@ -335,8 +335,8 @@ define void @rt_stride_1_no_reordering(ptr %pl, i64 %stride, ptr %ps) {
; CHECK-NEXT: [[GEP_L0:%.*]] = getelementptr inbounds i8, ptr [[PL]], i64 [[STRIDE0]]
; CHECK-NEXT: [[GEP_S0:%.*]] = getelementptr inbounds i8, ptr [[PS]], i64 0
; CHECK-NEXT: [[TMP1:%.*]] = mul i64 [[STRIDE]], 1
-; CHECK-NEXT: [[TMP2:%.*]] = call <16 x i8> @llvm.experimental.vp.strided.load.v16i8.p0.i64(ptr align 16 [[GEP_L0]], i64 [[TMP1]], <16 x i1> splat (i1 true), i32 16)
-; CHECK-NEXT: store <16 x i8> [[TMP2]], ptr [[GEP_S0]], align 16
+; CHECK-NEXT: [[TMP2:%.*]] = call <16 x i8> @llvm.experimental.vp.strided.load.v16i8.p0.i64(ptr align 1 [[GEP_L0]], i64 [[TMP1]], <16 x i1> splat (i1 true), i32 16)
+; CHECK-NEXT: store <16 x i8> [[TMP2]], ptr [[GEP_S0]], align 1
; CHECK-NEXT: ret void
;
%stride0 = mul nsw i64 %stride, 0
@@ -373,22 +373,22 @@ define void @rt_stride_1_no_reordering(ptr %pl, i64 %stride, ptr %ps) {
%gep_l14 = getelementptr inbounds i8, ptr %pl, i64 %stride14
%gep_l15 = getelementptr inbounds i8, ptr %pl, i64 %stride15
- %load0 = load i8, ptr %gep_l0 , align 16
- %load1 = load i8, ptr %gep_l1 , align 16
- %load2 = load i8, ptr %gep_l2 , align 16
- %load3 = load i8, ptr %gep_l3 , align 16
- %load4 = load i8, ptr %gep_l4 , align 16
- %load5 = load i8, ptr %gep_l5 , align 16
- %load6 = load i8, ptr %gep_l6 , align 16
- %load7 = load i8, ptr %gep_l7 , align 16
- %load8 = load i8, ptr %gep_l8 , align 16
- %load9 = load i8, ptr %gep_l9 , align 16
- %load10 = load i8, ptr %gep_l10, align 16
- %load11 = load i8, ptr %gep_l11, align 16
- %load12 = load i8, ptr %gep_l12, align 16
- %load13 = load i8, ptr %gep_l13, align 16
- %load14 = load i8, ptr %gep_l14, align 16
- %load15 = load i8, ptr %gep_l15, align 16
+ %load0 = load i8, ptr %gep_l0
+ %load1 = load i8, ptr %gep_l1
+ %load2 = load i8, ptr %gep_l2
+ %load3 = load i8, ptr %gep_l3
+ %load4 = load i8, ptr %gep_l4
+ %load5 = load i8, ptr %gep_l5
+ %load6 = load i8, ptr %gep_l6
+ %load7 = load i8, ptr %gep_l7
+ %load8 = load i8, ptr %gep_l8
+ %load9 = load i8, ptr %gep_l9
+ %load10 = load i8, ptr %gep_l10
+ %load11 = load i8, ptr %gep_l11
+ %load12 = load i8, ptr %gep_l12
+ %load13 = load i8, ptr %gep_l13
+ %load14 = load i8, ptr %gep_l14
+ %load15 = load i8, ptr %gep_l15
%gep_s0 = getelementptr inbounds i8, ptr %ps, i64 0
%gep_s1 = getelementptr inbounds i8, ptr %ps, i64 1
@@ -407,22 +407,22 @@ define void @rt_stride_1_no_reordering(ptr %pl, i64 %stride, ptr %ps) {
%gep_s14 = getelementptr inbounds i8, ptr %ps, i64 14
%gep_s15 = getelementptr inbounds i8, ptr %ps, i64 15
- store i8 %load0, ptr %gep_s0, align 16
- store i8 %load1, ptr %gep_s1, align 16
- store i8 %load2, ptr %gep_s2, align 16
- store i8 %load3, ptr %gep_s3, align 16
- store i8 %load4, ptr %gep_s4, align 16
- store i8 %load5, ptr %gep_s5, align 16
- store i8 %load6, ptr %gep_s6, align 16
- store i8 %load7, ptr %gep_s7, align 16
- store i8 %load8, ptr %gep_s8, align 16
- store i8 %load9, ptr %gep_s9, align 16
- store i8 %load10, ptr %gep_s10, align 16
- store i8 %load11, ptr %...
[truncated]
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
make it align 1, do not drop alignment completely
LLVM Buildbot has detected a new failure on builder Full details are available at: https://lab.llvm.org/buildbot/#/builders/116/builds/19061 Here is the relevant piece of the build log for the reference
|
It is not necessary.
It is not necessary.