diff --git a/llvm/test/Transforms/InstCombine/bswap-fold.ll b/llvm/test/Transforms/InstCombine/bswap-fold.ll index 8fdecb628b858e..5e9c0923d6c5a4 100644 --- a/llvm/test/Transforms/InstCombine/bswap-fold.ll +++ b/llvm/test/Transforms/InstCombine/bswap-fold.ll @@ -5,29 +5,29 @@ ; A & 255 define i32 @test4(i32 %a) nounwind { ; CHECK-LABEL: @test4( -; CHECK-NEXT: [[TMP2:%.*]] = and i32 %a, 255 -; CHECK-NEXT: ret i32 [[TMP2]] +; CHECK-NEXT: [[T2:%.*]] = and i32 [[A:%.*]], 255 +; CHECK-NEXT: ret i32 [[T2]] ; - %tmp2 = tail call i32 @llvm.bswap.i32( i32 %a ) - %tmp4 = lshr i32 %tmp2, 24 - ret i32 %tmp4 + %t2 = tail call i32 @llvm.bswap.i32( i32 %a ) + %t4 = lshr i32 %t2, 24 + ret i32 %t4 } ; a >> 24 define i32 @test6(i32 %a) nounwind { ; CHECK-LABEL: @test6( -; CHECK-NEXT: [[TMP2:%.*]] = lshr i32 %a, 24 -; CHECK-NEXT: ret i32 [[TMP2]] +; CHECK-NEXT: [[T2:%.*]] = lshr i32 [[A:%.*]], 24 +; CHECK-NEXT: ret i32 [[T2]] ; - %tmp2 = tail call i32 @llvm.bswap.i32( i32 %a ) - %tmp4 = and i32 %tmp2, 255 - ret i32 %tmp4 + %t2 = tail call i32 @llvm.bswap.i32( i32 %a ) + %t4 = and i32 %t2, 255 + ret i32 %t4 } ; PR5284 define i16 @test7(i32 %A) { ; CHECK-LABEL: @test7( -; CHECK-NEXT: [[TMP1:%.*]] = lshr i32 %A, 16 +; CHECK-NEXT: [[TMP1:%.*]] = lshr i32 [[A:%.*]], 16 ; CHECK-NEXT: [[D:%.*]] = trunc i32 [[TMP1]] to i16 ; CHECK-NEXT: ret i16 [[D]] ; @@ -39,7 +39,7 @@ define i16 @test7(i32 %A) { define i16 @test8(i64 %A) { ; CHECK-LABEL: @test8( -; CHECK-NEXT: [[TMP1:%.*]] = lshr i64 %A, 48 +; CHECK-NEXT: [[TMP1:%.*]] = lshr i64 [[A:%.*]], 48 ; CHECK-NEXT: [[D:%.*]] = trunc i64 [[TMP1]] to i16 ; CHECK-NEXT: ret i16 [[D]] ; @@ -63,7 +63,7 @@ define i64 @foo() { ; Fold: OP( BSWAP(x), CONSTANT ) -> BSWAP( OP(x, BSWAP(CONSTANT) ) ) define i16 @bs_and16i(i16 %a, i16 %b) #0 { ; CHECK-LABEL: @bs_and16i( -; CHECK-NEXT: [[TMP1:%.*]] = and i16 %a, 4391 +; CHECK-NEXT: [[TMP1:%.*]] = and i16 [[A:%.*]], 4391 ; CHECK-NEXT: [[TMP2:%.*]] = call i16 @llvm.bswap.i16(i16 [[TMP1]]) ; CHECK-NEXT: ret i16 [[TMP2]] ; @@ -74,132 +74,132 @@ define i16 @bs_and16i(i16 %a, i16 %b) #0 { define i16 @bs_and16(i16 %a, i16 %b) #0 { ; CHECK-LABEL: @bs_and16( -; CHECK-NEXT: [[TMP1:%.*]] = and i16 %a, %b +; CHECK-NEXT: [[TMP1:%.*]] = and i16 [[A:%.*]], [[B:%.*]] ; CHECK-NEXT: [[TMP2:%.*]] = call i16 @llvm.bswap.i16(i16 [[TMP1]]) ; CHECK-NEXT: ret i16 [[TMP2]] ; - %tmp1 = tail call i16 @llvm.bswap.i16(i16 %a) - %tmp2 = tail call i16 @llvm.bswap.i16(i16 %b) - %tmp3 = and i16 %tmp1, %tmp2 - ret i16 %tmp3 + %t1 = tail call i16 @llvm.bswap.i16(i16 %a) + %t2 = tail call i16 @llvm.bswap.i16(i16 %b) + %t3 = and i16 %t1, %t2 + ret i16 %t3 } define i16 @bs_or16(i16 %a, i16 %b) #0 { ; CHECK-LABEL: @bs_or16( -; CHECK-NEXT: [[TMP1:%.*]] = or i16 %a, %b +; CHECK-NEXT: [[TMP1:%.*]] = or i16 [[A:%.*]], [[B:%.*]] ; CHECK-NEXT: [[TMP2:%.*]] = call i16 @llvm.bswap.i16(i16 [[TMP1]]) ; CHECK-NEXT: ret i16 [[TMP2]] ; - %tmp1 = tail call i16 @llvm.bswap.i16(i16 %a) - %tmp2 = tail call i16 @llvm.bswap.i16(i16 %b) - %tmp3 = or i16 %tmp1, %tmp2 - ret i16 %tmp3 + %t1 = tail call i16 @llvm.bswap.i16(i16 %a) + %t2 = tail call i16 @llvm.bswap.i16(i16 %b) + %t3 = or i16 %t1, %t2 + ret i16 %t3 } define i16 @bs_xor16(i16 %a, i16 %b) #0 { ; CHECK-LABEL: @bs_xor16( -; CHECK-NEXT: [[TMP1:%.*]] = xor i16 %a, %b +; CHECK-NEXT: [[TMP1:%.*]] = xor i16 [[A:%.*]], [[B:%.*]] ; CHECK-NEXT: [[TMP2:%.*]] = call i16 @llvm.bswap.i16(i16 [[TMP1]]) ; CHECK-NEXT: ret i16 [[TMP2]] ; - %tmp1 = tail call i16 @llvm.bswap.i16(i16 %a) - %tmp2 = tail call i16 @llvm.bswap.i16(i16 %b) - %tmp3 = xor i16 %tmp1, %tmp2 - ret i16 %tmp3 + %t1 = tail call i16 @llvm.bswap.i16(i16 %a) + %t2 = tail call i16 @llvm.bswap.i16(i16 %b) + %t3 = xor i16 %t1, %t2 + ret i16 %t3 } define i32 @bs_and32i(i32 %a, i32 %b) #0 { ; CHECK-LABEL: @bs_and32i( -; CHECK-NEXT: [[TMP1:%.*]] = and i32 %a, -1585053440 +; CHECK-NEXT: [[TMP1:%.*]] = and i32 [[A:%.*]], -1585053440 ; CHECK-NEXT: [[TMP2:%.*]] = call i32 @llvm.bswap.i32(i32 [[TMP1]]) ; CHECK-NEXT: ret i32 [[TMP2]] ; - %tmp1 = tail call i32 @llvm.bswap.i32(i32 %a) - %tmp2 = and i32 %tmp1, 100001 - ret i32 %tmp2 + %t1 = tail call i32 @llvm.bswap.i32(i32 %a) + %t2 = and i32 %t1, 100001 + ret i32 %t2 } define i32 @bs_and32(i32 %a, i32 %b) #0 { ; CHECK-LABEL: @bs_and32( -; CHECK-NEXT: [[TMP1:%.*]] = and i32 %a, %b +; CHECK-NEXT: [[TMP1:%.*]] = and i32 [[A:%.*]], [[B:%.*]] ; CHECK-NEXT: [[TMP2:%.*]] = call i32 @llvm.bswap.i32(i32 [[TMP1]]) ; CHECK-NEXT: ret i32 [[TMP2]] ; - %tmp1 = tail call i32 @llvm.bswap.i32(i32 %a) - %tmp2 = tail call i32 @llvm.bswap.i32(i32 %b) - %tmp3 = and i32 %tmp1, %tmp2 - ret i32 %tmp3 + %t1 = tail call i32 @llvm.bswap.i32(i32 %a) + %t2 = tail call i32 @llvm.bswap.i32(i32 %b) + %t3 = and i32 %t1, %t2 + ret i32 %t3 } define i32 @bs_or32(i32 %a, i32 %b) #0 { ; CHECK-LABEL: @bs_or32( -; CHECK-NEXT: [[TMP1:%.*]] = or i32 %a, %b +; CHECK-NEXT: [[TMP1:%.*]] = or i32 [[A:%.*]], [[B:%.*]] ; CHECK-NEXT: [[TMP2:%.*]] = call i32 @llvm.bswap.i32(i32 [[TMP1]]) ; CHECK-NEXT: ret i32 [[TMP2]] ; - %tmp1 = tail call i32 @llvm.bswap.i32(i32 %a) - %tmp2 = tail call i32 @llvm.bswap.i32(i32 %b) - %tmp3 = or i32 %tmp1, %tmp2 - ret i32 %tmp3 + %t1 = tail call i32 @llvm.bswap.i32(i32 %a) + %t2 = tail call i32 @llvm.bswap.i32(i32 %b) + %t3 = or i32 %t1, %t2 + ret i32 %t3 } define i32 @bs_xor32(i32 %a, i32 %b) #0 { ; CHECK-LABEL: @bs_xor32( -; CHECK-NEXT: [[TMP1:%.*]] = xor i32 %a, %b +; CHECK-NEXT: [[TMP1:%.*]] = xor i32 [[A:%.*]], [[B:%.*]] ; CHECK-NEXT: [[TMP2:%.*]] = call i32 @llvm.bswap.i32(i32 [[TMP1]]) ; CHECK-NEXT: ret i32 [[TMP2]] ; - %tmp1 = tail call i32 @llvm.bswap.i32(i32 %a) - %tmp2 = tail call i32 @llvm.bswap.i32(i32 %b) - %tmp3 = xor i32 %tmp1, %tmp2 - ret i32 %tmp3 + %t1 = tail call i32 @llvm.bswap.i32(i32 %a) + %t2 = tail call i32 @llvm.bswap.i32(i32 %b) + %t3 = xor i32 %t1, %t2 + ret i32 %t3 } define i64 @bs_and64i(i64 %a, i64 %b) #0 { ; CHECK-LABEL: @bs_and64i( -; CHECK-NEXT: [[TMP1:%.*]] = and i64 %a, 129085117527228416 +; CHECK-NEXT: [[TMP1:%.*]] = and i64 [[A:%.*]], 129085117527228416 ; CHECK-NEXT: [[TMP2:%.*]] = call i64 @llvm.bswap.i64(i64 [[TMP1]]) ; CHECK-NEXT: ret i64 [[TMP2]] ; - %tmp1 = tail call i64 @llvm.bswap.i64(i64 %a) - %tmp2 = and i64 %tmp1, 1000000001 - ret i64 %tmp2 + %t1 = tail call i64 @llvm.bswap.i64(i64 %a) + %t2 = and i64 %t1, 1000000001 + ret i64 %t2 } define i64 @bs_and64(i64 %a, i64 %b) #0 { ; CHECK-LABEL: @bs_and64( -; CHECK-NEXT: [[TMP1:%.*]] = and i64 %a, %b +; CHECK-NEXT: [[TMP1:%.*]] = and i64 [[A:%.*]], [[B:%.*]] ; CHECK-NEXT: [[TMP2:%.*]] = call i64 @llvm.bswap.i64(i64 [[TMP1]]) ; CHECK-NEXT: ret i64 [[TMP2]] ; - %tmp1 = tail call i64 @llvm.bswap.i64(i64 %a) - %tmp2 = tail call i64 @llvm.bswap.i64(i64 %b) - %tmp3 = and i64 %tmp1, %tmp2 - ret i64 %tmp3 + %t1 = tail call i64 @llvm.bswap.i64(i64 %a) + %t2 = tail call i64 @llvm.bswap.i64(i64 %b) + %t3 = and i64 %t1, %t2 + ret i64 %t3 } define i64 @bs_or64(i64 %a, i64 %b) #0 { ; CHECK-LABEL: @bs_or64( -; CHECK-NEXT: [[TMP1:%.*]] = or i64 %a, %b +; CHECK-NEXT: [[TMP1:%.*]] = or i64 [[A:%.*]], [[B:%.*]] ; CHECK-NEXT: [[TMP2:%.*]] = call i64 @llvm.bswap.i64(i64 [[TMP1]]) ; CHECK-NEXT: ret i64 [[TMP2]] ; - %tmp1 = tail call i64 @llvm.bswap.i64(i64 %a) - %tmp2 = tail call i64 @llvm.bswap.i64(i64 %b) - %tmp3 = or i64 %tmp1, %tmp2 - ret i64 %tmp3 + %t1 = tail call i64 @llvm.bswap.i64(i64 %a) + %t2 = tail call i64 @llvm.bswap.i64(i64 %b) + %t3 = or i64 %t1, %t2 + ret i64 %t3 } define i64 @bs_xor64(i64 %a, i64 %b) #0 { ; CHECK-LABEL: @bs_xor64( -; CHECK-NEXT: [[TMP1:%.*]] = xor i64 %a, %b +; CHECK-NEXT: [[TMP1:%.*]] = xor i64 [[A:%.*]], [[B:%.*]] ; CHECK-NEXT: [[TMP2:%.*]] = call i64 @llvm.bswap.i64(i64 [[TMP1]]) ; CHECK-NEXT: ret i64 [[TMP2]] ; - %tmp1 = tail call i64 @llvm.bswap.i64(i64 %a) - %tmp2 = tail call i64 @llvm.bswap.i64(i64 %b) - %tmp3 = xor i64 %tmp1, %tmp2 - ret i64 %tmp3 + %t1 = tail call i64 @llvm.bswap.i64(i64 %a) + %t2 = tail call i64 @llvm.bswap.i64(i64 %b) + %t3 = xor i64 %t1, %t2 + ret i64 %t3 } define <2 x i32> @bs_and32vec(<2 x i32> %a, <2 x i32> %b) #0 { @@ -208,10 +208,10 @@ define <2 x i32> @bs_and32vec(<2 x i32> %a, <2 x i32> %b) #0 { ; CHECK-NEXT: [[TMP2:%.*]] = call <2 x i32> @llvm.bswap.v2i32(<2 x i32> [[TMP1]]) ; CHECK-NEXT: ret <2 x i32> [[TMP2]] ; - %tmp1 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %a) - %tmp2 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %b) - %tmp3 = and <2 x i32> %tmp1, %tmp2 - ret <2 x i32> %tmp3 + %t1 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %a) + %t2 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %b) + %t3 = and <2 x i32> %t1, %t2 + ret <2 x i32> %t3 } define <2 x i32> @bs_or32vec(<2 x i32> %a, <2 x i32> %b) #0 { @@ -220,10 +220,10 @@ define <2 x i32> @bs_or32vec(<2 x i32> %a, <2 x i32> %b) #0 { ; CHECK-NEXT: [[TMP2:%.*]] = call <2 x i32> @llvm.bswap.v2i32(<2 x i32> [[TMP1]]) ; CHECK-NEXT: ret <2 x i32> [[TMP2]] ; - %tmp1 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %a) - %tmp2 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %b) - %tmp3 = or <2 x i32> %tmp1, %tmp2 - ret <2 x i32> %tmp3 + %t1 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %a) + %t2 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %b) + %t3 = or <2 x i32> %t1, %t2 + ret <2 x i32> %t3 } define <2 x i32> @bs_xor32vec(<2 x i32> %a, <2 x i32> %b) #0 { @@ -232,10 +232,10 @@ define <2 x i32> @bs_xor32vec(<2 x i32> %a, <2 x i32> %b) #0 { ; CHECK-NEXT: [[TMP2:%.*]] = call <2 x i32> @llvm.bswap.v2i32(<2 x i32> [[TMP1]]) ; CHECK-NEXT: ret <2 x i32> [[TMP2]] ; - %tmp1 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %a) - %tmp2 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %b) - %tmp3 = xor <2 x i32> %tmp1, %tmp2 - ret <2 x i32> %tmp3 + %t1 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %a) + %t2 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %b) + %t3 = xor <2 x i32> %t1, %t2 + ret <2 x i32> %t3 } define <2 x i32> @bs_and32ivec(<2 x i32> %a, <2 x i32> %b) #0 { @@ -244,9 +244,9 @@ define <2 x i32> @bs_and32ivec(<2 x i32> %a, <2 x i32> %b) #0 { ; CHECK-NEXT: [[TMP2:%.*]] = call <2 x i32> @llvm.bswap.v2i32(<2 x i32> [[TMP1]]) ; CHECK-NEXT: ret <2 x i32> [[TMP2]] ; - %tmp1 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %a) - %tmp2 = and <2 x i32> %tmp1, - ret <2 x i32> %tmp2 + %t1 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %a) + %t2 = and <2 x i32> %t1, + ret <2 x i32> %t2 } define <2 x i32> @bs_or32ivec(<2 x i32> %a, <2 x i32> %b) #0 { @@ -255,9 +255,9 @@ define <2 x i32> @bs_or32ivec(<2 x i32> %a, <2 x i32> %b) #0 { ; CHECK-NEXT: [[TMP2:%.*]] = call <2 x i32> @llvm.bswap.v2i32(<2 x i32> [[TMP1]]) ; CHECK-NEXT: ret <2 x i32> [[TMP2]] ; - %tmp1 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %a) - %tmp2 = or <2 x i32> %tmp1, - ret <2 x i32> %tmp2 + %t1 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %a) + %t2 = or <2 x i32> %t1, + ret <2 x i32> %t2 } define <2 x i32> @bs_xor32ivec(<2 x i32> %a, <2 x i32> %b) #0 { @@ -266,69 +266,69 @@ define <2 x i32> @bs_xor32ivec(<2 x i32> %a, <2 x i32> %b) #0 { ; CHECK-NEXT: [[TMP2:%.*]] = call <2 x i32> @llvm.bswap.v2i32(<2 x i32> [[TMP1]]) ; CHECK-NEXT: ret <2 x i32> [[TMP2]] ; - %tmp1 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %a) - %tmp2 = xor <2 x i32> %tmp1, - ret <2 x i32> %tmp2 + %t1 = tail call <2 x i32> @llvm.bswap.v2i32(<2 x i32> %a) + %t2 = xor <2 x i32> %t1, + ret <2 x i32> %t2 } define i64 @bs_and64_multiuse1(i64 %a, i64 %b) #0 { ; CHECK-LABEL: @bs_and64_multiuse1( -; CHECK-NEXT: [[TMP1:%.*]] = tail call i64 @llvm.bswap.i64(i64 [[A:%.*]]) -; CHECK-NEXT: [[TMP2:%.*]] = tail call i64 @llvm.bswap.i64(i64 [[B:%.*]]) -; CHECK-NEXT: [[TMP3:%.*]] = and i64 [[TMP1]], [[TMP2]] -; CHECK-NEXT: [[TMP4:%.*]] = mul i64 [[TMP3]], [[TMP1]] -; CHECK-NEXT: [[TMP5:%.*]] = mul i64 [[TMP4]], [[TMP2]] -; CHECK-NEXT: ret i64 [[TMP5]] +; CHECK-NEXT: [[T1:%.*]] = tail call i64 @llvm.bswap.i64(i64 [[A:%.*]]) +; CHECK-NEXT: [[T2:%.*]] = tail call i64 @llvm.bswap.i64(i64 [[B:%.*]]) +; CHECK-NEXT: [[T3:%.*]] = and i64 [[T1]], [[T2]] +; CHECK-NEXT: [[T4:%.*]] = mul i64 [[T3]], [[T1]] +; CHECK-NEXT: [[T5:%.*]] = mul i64 [[T4]], [[T2]] +; CHECK-NEXT: ret i64 [[T5]] ; - %tmp1 = tail call i64 @llvm.bswap.i64(i64 %a) - %tmp2 = tail call i64 @llvm.bswap.i64(i64 %b) - %tmp3 = and i64 %tmp1, %tmp2 - %tmp4 = mul i64 %tmp3, %tmp1 ; to increase use count of the bswaps - %tmp5 = mul i64 %tmp4, %tmp2 ; to increase use count of the bswaps - ret i64 %tmp5 + %t1 = tail call i64 @llvm.bswap.i64(i64 %a) + %t2 = tail call i64 @llvm.bswap.i64(i64 %b) + %t3 = and i64 %t1, %t2 + %t4 = mul i64 %t3, %t1 ; to increase use count of the bswaps + %t5 = mul i64 %t4, %t2 ; to increase use count of the bswaps + ret i64 %t5 } define i64 @bs_and64_multiuse2(i64 %a, i64 %b) #0 { ; CHECK-LABEL: @bs_and64_multiuse2( -; CHECK-NEXT: [[TMP1:%.*]] = tail call i64 @llvm.bswap.i64(i64 [[A:%.*]]) -; CHECK-NEXT: [[TMP2:%.*]] = and i64 [[A]], [[B:%.*]] -; CHECK-NEXT: [[TMP3:%.*]] = call i64 @llvm.bswap.i64(i64 [[TMP2]]) -; CHECK-NEXT: [[TMP4:%.*]] = mul i64 [[TMP3]], [[TMP1]] -; CHECK-NEXT: ret i64 [[TMP4]] +; CHECK-NEXT: [[T1:%.*]] = tail call i64 @llvm.bswap.i64(i64 [[A:%.*]]) +; CHECK-NEXT: [[TMP1:%.*]] = and i64 [[A]], [[B:%.*]] +; CHECK-NEXT: [[TMP2:%.*]] = call i64 @llvm.bswap.i64(i64 [[TMP1]]) +; CHECK-NEXT: [[T4:%.*]] = mul i64 [[TMP2]], [[T1]] +; CHECK-NEXT: ret i64 [[T4]] ; - %tmp1 = tail call i64 @llvm.bswap.i64(i64 %a) - %tmp2 = tail call i64 @llvm.bswap.i64(i64 %b) - %tmp3 = and i64 %tmp1, %tmp2 - %tmp4 = mul i64 %tmp3, %tmp1 ; to increase use count of the bswaps - ret i64 %tmp4 + %t1 = tail call i64 @llvm.bswap.i64(i64 %a) + %t2 = tail call i64 @llvm.bswap.i64(i64 %b) + %t3 = and i64 %t1, %t2 + %t4 = mul i64 %t3, %t1 ; to increase use count of the bswaps + ret i64 %t4 } define i64 @bs_and64_multiuse3(i64 %a, i64 %b) #0 { ; CHECK-LABEL: @bs_and64_multiuse3( -; CHECK-NEXT: [[TMP2:%.*]] = tail call i64 @llvm.bswap.i64(i64 [[B:%.*]]) +; CHECK-NEXT: [[T2:%.*]] = tail call i64 @llvm.bswap.i64(i64 [[B:%.*]]) ; CHECK-NEXT: [[TMP1:%.*]] = and i64 [[A:%.*]], [[B]] -; CHECK-NEXT: [[TMP3:%.*]] = call i64 @llvm.bswap.i64(i64 [[TMP1]]) -; CHECK-NEXT: [[TMP4:%.*]] = mul i64 [[TMP3]], [[TMP2]] -; CHECK-NEXT: ret i64 [[TMP4]] +; CHECK-NEXT: [[TMP2:%.*]] = call i64 @llvm.bswap.i64(i64 [[TMP1]]) +; CHECK-NEXT: [[T4:%.*]] = mul i64 [[TMP2]], [[T2]] +; CHECK-NEXT: ret i64 [[T4]] ; - %tmp1 = tail call i64 @llvm.bswap.i64(i64 %a) - %tmp2 = tail call i64 @llvm.bswap.i64(i64 %b) - %tmp3 = and i64 %tmp1, %tmp2 - %tmp4 = mul i64 %tmp3, %tmp2 ; to increase use count of the bswaps - ret i64 %tmp4 + %t1 = tail call i64 @llvm.bswap.i64(i64 %a) + %t2 = tail call i64 @llvm.bswap.i64(i64 %b) + %t3 = and i64 %t1, %t2 + %t4 = mul i64 %t3, %t2 ; to increase use count of the bswaps + ret i64 %t4 } define i64 @bs_and64i_multiuse(i64 %a, i64 %b) #0 { ; CHECK-LABEL: @bs_and64i_multiuse( -; CHECK-NEXT: [[TMP1:%.*]] = tail call i64 @llvm.bswap.i64(i64 [[A:%.*]]) -; CHECK-NEXT: [[TMP2:%.*]] = and i64 [[TMP1]], 1000000001 -; CHECK-NEXT: [[TMP3:%.*]] = mul i64 [[TMP2]], [[TMP1]] -; CHECK-NEXT: ret i64 [[TMP3]] +; CHECK-NEXT: [[T1:%.*]] = tail call i64 @llvm.bswap.i64(i64 [[A:%.*]]) +; CHECK-NEXT: [[T2:%.*]] = and i64 [[T1]], 1000000001 +; CHECK-NEXT: [[T3:%.*]] = mul i64 [[T2]], [[T1]] +; CHECK-NEXT: ret i64 [[T3]] ; - %tmp1 = tail call i64 @llvm.bswap.i64(i64 %a) - %tmp2 = and i64 %tmp1, 1000000001 - %tmp3 = mul i64 %tmp2, %tmp1 ; to increase use count of the bswap - ret i64 %tmp3 + %t1 = tail call i64 @llvm.bswap.i64(i64 %a) + %t2 = and i64 %t1, 1000000001 + %t3 = mul i64 %t2, %t1 ; to increase use count of the bswap + ret i64 %t3 } declare i16 @llvm.bswap.i16(i16)