134 changes: 76 additions & 58 deletions llvm/test/CodeGen/LoongArch/stack-realignment.ll
Original file line number Diff line number Diff line change
Expand Up @@ -305,42 +305,42 @@ define void @caller_no_realign256() "no-realign-stack" {
define void @caller512() {
; LA32-LABEL: caller512:
; LA32: # %bb.0:
; LA32-NEXT: addi.w $sp, $sp, -512
; LA32-NEXT: .cfi_def_cfa_offset 512
; LA32-NEXT: st.w $ra, $sp, 508 # 4-byte Folded Spill
; LA32-NEXT: st.w $fp, $sp, 504 # 4-byte Folded Spill
; LA32-NEXT: addi.w $sp, $sp, -1024
; LA32-NEXT: .cfi_def_cfa_offset 1024
; LA32-NEXT: st.w $ra, $sp, 1020 # 4-byte Folded Spill
; LA32-NEXT: st.w $fp, $sp, 1016 # 4-byte Folded Spill
; LA32-NEXT: .cfi_offset 1, -4
; LA32-NEXT: .cfi_offset 22, -8
; LA32-NEXT: addi.w $fp, $sp, 512
; LA32-NEXT: addi.w $fp, $sp, 1024
; LA32-NEXT: .cfi_def_cfa 22, 0
; LA32-NEXT: srli.w $a0, $sp, 9
; LA32-NEXT: slli.w $sp, $a0, 9
; LA32-NEXT: addi.w $a0, $sp, 0
; LA32-NEXT: addi.w $a0, $sp, 512
; LA32-NEXT: bl %plt(callee)
; LA32-NEXT: addi.w $sp, $fp, -512
; LA32-NEXT: ld.w $fp, $sp, 504 # 4-byte Folded Reload
; LA32-NEXT: ld.w $ra, $sp, 508 # 4-byte Folded Reload
; LA32-NEXT: addi.w $sp, $sp, 512
; LA32-NEXT: addi.w $sp, $fp, -1024
; LA32-NEXT: ld.w $fp, $sp, 1016 # 4-byte Folded Reload
; LA32-NEXT: ld.w $ra, $sp, 1020 # 4-byte Folded Reload
; LA32-NEXT: addi.w $sp, $sp, 1024
; LA32-NEXT: ret
;
; LA64-LABEL: caller512:
; LA64: # %bb.0:
; LA64-NEXT: addi.d $sp, $sp, -512
; LA64-NEXT: .cfi_def_cfa_offset 512
; LA64-NEXT: st.d $ra, $sp, 504 # 8-byte Folded Spill
; LA64-NEXT: st.d $fp, $sp, 496 # 8-byte Folded Spill
; LA64-NEXT: addi.d $sp, $sp, -1024
; LA64-NEXT: .cfi_def_cfa_offset 1024
; LA64-NEXT: st.d $ra, $sp, 1016 # 8-byte Folded Spill
; LA64-NEXT: st.d $fp, $sp, 1008 # 8-byte Folded Spill
; LA64-NEXT: .cfi_offset 1, -8
; LA64-NEXT: .cfi_offset 22, -16
; LA64-NEXT: addi.d $fp, $sp, 512
; LA64-NEXT: addi.d $fp, $sp, 1024
; LA64-NEXT: .cfi_def_cfa 22, 0
; LA64-NEXT: srli.d $a0, $sp, 9
; LA64-NEXT: slli.d $sp, $a0, 9
; LA64-NEXT: addi.d $a0, $sp, 0
; LA64-NEXT: addi.d $a0, $sp, 512
; LA64-NEXT: bl %plt(callee)
; LA64-NEXT: addi.d $sp, $fp, -512
; LA64-NEXT: ld.d $fp, $sp, 496 # 8-byte Folded Reload
; LA64-NEXT: ld.d $ra, $sp, 504 # 8-byte Folded Reload
; LA64-NEXT: addi.d $sp, $sp, 512
; LA64-NEXT: addi.d $sp, $fp, -1024
; LA64-NEXT: ld.d $fp, $sp, 1008 # 8-byte Folded Reload
; LA64-NEXT: ld.d $ra, $sp, 1016 # 8-byte Folded Reload
; LA64-NEXT: addi.d $sp, $sp, 1024
; LA64-NEXT: ret
%1 = alloca i8, align 512
call void @callee(i8* %1)
Expand Down Expand Up @@ -379,42 +379,46 @@ define void @caller_no_realign512() "no-realign-stack" {
define void @caller1024() {
; LA32-LABEL: caller1024:
; LA32: # %bb.0:
; LA32-NEXT: addi.w $sp, $sp, -1024
; LA32-NEXT: .cfi_def_cfa_offset 1024
; LA32-NEXT: st.w $ra, $sp, 1020 # 4-byte Folded Spill
; LA32-NEXT: st.w $fp, $sp, 1016 # 4-byte Folded Spill
; LA32-NEXT: addi.w $sp, $sp, -2032
; LA32-NEXT: .cfi_def_cfa_offset 2032
; LA32-NEXT: st.w $ra, $sp, 2028 # 4-byte Folded Spill
; LA32-NEXT: st.w $fp, $sp, 2024 # 4-byte Folded Spill
; LA32-NEXT: .cfi_offset 1, -4
; LA32-NEXT: .cfi_offset 22, -8
; LA32-NEXT: addi.w $fp, $sp, 1024
; LA32-NEXT: addi.w $fp, $sp, 2032
; LA32-NEXT: .cfi_def_cfa 22, 0
; LA32-NEXT: addi.w $sp, $sp, -16
; LA32-NEXT: srli.w $a0, $sp, 10
; LA32-NEXT: slli.w $sp, $a0, 10
; LA32-NEXT: addi.w $a0, $sp, 0
; LA32-NEXT: addi.w $a0, $sp, 1024
; LA32-NEXT: bl %plt(callee)
; LA32-NEXT: addi.w $sp, $fp, -1024
; LA32-NEXT: ld.w $fp, $sp, 1016 # 4-byte Folded Reload
; LA32-NEXT: ld.w $ra, $sp, 1020 # 4-byte Folded Reload
; LA32-NEXT: addi.w $sp, $sp, 1024
; LA32-NEXT: addi.w $sp, $fp, -2048
; LA32-NEXT: addi.w $sp, $sp, 16
; LA32-NEXT: ld.w $fp, $sp, 2024 # 4-byte Folded Reload
; LA32-NEXT: ld.w $ra, $sp, 2028 # 4-byte Folded Reload
; LA32-NEXT: addi.w $sp, $sp, 2032
; LA32-NEXT: ret
;
; LA64-LABEL: caller1024:
; LA64: # %bb.0:
; LA64-NEXT: addi.d $sp, $sp, -1024
; LA64-NEXT: .cfi_def_cfa_offset 1024
; LA64-NEXT: st.d $ra, $sp, 1016 # 8-byte Folded Spill
; LA64-NEXT: st.d $fp, $sp, 1008 # 8-byte Folded Spill
; LA64-NEXT: addi.d $sp, $sp, -2032
; LA64-NEXT: .cfi_def_cfa_offset 2032
; LA64-NEXT: st.d $ra, $sp, 2024 # 8-byte Folded Spill
; LA64-NEXT: st.d $fp, $sp, 2016 # 8-byte Folded Spill
; LA64-NEXT: .cfi_offset 1, -8
; LA64-NEXT: .cfi_offset 22, -16
; LA64-NEXT: addi.d $fp, $sp, 1024
; LA64-NEXT: addi.d $fp, $sp, 2032
; LA64-NEXT: .cfi_def_cfa 22, 0
; LA64-NEXT: addi.d $sp, $sp, -16
; LA64-NEXT: srli.d $a0, $sp, 10
; LA64-NEXT: slli.d $sp, $a0, 10
; LA64-NEXT: addi.d $a0, $sp, 0
; LA64-NEXT: addi.d $a0, $sp, 1024
; LA64-NEXT: bl %plt(callee)
; LA64-NEXT: addi.d $sp, $fp, -1024
; LA64-NEXT: ld.d $fp, $sp, 1008 # 8-byte Folded Reload
; LA64-NEXT: ld.d $ra, $sp, 1016 # 8-byte Folded Reload
; LA64-NEXT: addi.d $sp, $sp, 1024
; LA64-NEXT: addi.d $sp, $fp, -2048
; LA64-NEXT: addi.d $sp, $sp, 16
; LA64-NEXT: ld.d $fp, $sp, 2016 # 8-byte Folded Reload
; LA64-NEXT: ld.d $ra, $sp, 2024 # 8-byte Folded Reload
; LA64-NEXT: addi.d $sp, $sp, 2032
; LA64-NEXT: ret
%1 = alloca i8, align 1024
call void @callee(i8* %1)
Expand Down Expand Up @@ -461,13 +465,17 @@ define void @caller2048() {
; LA32-NEXT: .cfi_offset 22, -8
; LA32-NEXT: addi.w $fp, $sp, 2032
; LA32-NEXT: .cfi_def_cfa 22, 0
; LA32-NEXT: addi.w $sp, $sp, -2048
; LA32-NEXT: addi.w $sp, $sp, -16
; LA32-NEXT: srli.w $a0, $sp, 11
; LA32-NEXT: slli.w $sp, $a0, 11
; LA32-NEXT: addi.w $a0, $sp, 0
; LA32-NEXT: ori $a0, $zero, 2048
; LA32-NEXT: add.w $a0, $sp, $a0
; LA32-NEXT: bl %plt(callee)
; LA32-NEXT: addi.w $sp, $fp, -2048
; LA32-NEXT: addi.w $sp, $sp, 16
; LA32-NEXT: lu12i.w $a0, 1
; LA32-NEXT: sub.w $sp, $fp, $a0
; LA32-NEXT: addi.w $sp, $sp, 2032
; LA32-NEXT: addi.w $sp, $sp, 32
; LA32-NEXT: ld.w $fp, $sp, 2024 # 4-byte Folded Reload
; LA32-NEXT: ld.w $ra, $sp, 2028 # 4-byte Folded Reload
; LA32-NEXT: addi.w $sp, $sp, 2032
Expand All @@ -483,13 +491,17 @@ define void @caller2048() {
; LA64-NEXT: .cfi_offset 22, -16
; LA64-NEXT: addi.d $fp, $sp, 2032
; LA64-NEXT: .cfi_def_cfa 22, 0
; LA64-NEXT: addi.d $sp, $sp, -2048
; LA64-NEXT: addi.d $sp, $sp, -16
; LA64-NEXT: srli.d $a0, $sp, 11
; LA64-NEXT: slli.d $sp, $a0, 11
; LA64-NEXT: addi.d $a0, $sp, 0
; LA64-NEXT: ori $a0, $zero, 2048
; LA64-NEXT: add.d $a0, $sp, $a0
; LA64-NEXT: bl %plt(callee)
; LA64-NEXT: addi.d $sp, $fp, -2048
; LA64-NEXT: addi.d $sp, $sp, 16
; LA64-NEXT: lu12i.w $a0, 1
; LA64-NEXT: sub.d $sp, $fp, $a0
; LA64-NEXT: addi.d $sp, $sp, 2032
; LA64-NEXT: addi.d $sp, $sp, 32
; LA64-NEXT: ld.d $fp, $sp, 2016 # 8-byte Folded Reload
; LA64-NEXT: ld.d $ra, $sp, 2024 # 8-byte Folded Reload
; LA64-NEXT: addi.d $sp, $sp, 2032
Expand Down Expand Up @@ -539,16 +551,19 @@ define void @caller4096() {
; LA32-NEXT: .cfi_offset 22, -8
; LA32-NEXT: addi.w $fp, $sp, 2032
; LA32-NEXT: .cfi_def_cfa 22, 0
; LA32-NEXT: addi.w $sp, $sp, -2048
; LA32-NEXT: addi.w $sp, $sp, -16
; LA32-NEXT: lu12i.w $a0, 1
; LA32-NEXT: ori $a0, $a0, 2064
; LA32-NEXT: sub.w $sp, $sp, $a0
; LA32-NEXT: srli.w $a0, $sp, 12
; LA32-NEXT: slli.w $sp, $a0, 12
; LA32-NEXT: addi.w $a0, $sp, 0
; LA32-NEXT: bl %plt(callee)
; LA32-NEXT: lu12i.w $a0, 1
; LA32-NEXT: add.w $a0, $sp, $a0
; LA32-NEXT: bl %plt(callee)
; LA32-NEXT: lu12i.w $a0, 2
; LA32-NEXT: sub.w $sp, $fp, $a0
; LA32-NEXT: addi.w $sp, $sp, 2032
; LA32-NEXT: addi.w $sp, $sp, 32
; LA32-NEXT: lu12i.w $a0, 1
; LA32-NEXT: ori $a0, $a0, 2064
; LA32-NEXT: add.w $sp, $sp, $a0
; LA32-NEXT: ld.w $fp, $sp, 2024 # 4-byte Folded Reload
; LA32-NEXT: ld.w $ra, $sp, 2028 # 4-byte Folded Reload
; LA32-NEXT: addi.w $sp, $sp, 2032
Expand All @@ -564,16 +579,19 @@ define void @caller4096() {
; LA64-NEXT: .cfi_offset 22, -16
; LA64-NEXT: addi.d $fp, $sp, 2032
; LA64-NEXT: .cfi_def_cfa 22, 0
; LA64-NEXT: addi.d $sp, $sp, -2048
; LA64-NEXT: addi.d $sp, $sp, -16
; LA64-NEXT: lu12i.w $a0, 1
; LA64-NEXT: ori $a0, $a0, 2064
; LA64-NEXT: sub.d $sp, $sp, $a0
; LA64-NEXT: srli.d $a0, $sp, 12
; LA64-NEXT: slli.d $sp, $a0, 12
; LA64-NEXT: addi.d $a0, $sp, 0
; LA64-NEXT: bl %plt(callee)
; LA64-NEXT: lu12i.w $a0, 1
; LA64-NEXT: add.d $a0, $sp, $a0
; LA64-NEXT: bl %plt(callee)
; LA64-NEXT: lu12i.w $a0, 2
; LA64-NEXT: sub.d $sp, $fp, $a0
; LA64-NEXT: addi.d $sp, $sp, 2032
; LA64-NEXT: addi.d $sp, $sp, 32
; LA64-NEXT: lu12i.w $a0, 1
; LA64-NEXT: ori $a0, $a0, 2064
; LA64-NEXT: add.d $sp, $sp, $a0
; LA64-NEXT: ld.d $fp, $sp, 2016 # 8-byte Folded Reload
; LA64-NEXT: ld.d $ra, $sp, 2024 # 8-byte Folded Reload
; LA64-NEXT: addi.d $sp, $sp, 2032
Expand Down