74 lines
2.6 KiB
LLVM
74 lines
2.6 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc --mtriple=loongarch32 --verify-machineinstrs < %s \
|
|
; RUN: | FileCheck %s --check-prefix=LA32
|
|
; RUN: llc --mtriple=loongarch64 --verify-machineinstrs < %s \
|
|
; RUN: | FileCheck %s --check-prefix=LA64
|
|
|
|
declare void @callee(i8*, i32*)
|
|
|
|
define void @caller(i32 %n) {
|
|
; LA32-LABEL: caller:
|
|
; LA32: # %bb.0:
|
|
; LA32-NEXT: addi.w $sp, $sp, -64
|
|
; LA32-NEXT: .cfi_def_cfa_offset 64
|
|
; LA32-NEXT: st.w $ra, $sp, 60 # 4-byte Folded Spill
|
|
; LA32-NEXT: st.w $fp, $sp, 56 # 4-byte Folded Spill
|
|
; LA32-NEXT: st.w $s8, $sp, 52 # 4-byte Folded Spill
|
|
; LA32-NEXT: .cfi_offset 1, -4
|
|
; LA32-NEXT: .cfi_offset 22, -8
|
|
; LA32-NEXT: .cfi_offset 31, -12
|
|
; LA32-NEXT: addi.w $fp, $sp, 64
|
|
; LA32-NEXT: .cfi_def_cfa 22, 0
|
|
; LA32-NEXT: srli.w $a1, $sp, 6
|
|
; LA32-NEXT: slli.w $sp, $a1, 6
|
|
; LA32-NEXT: move $s8, $sp
|
|
; LA32-NEXT: addi.w $a0, $a0, 15
|
|
; LA32-NEXT: addi.w $a1, $zero, -16
|
|
; LA32-NEXT: and $a0, $a0, $a1
|
|
; LA32-NEXT: sub.w $a0, $sp, $a0
|
|
; LA32-NEXT: move $sp, $a0
|
|
; LA32-NEXT: addi.w $a1, $s8, 0
|
|
; LA32-NEXT: bl %plt(callee)
|
|
; LA32-NEXT: addi.w $sp, $fp, -64
|
|
; LA32-NEXT: ld.w $s8, $sp, 52 # 4-byte Folded Reload
|
|
; LA32-NEXT: ld.w $fp, $sp, 56 # 4-byte Folded Reload
|
|
; LA32-NEXT: ld.w $ra, $sp, 60 # 4-byte Folded Reload
|
|
; LA32-NEXT: addi.w $sp, $sp, 64
|
|
; LA32-NEXT: ret
|
|
;
|
|
; LA64-LABEL: caller:
|
|
; LA64: # %bb.0:
|
|
; LA64-NEXT: addi.d $sp, $sp, -64
|
|
; LA64-NEXT: .cfi_def_cfa_offset 64
|
|
; LA64-NEXT: st.d $ra, $sp, 56 # 8-byte Folded Spill
|
|
; LA64-NEXT: st.d $fp, $sp, 48 # 8-byte Folded Spill
|
|
; LA64-NEXT: st.d $s8, $sp, 40 # 8-byte Folded Spill
|
|
; LA64-NEXT: .cfi_offset 1, -8
|
|
; LA64-NEXT: .cfi_offset 22, -16
|
|
; LA64-NEXT: .cfi_offset 31, -24
|
|
; LA64-NEXT: addi.d $fp, $sp, 64
|
|
; LA64-NEXT: .cfi_def_cfa 22, 0
|
|
; LA64-NEXT: srli.d $a1, $sp, 6
|
|
; LA64-NEXT: slli.d $sp, $a1, 6
|
|
; LA64-NEXT: move $s8, $sp
|
|
; LA64-NEXT: addi.w $a1, $zero, -16
|
|
; LA64-NEXT: lu32i.d $a1, 1
|
|
; LA64-NEXT: bstrpick.d $a0, $a0, 31, 0
|
|
; LA64-NEXT: addi.d $a0, $a0, 15
|
|
; LA64-NEXT: and $a0, $a0, $a1
|
|
; LA64-NEXT: sub.d $a0, $sp, $a0
|
|
; LA64-NEXT: move $sp, $a0
|
|
; LA64-NEXT: addi.d $a1, $s8, 0
|
|
; LA64-NEXT: bl %plt(callee)
|
|
; LA64-NEXT: addi.d $sp, $fp, -64
|
|
; LA64-NEXT: ld.d $s8, $sp, 40 # 8-byte Folded Reload
|
|
; LA64-NEXT: ld.d $fp, $sp, 48 # 8-byte Folded Reload
|
|
; LA64-NEXT: ld.d $ra, $sp, 56 # 8-byte Folded Reload
|
|
; LA64-NEXT: addi.d $sp, $sp, 64
|
|
; LA64-NEXT: ret
|
|
%1 = alloca i8, i32 %n
|
|
%2 = alloca i32, align 64
|
|
call void @callee(i8* %1, i32 *%2)
|
|
ret void
|
|
}
|