58 lines
2.7 KiB
LLVM
58 lines
2.7 KiB
LLVM
; RUN: opt -mtriple=amdgcn-- -amdgpu-use-legacy-divergence-analysis -passes='print<divergence>' 2>&1 -disable-output %s | FileCheck %s
|
|
|
|
; CHECK: DIVERGENT: %orig = atomicrmw xchg ptr %ptr, i32 %val seq_cst
|
|
define amdgpu_kernel void @test1(ptr %ptr, i32 %val) #0 {
|
|
%orig = atomicrmw xchg ptr %ptr, i32 %val seq_cst
|
|
store i32 %orig, ptr %ptr
|
|
ret void
|
|
}
|
|
|
|
; CHECK: DIVERGENT: %orig = cmpxchg ptr %ptr, i32 %cmp, i32 %new seq_cst seq_cst
|
|
define amdgpu_kernel void @test2(ptr %ptr, i32 %cmp, i32 %new) {
|
|
%orig = cmpxchg ptr %ptr, i32 %cmp, i32 %new seq_cst seq_cst
|
|
%val = extractvalue { i32, i1 } %orig, 0
|
|
store i32 %val, ptr %ptr
|
|
ret void
|
|
}
|
|
|
|
; CHECK: DIVERGENT: %ret = call i32 @llvm.amdgcn.atomic.inc.i32.p1(ptr addrspace(1) %ptr, i32 %val, i32 0, i32 0, i1 false)
|
|
define i32 @test_atomic_inc_i32(ptr addrspace(1) %ptr, i32 %val) #0 {
|
|
%ret = call i32 @llvm.amdgcn.atomic.inc.i32.p1(ptr addrspace(1) %ptr, i32 %val, i32 0, i32 0, i1 false)
|
|
ret i32 %ret
|
|
}
|
|
|
|
; CHECK: DIVERGENT: %ret = call i64 @llvm.amdgcn.atomic.inc.i64.p1(ptr addrspace(1) %ptr, i64 %val, i32 0, i32 0, i1 false)
|
|
define i64 @test_atomic_inc_i64(ptr addrspace(1) %ptr, i64 %val) #0 {
|
|
%ret = call i64 @llvm.amdgcn.atomic.inc.i64.p1(ptr addrspace(1) %ptr, i64 %val, i32 0, i32 0, i1 false)
|
|
ret i64 %ret
|
|
}
|
|
|
|
; CHECK: DIVERGENT: %ret = call i32 @llvm.amdgcn.atomic.dec.i32.p1(ptr addrspace(1) %ptr, i32 %val, i32 0, i32 0, i1 false)
|
|
define i32 @test_atomic_dec_i32(ptr addrspace(1) %ptr, i32 %val) #0 {
|
|
%ret = call i32 @llvm.amdgcn.atomic.dec.i32.p1(ptr addrspace(1) %ptr, i32 %val, i32 0, i32 0, i1 false)
|
|
ret i32 %ret
|
|
}
|
|
|
|
; CHECK: DIVERGENT: %ret = call i64 @llvm.amdgcn.atomic.dec.i64.p1(ptr addrspace(1) %ptr, i64 %val, i32 0, i32 0, i1 false)
|
|
define i64 @test_atomic_dec_i64(ptr addrspace(1) %ptr, i64 %val) #0 {
|
|
%ret = call i64 @llvm.amdgcn.atomic.dec.i64.p1(ptr addrspace(1) %ptr, i64 %val, i32 0, i32 0, i1 false)
|
|
ret i64 %ret
|
|
}
|
|
|
|
declare i32 @llvm.amdgcn.atomic.inc.i32.p1(ptr addrspace(1) nocapture, i32, i32, i32, i1) #1
|
|
declare i64 @llvm.amdgcn.atomic.inc.i64.p1(ptr addrspace(1) nocapture, i64, i32, i32, i1) #1
|
|
declare i32 @llvm.amdgcn.atomic.dec.i32.p1(ptr addrspace(1) nocapture, i32, i32, i32, i1) #1
|
|
declare i64 @llvm.amdgcn.atomic.dec.i64.p1(ptr addrspace(1) nocapture, i64, i32, i32, i1) #1
|
|
|
|
; CHECK: DIVERGENT: %ret = call i32 @llvm.amdgcn.global.atomic.csub.p1(ptr addrspace(1) %ptr, i32 %val)
|
|
define amdgpu_kernel void @test_atomic_csub_i32(ptr addrspace(1) %ptr, i32 %val) #0 {
|
|
%ret = call i32 @llvm.amdgcn.global.atomic.csub.p1(ptr addrspace(1) %ptr, i32 %val)
|
|
store i32 %ret, ptr addrspace(1) %ptr, align 4
|
|
ret void
|
|
}
|
|
|
|
declare i32 @llvm.amdgcn.global.atomic.csub.p1(ptr addrspace(1) nocapture, i32) #1
|
|
|
|
attributes #0 = { nounwind }
|
|
attributes #1 = { argmemonly nounwind willreturn }
|