xref: /llvm-project/clang/test/OpenMP/amdgpu-unsafe-fp-atomics.cpp (revision e108853ac8fad27ff22be9303c87d90bcdf0ef53)
1*e108853aSMatt Arsenault // NOTE: Assertions have been autogenerated by utils/update_cc_test_checks.py UTC_ARGS: --version 5
2*e108853aSMatt Arsenault // RUN: %clang_cc1 -fopenmp -x c++ -std=c++11 -triple amdgcn-amd-amdhsa -fopenmp-targets=amdgcn-amd-amdhsa -emit-llvm %s -fopenmp-is-target-device -o - | FileCheck -check-prefix=DEFAULT %s
3*e108853aSMatt Arsenault // RUN: %clang_cc1 -fopenmp -x c++ -std=c++11 -triple amdgcn-amd-amdhsa -fopenmp-targets=amdgcn-amd-amdhsa -munsafe-fp-atomics -emit-llvm %s -fopenmp-is-target-device -o - | FileCheck -check-prefix=UNSAFE-FP-ATOMICS %s
4*e108853aSMatt Arsenault 
5*e108853aSMatt Arsenault #pragma omp declare target
6*e108853aSMatt Arsenault 
7*e108853aSMatt Arsenault float fv, fx;
8*e108853aSMatt Arsenault double dv, dx;
9*e108853aSMatt Arsenault 
10*e108853aSMatt Arsenault // DEFAULT-LABEL: define hidden void @_Z15atomic_fadd_f32v(
11*e108853aSMatt Arsenault // DEFAULT-SAME: ) #[[ATTR0:[0-9]+]] {
12*e108853aSMatt Arsenault // DEFAULT-NEXT:  [[ENTRY:.*:]]
13*e108853aSMatt Arsenault // DEFAULT-NEXT:    [[TMP0:%.*]] = load float, ptr addrspacecast (ptr addrspace(1) @fv to ptr), align 4
14*e108853aSMatt Arsenault // DEFAULT-NEXT:    [[TMP1:%.*]] = atomicrmw fadd ptr addrspacecast (ptr addrspace(1) @fx to ptr), float [[TMP0]] monotonic, align 4
15*e108853aSMatt Arsenault // DEFAULT-NEXT:    [[ADD:%.*]] = fadd float [[TMP1]], [[TMP0]]
16*e108853aSMatt Arsenault // DEFAULT-NEXT:    store float [[ADD]], ptr addrspacecast (ptr addrspace(1) @fv to ptr), align 4
17*e108853aSMatt Arsenault // DEFAULT-NEXT:    ret void
18*e108853aSMatt Arsenault //
19*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-LABEL: define hidden void @_Z15atomic_fadd_f32v(
20*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-SAME: ) #[[ATTR0:[0-9]+]] {
21*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-NEXT:  [[ENTRY:.*:]]
22*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-NEXT:    [[TMP0:%.*]] = load float, ptr addrspacecast (ptr addrspace(1) @fv to ptr), align 4
23*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-NEXT:    [[TMP1:%.*]] = atomicrmw fadd ptr addrspacecast (ptr addrspace(1) @fx to ptr), float [[TMP0]] monotonic, align 4, !amdgpu.no.fine.grained.memory [[META5:![0-9]+]], !amdgpu.ignore.denormal.mode [[META5]]
24*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-NEXT:    [[ADD:%.*]] = fadd float [[TMP1]], [[TMP0]]
25*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-NEXT:    store float [[ADD]], ptr addrspacecast (ptr addrspace(1) @fv to ptr), align 4
26*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-NEXT:    ret void
27*e108853aSMatt Arsenault //
28*e108853aSMatt Arsenault void atomic_fadd_f32() {
29*e108853aSMatt Arsenault #pragma omp atomic capture
30*e108853aSMatt Arsenault   fv = fx = fx + fv;
31*e108853aSMatt Arsenault }
32*e108853aSMatt Arsenault 
33*e108853aSMatt Arsenault // DEFAULT-LABEL: define hidden void @_Z15atomic_fadd_f64v(
34*e108853aSMatt Arsenault // DEFAULT-SAME: ) #[[ATTR0]] {
35*e108853aSMatt Arsenault // DEFAULT-NEXT:  [[ENTRY:.*:]]
36*e108853aSMatt Arsenault // DEFAULT-NEXT:    [[TMP0:%.*]] = load double, ptr addrspacecast (ptr addrspace(1) @dv to ptr), align 8
37*e108853aSMatt Arsenault // DEFAULT-NEXT:    [[TMP1:%.*]] = atomicrmw fadd ptr addrspacecast (ptr addrspace(1) @dx to ptr), double [[TMP0]] monotonic, align 8
38*e108853aSMatt Arsenault // DEFAULT-NEXT:    [[ADD:%.*]] = fadd double [[TMP1]], [[TMP0]]
39*e108853aSMatt Arsenault // DEFAULT-NEXT:    store double [[ADD]], ptr addrspacecast (ptr addrspace(1) @dv to ptr), align 8
40*e108853aSMatt Arsenault // DEFAULT-NEXT:    ret void
41*e108853aSMatt Arsenault //
42*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-LABEL: define hidden void @_Z15atomic_fadd_f64v(
43*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-SAME: ) #[[ATTR0]] {
44*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-NEXT:  [[ENTRY:.*:]]
45*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-NEXT:    [[TMP0:%.*]] = load double, ptr addrspacecast (ptr addrspace(1) @dv to ptr), align 8
46*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-NEXT:    [[TMP1:%.*]] = atomicrmw fadd ptr addrspacecast (ptr addrspace(1) @dx to ptr), double [[TMP0]] monotonic, align 8, !amdgpu.no.fine.grained.memory [[META5]]
47*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-NEXT:    [[ADD:%.*]] = fadd double [[TMP1]], [[TMP0]]
48*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-NEXT:    store double [[ADD]], ptr addrspacecast (ptr addrspace(1) @dv to ptr), align 8
49*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS-NEXT:    ret void
50*e108853aSMatt Arsenault //
51*e108853aSMatt Arsenault void atomic_fadd_f64() {
52*e108853aSMatt Arsenault #pragma omp atomic capture
53*e108853aSMatt Arsenault   dv = dx = dx + dv;
54*e108853aSMatt Arsenault }
55*e108853aSMatt Arsenault 
56*e108853aSMatt Arsenault #pragma omp end declare target
57*e108853aSMatt Arsenault //.
58*e108853aSMatt Arsenault // UNSAFE-FP-ATOMICS: [[META5]] = !{}
59*e108853aSMatt Arsenault //.
60