xref: /llvm-project/mlir/test/Target/LLVMIR/omptarget-region-device-llvm.mlir (revision d2d4c3bd598453a1d871096466953ec25483034f)
1// RUN: mlir-translate -mlir-to-llvmir %s | FileCheck %s
2
3module attributes {llvm.target_triple = "amdgcn-amd-amdhsa", omp.is_target_device = true} {
4  llvm.func @omp_target_region_() {
5    %0 = llvm.mlir.constant(20 : i32) : i32
6    %1 = llvm.mlir.constant(10 : i32) : i32
7    %2 = llvm.mlir.constant(1 : i64) : i64
8    %3 = llvm.alloca %2 x i32 {bindc_name = "a", in_type = i32, operandSegmentSizes = array<i32: 0, 0>, uniq_name = "_QFomp_target_regionEa"} : (i64) -> !llvm.ptr
9    %4 = llvm.mlir.constant(1 : i64) : i64
10    %5 = llvm.alloca %4 x i32 {bindc_name = "b", in_type = i32, operandSegmentSizes = array<i32: 0, 0>, uniq_name = "_QFomp_target_regionEb"} : (i64) -> !llvm.ptr
11    %6 = llvm.mlir.constant(1 : i64) : i64
12    %7 = llvm.alloca %6 x i32 {bindc_name = "c", in_type = i32, operandSegmentSizes = array<i32: 0, 0>, uniq_name = "_QFomp_target_regionEc"} : (i64) -> !llvm.ptr
13    llvm.store %1, %3 : i32, !llvm.ptr
14    llvm.store %0, %5 : i32, !llvm.ptr
15    %map1 = omp.map.info var_ptr(%3 : !llvm.ptr, i32)   map_clauses(tofrom) capture(ByRef) -> !llvm.ptr {name = ""}
16    %map2 = omp.map.info var_ptr(%5 : !llvm.ptr, i32)   map_clauses(tofrom) capture(ByRef) -> !llvm.ptr {name = ""}
17    %map3 = omp.map.info var_ptr(%7 : !llvm.ptr, i32)   map_clauses(tofrom) capture(ByRef) -> !llvm.ptr {name = ""}
18    omp.target map_entries(%map1 -> %arg0, %map2 -> %arg1, %map3 -> %arg2 : !llvm.ptr, !llvm.ptr, !llvm.ptr) {
19      %8 = llvm.load %arg0 : !llvm.ptr -> i32
20      %9 = llvm.load %arg1 : !llvm.ptr -> i32
21      %10 = llvm.add %8, %9  : i32
22      llvm.store %10, %arg2 : i32, !llvm.ptr
23      omp.terminator
24    }
25    llvm.return
26  }
27}
28
29// CHECK:      @[[SRC_LOC:.*]] = private unnamed_addr constant [23 x i8] c"{{[^"]*}}", align 1
30// CHECK:      @[[IDENT:.*]] = private unnamed_addr constant %struct.ident_t { i32 0, i32 2, i32 0, i32 22, ptr @[[SRC_LOC]] }, align 8
31// CHECK:      @[[DYNA_ENV:.*]] = weak_odr protected global %struct.DynamicEnvironmentTy zeroinitializer
32// CHECK:      @[[KERNEL_ENV:.*]] = weak_odr protected constant %struct.KernelEnvironmentTy { %struct.ConfigurationEnvironmentTy { i8 1, i8 1, i8 1, i32 1, i32 256, i32 -1, i32 -1, i32 0, i32 0 }, ptr @[[IDENT]], ptr @[[DYNA_ENV]] }
33// CHECK:      define weak_odr protected amdgpu_kernel void @__omp_offloading_{{[^_]+}}_{{[^_]+}}_omp_target_region__l{{[0-9]+}}(ptr %[[DYN_PTR:.*]], ptr %[[ADDR_A:.*]], ptr %[[ADDR_B:.*]], ptr %[[ADDR_C:.*]])
34// CHECK:        %[[TMP_A:.*]] = alloca ptr, align 8
35// CHECK:        store ptr %[[ADDR_A]], ptr %[[TMP_A]], align 8
36// CHECK:        %[[TMP_B:.*]] = alloca ptr, align 8
37// CHECK:        store ptr %[[ADDR_B]], ptr %[[TMP_B]], align 8
38// CHECK:        %[[TMP_C:.*]] = alloca ptr, align 8
39// CHECK:        store ptr %[[ADDR_C]], ptr %[[TMP_C]], align 8
40// CHECK:        %[[INIT:.*]] = call i32 @__kmpc_target_init(ptr @[[KERNEL_ENV]], ptr %[[DYN_PTR]])
41// CHECK-NEXT:   %[[CMP:.*]] = icmp eq i32 %[[INIT]], -1
42// CHECK-NEXT:   br i1 %[[CMP]], label %[[LABEL_ENTRY:.*]], label %[[LABEL_EXIT:.*]]
43// CHECK:        [[LABEL_ENTRY]]:
44// CHECK:        %[[PTR_A:.*]] = load ptr, ptr %[[TMP_A]], align 8
45// CHECK:        %[[PTR_B:.*]] = load ptr, ptr %[[TMP_B]], align 8
46// CHECK:        %[[PTR_C:.*]] = load ptr, ptr %[[TMP_C]], align 8
47// CHECK-NEXT:   br label %[[LABEL_TARGET:.*]]
48// CHECK:        [[LABEL_TARGET]]:
49// CHECK:        %[[A:.*]] = load i32, ptr %[[PTR_A]], align 4
50// CHECK:        %[[B:.*]] = load i32, ptr %[[PTR_B]], align 4
51// CHECK:        %[[C:.*]] = add i32 %[[A]], %[[B]]
52// CHECK:        store i32 %[[C]], ptr %[[PTR_C]], align 4
53// CHECK:        br label %[[LABEL_DEINIT:.*]]
54// CHECK:        [[LABEL_DEINIT]]:
55// CHECK-NEXT:   call void @__kmpc_target_deinit()
56// CHECK-NEXT:   ret void
57// CHECK:        [[LABEL_EXIT]]:
58// CHECK-NEXT:   ret void
59