1// RUN: mlir-translate -mlir-to-llvmir %s | FileCheck %s 2 3// This tests checks that a target op inside a data op 4// We are only interested in ensuring that the -mlir-to-llmvir pass doesn't crash. 5// CHECK: {{.*}} = add i32 {{.*}}, 1 6module attributes { } { 7 llvm.mlir.global weak_odr hidden local_unnamed_addr constant @__oclc_ABI_version(400 : i32) {addr_space = 4 : i32} : i32 8 llvm.func @_QQmain() attributes {fir.bindc_name = "main", omp.declare_target = #omp.declaretarget<device_type = (host), capture_clause = (to)>} { 9 %0 = llvm.mlir.constant(99 : index) : i64 10 %1 = llvm.mlir.constant(0 : index) : i64 11 %2 = llvm.mlir.constant(1 : index) : i64 12 %3 = llvm.mlir.constant(100 : index) : i64 13 %4 = llvm.mlir.constant(1 : i64) : i64 14 %5 = llvm.alloca %4 x i32 {bindc_name = "array_length"} : (i64) -> !llvm.ptr<5> 15 %6 = llvm.addrspacecast %5 : !llvm.ptr<5> to !llvm.ptr 16 %7 = llvm.mlir.constant(1 : i64) : i64 17 %8 = llvm.alloca %7 x i32 {bindc_name = "index_"} : (i64) -> !llvm.ptr<5> 18 %9 = llvm.addrspacecast %8 : !llvm.ptr<5> to !llvm.ptr 19 %10 = llvm.mlir.addressof @_QFEint_array : !llvm.ptr 20 %11 = omp.map.bounds lower_bound(%1 : i64) upper_bound(%0 : i64) extent(%3 : i64) stride(%2 : i64) start_idx(%2 : i64) 21 %12 = omp.map.info var_ptr(%10 : !llvm.ptr, !llvm.array<100 x i32>) map_clauses(from) capture(ByRef) bounds(%11) -> !llvm.ptr {name = "int_array"} 22 omp.target_data map_entries(%12 : !llvm.ptr) { 23 %13 = omp.map.info var_ptr(%10 : !llvm.ptr, !llvm.array<100 x i32>) map_clauses(from) capture(ByRef) bounds(%11) -> !llvm.ptr {name = "int_array"} 24 %14 = omp.map.info var_ptr(%9 : !llvm.ptr, i32) map_clauses(implicit, exit_release_or_enter_alloc) capture(ByCopy) -> !llvm.ptr {name = "index_"} 25 omp.target map_entries(%13 -> %arg0, %14 -> %arg1 : !llvm.ptr, !llvm.ptr) { 26 %15 = llvm.mlir.constant(100 : i32) : i32 27 %16 = llvm.mlir.constant(1 : i32) : i32 28 %17 = llvm.mlir.constant(100 : index) : i64 29 omp.parallel { 30 %18 = llvm.mlir.constant(1 : i64) : i64 31 %19 = llvm.alloca %18 x i32 {pinned} : (i64) -> !llvm.ptr<5> 32 %20 = llvm.addrspacecast %19 : !llvm.ptr<5> to !llvm.ptr 33 omp.wsloop { 34 omp.loop_nest (%arg2) : i32 = (%16) to (%15) inclusive step (%16) { 35 llvm.store %arg2, %20 : i32, !llvm.ptr 36 %21 = llvm.load %20 : !llvm.ptr -> i32 37 %22 = llvm.sext %21 : i32 to i64 38 %23 = llvm.mlir.constant(1 : i64) : i64 39 %24 = llvm.mlir.constant(0 : i64) : i64 40 %25 = llvm.sub %22, %23 overflow<nsw> : i64 41 %26 = llvm.mul %25, %23 overflow<nsw> : i64 42 %27 = llvm.mul %26, %23 overflow<nsw> : i64 43 %28 = llvm.add %27, %24 overflow<nsw> : i64 44 %29 = llvm.mul %23, %17 overflow<nsw> : i64 45 %30 = llvm.getelementptr %arg0[%28] : (!llvm.ptr, i64) -> !llvm.ptr, i32 46 llvm.store %21, %30 : i32, !llvm.ptr 47 omp.yield 48 } 49 } 50 omp.terminator 51 } 52 omp.terminator 53 } 54 omp.terminator 55 } 56 llvm.return 57 } 58 llvm.mlir.global internal @_QFEint_array() {addr_space = 0 : i32} : !llvm.array<100 x i32> { 59 %0 = llvm.mlir.zero : !llvm.array<100 x i32> 60 llvm.return %0 : !llvm.array<100 x i32> 61 } 62} 63