1; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py 2; RUN: llc -verify-machineinstrs -mtriple powerpc64le-unknown-linux-gnu \ 3; RUN: -mcpu=pwr10 -ppc-vsr-nums-as-vr -ppc-asm-full-reg-names < %s \ 4; RUN: | FileCheck %s 5; RUN: llc -verify-machineinstrs -mtriple powerpc64le-unknown-linux-gnu \ 6; RUN: -mcpu=pwr10 -ppc-vsr-nums-as-vr -ppc-asm-full-reg-names \ 7; RUN: -enable-subreg-liveness < %s | FileCheck %s --check-prefix=TRACKLIVE 8 9%0 = type <{ double }> 10%1 = type <{ double }> 11 12define void @acc_regalloc(ptr %arg, ptr %arg1, ptr %arg2) local_unnamed_addr { 13; CHECK-LABEL: acc_regalloc: 14; CHECK: # %bb.0: # %bb 15; CHECK-NEXT: lwz r3, 0(r3) 16; CHECK-NEXT: lxv v4, 0(0) 17; CHECK-NEXT: xxlxor v0, v0, v0 18; CHECK-NEXT: xxlxor v1, v1, v1 19; CHECK-NEXT: xxlxor v2, v2, v2 20; CHECK-NEXT: li r6, 1 21; CHECK-NEXT: li r4, 16 22; CHECK-NEXT: stfd f14, -144(r1) # 8-byte Folded Spill 23; CHECK-NEXT: stfd f15, -136(r1) # 8-byte Folded Spill 24; CHECK-NEXT: extswsli r3, r3, 3 25; CHECK-NEXT: xvmaddadp v1, v4, v1 26; CHECK-NEXT: lxvdsx v5, 0, r3 27; CHECK-NEXT: xvmaddadp v0, v5, v0 28; CHECK-NEXT: .p2align 4 29; CHECK-NEXT: .LBB0_1: # %bb9 30; CHECK-NEXT: # 31; CHECK-NEXT: addi r6, r6, 2 32; CHECK-NEXT: lxv vs0, 16(0) 33; CHECK-NEXT: lxv vs1, -64(r5) 34; CHECK-NEXT: xxlxor vs7, vs7, vs7 35; CHECK-NEXT: xxlor vs3, v0, v0 36; CHECK-NEXT: xxlxor vs2, vs2, vs2 37; CHECK-NEXT: xxlxor vs12, vs12, vs12 38; CHECK-NEXT: mulld r6, r6, r3 39; CHECK-NEXT: xxlor vs10, v2, v2 40; CHECK-NEXT: xxlor vs4, v2, v2 41; CHECK-NEXT: xxlor vs8, vs10, vs10 42; CHECK-NEXT: xxlor vs10, v1, v1 43; CHECK-NEXT: xvmaddadp vs7, vs0, v5 44; CHECK-NEXT: xvmuldp vs6, vs0, v2 45; CHECK-NEXT: lxv vs0, -16(r5) 46; CHECK-NEXT: xvmaddadp vs3, vs1, v2 47; CHECK-NEXT: xvmaddadp vs2, vs1, vs2 48; CHECK-NEXT: lxvdsx v6, r6, r4 49; CHECK-NEXT: li r6, 0 50; CHECK-NEXT: xvmaddadp vs7, v2, v2 51; CHECK-NEXT: xvmaddadp vs6, v2, v2 52; CHECK-NEXT: xvmaddadp vs12, vs0, vs12 53; CHECK-NEXT: xvmuldp v3, vs1, v6 54; CHECK-NEXT: xvmuldp vs11, v4, v6 55; CHECK-NEXT: xvmuldp vs13, vs0, v6 56; CHECK-NEXT: xvmuldp vs5, v6, v2 57; CHECK-NEXT: xxlor vs0, v2, v2 58; CHECK-NEXT: xxlor vs14, vs12, vs12 59; CHECK-NEXT: xxlor vs12, v2, v2 60; CHECK-NEXT: xxlor vs1, v3, v3 61; CHECK-NEXT: xxlor vs9, vs11, vs11 62; CHECK-NEXT: xxlor vs15, vs13, vs13 63; CHECK-NEXT: xxmtacc acc1 64; CHECK-NEXT: xxmtacc acc0 65; CHECK-NEXT: xxmtacc acc2 66; CHECK-NEXT: xxmtacc acc3 67; CHECK-NEXT: xvf64gerpp acc0, vsp34, vs0 68; CHECK-NEXT: xvf64gerpp acc1, vsp34, vs0 69; CHECK-NEXT: xvf64gerpp acc2, vsp34, vs0 70; CHECK-NEXT: xvf64gerpp acc3, vsp34, vs0 71; CHECK-NEXT: xvf64gerpp acc0, vsp34, vs0 72; CHECK-NEXT: xvf64gerpp acc1, vsp34, vs0 73; CHECK-NEXT: xvf64gerpp acc2, vsp34, vs0 74; CHECK-NEXT: xvf64gerpp acc3, vsp34, vs0 75; CHECK-NEXT: xvf64gerpp acc0, vsp34, vs0 76; CHECK-NEXT: xvf64gerpp acc1, vsp34, vs0 77; CHECK-NEXT: xvf64gerpp acc2, vsp34, vs0 78; CHECK-NEXT: xvf64gerpp acc3, vsp34, vs0 79; CHECK-NEXT: xvf64gerpp acc0, vsp34, vs0 80; CHECK-NEXT: xvf64gerpp acc1, vsp34, vs0 81; CHECK-NEXT: xvf64gerpp acc2, vsp34, vs0 82; CHECK-NEXT: xvf64gerpp acc3, vsp34, vs0 83; CHECK-NEXT: xvf64gerpp acc0, vsp34, vs0 84; CHECK-NEXT: xvf64gerpp acc1, vsp34, vs0 85; CHECK-NEXT: xvf64gerpp acc2, vsp34, vs0 86; CHECK-NEXT: xvf64gerpp acc3, vsp34, vs0 87; CHECK-NEXT: xvf64gerpp acc0, vsp34, vs0 88; CHECK-NEXT: xvf64gerpp acc1, vsp34, vs0 89; CHECK-NEXT: xvf64gerpp acc2, vsp34, vs0 90; CHECK-NEXT: xvf64gerpp acc3, vsp34, vs0 91; CHECK-NEXT: xvf64gerpp acc0, vsp34, vs0 92; CHECK-NEXT: xvf64gerpp acc1, vsp34, vs0 93; CHECK-NEXT: xvf64gerpp acc2, vsp34, vs0 94; CHECK-NEXT: xvf64gerpp acc3, vsp34, vs0 95; CHECK-NEXT: xxmfacc acc0 96; CHECK-NEXT: xxmfacc acc1 97; CHECK-NEXT: xxmfacc acc2 98; CHECK-NEXT: xxmfacc acc3 99; CHECK-NEXT: stxv vs1, 0(r3) 100; CHECK-NEXT: stxv vs9, 32(r3) 101; CHECK-NEXT: stxv vs4, 16(0) 102; CHECK-NEXT: stxv vs12, 48(0) 103; CHECK-NEXT: b .LBB0_1 104; 105; TRACKLIVE-LABEL: acc_regalloc: 106; TRACKLIVE: # %bb.0: # %bb 107; TRACKLIVE-NEXT: lwz r3, 0(r3) 108; TRACKLIVE-NEXT: lxv v4, 0(0) 109; TRACKLIVE-NEXT: xxlxor v0, v0, v0 110; TRACKLIVE-NEXT: xxlxor v1, v1, v1 111; TRACKLIVE-NEXT: xxlxor v2, v2, v2 112; TRACKLIVE-NEXT: li r6, 1 113; TRACKLIVE-NEXT: li r4, 16 114; TRACKLIVE-NEXT: stfd f14, -144(r1) # 8-byte Folded Spill 115; TRACKLIVE-NEXT: stfd f15, -136(r1) # 8-byte Folded Spill 116; TRACKLIVE-NEXT: extswsli r3, r3, 3 117; TRACKLIVE-NEXT: xvmaddadp v1, v4, v1 118; TRACKLIVE-NEXT: lxvdsx v5, 0, r3 119; TRACKLIVE-NEXT: xvmaddadp v0, v5, v0 120; TRACKLIVE-NEXT: .p2align 4 121; TRACKLIVE-NEXT: .LBB0_1: # %bb9 122; TRACKLIVE-NEXT: # 123; TRACKLIVE-NEXT: addi r6, r6, 2 124; TRACKLIVE-NEXT: lxv vs0, 16(0) 125; TRACKLIVE-NEXT: lxv vs1, -64(r5) 126; TRACKLIVE-NEXT: xxlxor vs7, vs7, vs7 127; TRACKLIVE-NEXT: xxlor vs3, v0, v0 128; TRACKLIVE-NEXT: xxlxor vs2, vs2, vs2 129; TRACKLIVE-NEXT: xxlxor vs12, vs12, vs12 130; TRACKLIVE-NEXT: mulld r6, r6, r3 131; TRACKLIVE-NEXT: xxlor vs10, v2, v2 132; TRACKLIVE-NEXT: xxlor vs4, v2, v2 133; TRACKLIVE-NEXT: xxlor vs8, vs10, vs10 134; TRACKLIVE-NEXT: xxlor vs10, v1, v1 135; TRACKLIVE-NEXT: xvmaddadp vs7, vs0, v5 136; TRACKLIVE-NEXT: xvmuldp vs6, vs0, v2 137; TRACKLIVE-NEXT: lxv vs0, -16(r5) 138; TRACKLIVE-NEXT: xvmaddadp vs3, vs1, v2 139; TRACKLIVE-NEXT: xvmaddadp vs2, vs1, vs2 140; TRACKLIVE-NEXT: lxvdsx v6, r6, r4 141; TRACKLIVE-NEXT: li r6, 0 142; TRACKLIVE-NEXT: xvmaddadp vs7, v2, v2 143; TRACKLIVE-NEXT: xvmaddadp vs6, v2, v2 144; TRACKLIVE-NEXT: xvmaddadp vs12, vs0, vs12 145; TRACKLIVE-NEXT: xvmuldp v3, vs1, v6 146; TRACKLIVE-NEXT: xvmuldp vs11, v4, v6 147; TRACKLIVE-NEXT: xvmuldp vs13, vs0, v6 148; TRACKLIVE-NEXT: xvmuldp vs5, v6, v2 149; TRACKLIVE-NEXT: xxlor vs0, v2, v2 150; TRACKLIVE-NEXT: xxlor vs14, vs12, vs12 151; TRACKLIVE-NEXT: xxlor vs12, v2, v2 152; TRACKLIVE-NEXT: xxlor vs1, v3, v3 153; TRACKLIVE-NEXT: xxlor vs9, vs11, vs11 154; TRACKLIVE-NEXT: xxlor vs15, vs13, vs13 155; TRACKLIVE-NEXT: xxmtacc acc1 156; TRACKLIVE-NEXT: xxmtacc acc0 157; TRACKLIVE-NEXT: xxmtacc acc2 158; TRACKLIVE-NEXT: xxmtacc acc3 159; TRACKLIVE-NEXT: xvf64gerpp acc0, vsp34, vs0 160; TRACKLIVE-NEXT: xvf64gerpp acc1, vsp34, vs0 161; TRACKLIVE-NEXT: xvf64gerpp acc2, vsp34, vs0 162; TRACKLIVE-NEXT: xvf64gerpp acc3, vsp34, vs0 163; TRACKLIVE-NEXT: xvf64gerpp acc0, vsp34, vs0 164; TRACKLIVE-NEXT: xvf64gerpp acc1, vsp34, vs0 165; TRACKLIVE-NEXT: xvf64gerpp acc2, vsp34, vs0 166; TRACKLIVE-NEXT: xvf64gerpp acc3, vsp34, vs0 167; TRACKLIVE-NEXT: xvf64gerpp acc0, vsp34, vs0 168; TRACKLIVE-NEXT: xvf64gerpp acc1, vsp34, vs0 169; TRACKLIVE-NEXT: xvf64gerpp acc2, vsp34, vs0 170; TRACKLIVE-NEXT: xvf64gerpp acc3, vsp34, vs0 171; TRACKLIVE-NEXT: xvf64gerpp acc0, vsp34, vs0 172; TRACKLIVE-NEXT: xvf64gerpp acc1, vsp34, vs0 173; TRACKLIVE-NEXT: xvf64gerpp acc2, vsp34, vs0 174; TRACKLIVE-NEXT: xvf64gerpp acc3, vsp34, vs0 175; TRACKLIVE-NEXT: xvf64gerpp acc0, vsp34, vs0 176; TRACKLIVE-NEXT: xvf64gerpp acc1, vsp34, vs0 177; TRACKLIVE-NEXT: xvf64gerpp acc2, vsp34, vs0 178; TRACKLIVE-NEXT: xvf64gerpp acc3, vsp34, vs0 179; TRACKLIVE-NEXT: xvf64gerpp acc0, vsp34, vs0 180; TRACKLIVE-NEXT: xvf64gerpp acc1, vsp34, vs0 181; TRACKLIVE-NEXT: xvf64gerpp acc2, vsp34, vs0 182; TRACKLIVE-NEXT: xvf64gerpp acc3, vsp34, vs0 183; TRACKLIVE-NEXT: xvf64gerpp acc0, vsp34, vs0 184; TRACKLIVE-NEXT: xvf64gerpp acc1, vsp34, vs0 185; TRACKLIVE-NEXT: xvf64gerpp acc2, vsp34, vs0 186; TRACKLIVE-NEXT: xvf64gerpp acc3, vsp34, vs0 187; TRACKLIVE-NEXT: xxmfacc acc0 188; TRACKLIVE-NEXT: xxmfacc acc1 189; TRACKLIVE-NEXT: xxmfacc acc2 190; TRACKLIVE-NEXT: xxmfacc acc3 191; TRACKLIVE-NEXT: stxv vs1, 0(r3) 192; TRACKLIVE-NEXT: stxv vs9, 32(r3) 193; TRACKLIVE-NEXT: stxv vs4, 16(0) 194; TRACKLIVE-NEXT: stxv vs12, 48(0) 195; TRACKLIVE-NEXT: b .LBB0_1 196bb: 197 %i = load i32, ptr %arg, align 4 198 %i3 = sext i32 %i to i64 199 %i4 = shl nsw i64 %i3, 3 200 %i6 = getelementptr i8, ptr %arg1, i64 undef 201 %i7 = getelementptr [0 x %1], ptr %arg2, i64 0, i64 -8 202 %i8 = getelementptr i8, ptr %i6, i64 undef 203 br label %bb9 204 205bb9: ; preds = %bb95, %bb 206 %i10 = phi i64 [ 1, %bb ], [ 0, %bb95 ] 207 %i11 = getelementptr %1, ptr null, i64 2 208 %i13 = load <2 x double>, ptr %i11, align 1 209 %i14 = add nuw nsw i64 %i10, 2 210 %i15 = getelementptr inbounds %1, ptr %i7, i64 undef 211 %i17 = load <2 x double>, ptr %i15, align 1 212 %i18 = load <2 x double>, ptr null, align 1 213 %i19 = getelementptr %1, ptr %i15, i64 6 214 %i21 = load <2 x double>, ptr %i19, align 1 215 %i22 = load i64, ptr undef, align 8 216 %i23 = insertelement <2 x i64> poison, i64 %i22, i32 0 217 %i24 = bitcast <2 x i64> %i23 to <2 x double> 218 %i25 = shufflevector <2 x double> %i24, <2 x double> undef, <2 x i32> zeroinitializer 219 %i26 = mul i64 %i14, %i4 220 %i27 = getelementptr i8, ptr null, i64 %i26 221 %i29 = getelementptr i8, ptr %i27, i64 16 222 %i31 = load i64, ptr %i29, align 8 223 %i32 = insertelement <2 x i64> poison, i64 %i31, i32 0 224 %i33 = bitcast <2 x i64> %i32 to <2 x double> 225 %i34 = shufflevector <2 x double> %i33, <2 x double> undef, <2 x i32> zeroinitializer 226 %i35 = tail call contract <2 x double> @llvm.fma.v2f64(<2 x double> zeroinitializer, <2 x double> %i25, <2 x double> zeroinitializer) 227 %i36 = tail call contract <2 x double> @llvm.fma.v2f64(<2 x double> %i13, <2 x double> %i25, <2 x double> zeroinitializer) 228 %i37 = fmul contract <2 x double> %i13, zeroinitializer 229 %i38 = tail call contract <2 x double> @llvm.fma.v2f64(<2 x double> %i17, <2 x double> zeroinitializer, <2 x double> %i35) 230 %i39 = tail call contract <2 x double> @llvm.fma.v2f64(<2 x double> zeroinitializer, <2 x double> zeroinitializer, <2 x double> %i36) 231 %i40 = tail call contract <2 x double> @llvm.fma.v2f64(<2 x double> %i17, <2 x double> zeroinitializer, <2 x double> zeroinitializer) 232 %i41 = tail call contract <2 x double> @llvm.fma.v2f64(<2 x double> zeroinitializer, <2 x double> zeroinitializer, <2 x double> %i37) 233 %i42 = tail call contract <2 x double> @llvm.fma.v2f64(<2 x double> %i18, <2 x double> zeroinitializer, <2 x double> zeroinitializer) 234 %i43 = tail call contract <2 x double> @llvm.fma.v2f64(<2 x double> %i21, <2 x double> zeroinitializer, <2 x double> zeroinitializer) 235 %i44 = fmul contract <2 x double> %i17, %i34 236 %i45 = fmul contract <2 x double> zeroinitializer, %i34 237 %i46 = fmul contract <2 x double> %i18, %i34 238 %i47 = fmul contract <2 x double> %i21, %i34 239 %i48 = bitcast <2 x double> %i44 to <16 x i8> 240 %i49 = bitcast <2 x double> %i40 to <16 x i8> 241 %i50 = bitcast <2 x double> %i38 to <16 x i8> 242 %i51 = tail call <512 x i1> @llvm.ppc.mma.assemble.acc(<16 x i8> zeroinitializer, <16 x i8> %i48, <16 x i8> %i49, <16 x i8> %i50) 243 %i52 = bitcast <2 x double> %i45 to <16 x i8> 244 %i53 = bitcast <2 x double> %i41 to <16 x i8> 245 %i54 = bitcast <2 x double> %i39 to <16 x i8> 246 %i55 = tail call <512 x i1> @llvm.ppc.mma.assemble.acc(<16 x i8> zeroinitializer, <16 x i8> %i52, <16 x i8> %i53, <16 x i8> %i54) 247 %i56 = bitcast <2 x double> %i46 to <16 x i8> 248 %i57 = bitcast <2 x double> %i42 to <16 x i8> 249 %i58 = tail call <512 x i1> @llvm.ppc.mma.assemble.acc(<16 x i8> zeroinitializer, <16 x i8> %i56, <16 x i8> %i57, <16 x i8> %i56) 250 %i59 = bitcast <2 x double> %i47 to <16 x i8> 251 %i60 = bitcast <2 x double> %i43 to <16 x i8> 252 %i61 = tail call <512 x i1> @llvm.ppc.mma.assemble.acc(<16 x i8> zeroinitializer, <16 x i8> %i59, <16 x i8> %i60, <16 x i8> %i59) 253 %i62 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i51, <256 x i1> undef, <16 x i8> undef) 254 %i63 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i55, <256 x i1> undef, <16 x i8> undef) 255 %i64 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i58, <256 x i1> undef, <16 x i8> undef) 256 %i65 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i61, <256 x i1> undef, <16 x i8> undef) 257 %i66 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i62, <256 x i1> undef, <16 x i8> undef) 258 %i67 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i63, <256 x i1> undef, <16 x i8> undef) 259 %i68 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i64, <256 x i1> undef, <16 x i8> undef) 260 %i69 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i65, <256 x i1> undef, <16 x i8> undef) 261 %i70 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i66, <256 x i1> undef, <16 x i8> undef) 262 %i71 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i67, <256 x i1> undef, <16 x i8> undef) 263 %i72 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i68, <256 x i1> undef, <16 x i8> undef) 264 %i73 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i69, <256 x i1> undef, <16 x i8> undef) 265 %i74 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i70, <256 x i1> undef, <16 x i8> undef) 266 %i75 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i71, <256 x i1> undef, <16 x i8> undef) 267 %i76 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i72, <256 x i1> undef, <16 x i8> undef) 268 %i77 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i73, <256 x i1> undef, <16 x i8> undef) 269 %i78 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i74, <256 x i1> undef, <16 x i8> undef) 270 %i79 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i75, <256 x i1> undef, <16 x i8> undef) 271 %i80 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i76, <256 x i1> undef, <16 x i8> undef) 272 %i81 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i77, <256 x i1> undef, <16 x i8> undef) 273 br label %bb82 274 275bb82: ; preds = %bb82, %bb9 276 %i83 = phi <512 x i1> [ %i94, %bb82 ], [ %i81, %bb9 ] 277 %i84 = phi <512 x i1> [ %i93, %bb82 ], [ %i80, %bb9 ] 278 %i85 = phi <512 x i1> [ %i92, %bb82 ], [ %i79, %bb9 ] 279 %i86 = phi <512 x i1> [ %i91, %bb82 ], [ %i78, %bb9 ] 280 %i87 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i86, <256 x i1> undef, <16 x i8> undef) 281 %i88 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i85, <256 x i1> undef, <16 x i8> undef) 282 %i89 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i84, <256 x i1> undef, <16 x i8> undef) 283 %i90 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i83, <256 x i1> undef, <16 x i8> undef) 284 %i91 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i87, <256 x i1> undef, <16 x i8> undef) 285 %i92 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i88, <256 x i1> undef, <16 x i8> undef) 286 %i93 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i89, <256 x i1> undef, <16 x i8> undef) 287 %i94 = tail call <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1> %i90, <256 x i1> undef, <16 x i8> undef) 288 br i1 undef, label %bb95, label %bb82 289 290bb95: ; preds = %bb82 291 %i96 = tail call { <16 x i8>, <16 x i8>, <16 x i8>, <16 x i8> } @llvm.ppc.mma.disassemble.acc(<512 x i1> %i91) 292 %i97 = extractvalue { <16 x i8>, <16 x i8>, <16 x i8>, <16 x i8> } %i96, 2 293 %i98 = tail call { <16 x i8>, <16 x i8>, <16 x i8>, <16 x i8> } @llvm.ppc.mma.disassemble.acc(<512 x i1> %i92) 294 %i99 = extractvalue { <16 x i8>, <16 x i8>, <16 x i8>, <16 x i8> } %i98, 3 295 %i100 = tail call { <16 x i8>, <16 x i8>, <16 x i8>, <16 x i8> } @llvm.ppc.mma.disassemble.acc(<512 x i1> %i93) 296 %i101 = extractvalue { <16 x i8>, <16 x i8>, <16 x i8>, <16 x i8> } %i100, 2 297 %i102 = tail call { <16 x i8>, <16 x i8>, <16 x i8>, <16 x i8> } @llvm.ppc.mma.disassemble.acc(<512 x i1> %i94) 298 %i103 = extractvalue { <16 x i8>, <16 x i8>, <16 x i8>, <16 x i8> } %i102, 3 299 %i104 = getelementptr inbounds i8, ptr %i8, i64 undef 300 store <16 x i8> %i97, ptr %i104, align 1 301 %i106 = getelementptr i8, ptr %i104, i64 32 302 store <16 x i8> %i101, ptr %i106, align 1 303 %i108 = getelementptr i8, ptr null, i64 16 304 store <16 x i8> %i99, ptr %i108, align 1 305 %i110 = getelementptr i8, ptr null, i64 48 306 store <16 x i8> %i103, ptr %i110, align 1 307 br label %bb9 308} 309 310declare <2 x double> @llvm.fma.v2f64(<2 x double>, <2 x double>, <2 x double>) 311declare <512 x i1> @llvm.ppc.mma.assemble.acc(<16 x i8>, <16 x i8>, <16 x i8>, <16 x i8>) 312declare <512 x i1> @llvm.ppc.mma.xvf64gerpp(<512 x i1>, <256 x i1>, <16 x i8>) 313declare { <16 x i8>, <16 x i8>, <16 x i8>, <16 x i8> } @llvm.ppc.mma.disassemble.acc(<512 x i1>) 314 315