16c5d5ce5SUlrich Weigand; Test strict vector addition. 26c5d5ce5SUlrich Weigand; 36c5d5ce5SUlrich Weigand; RUN: llc < %s -mtriple=s390x-linux-gnu -mcpu=z13 | FileCheck %s 46c5d5ce5SUlrich Weigand 56c5d5ce5SUlrich Weiganddeclare double @llvm.experimental.constrained.fadd.f64(double, double, metadata, metadata) 66c5d5ce5SUlrich Weiganddeclare <2 x double> @llvm.experimental.constrained.fadd.v2f64(<2 x double>, <2 x double>, metadata, metadata) 76c5d5ce5SUlrich Weigand 86c5d5ce5SUlrich Weigand; Test a v2f64 addition. 96c5d5ce5SUlrich Weiganddefine <2 x double> @f5(<2 x double> %dummy, <2 x double> %val1, 10*68b80521SKevin P. Neal <2 x double> %val2) strictfp { 116c5d5ce5SUlrich Weigand; CHECK-LABEL: f5: 126c5d5ce5SUlrich Weigand; CHECK: vfadb %v24, %v26, %v28 136c5d5ce5SUlrich Weigand; CHECK: br %r14 146c5d5ce5SUlrich Weigand %ret = call <2 x double> @llvm.experimental.constrained.fadd.v2f64( 156c5d5ce5SUlrich Weigand <2 x double> %val1, <2 x double> %val2, 166c5d5ce5SUlrich Weigand metadata !"round.dynamic", 17*68b80521SKevin P. Neal metadata !"fpexcept.strict") strictfp 186c5d5ce5SUlrich Weigand ret <2 x double> %ret 196c5d5ce5SUlrich Weigand} 206c5d5ce5SUlrich Weigand 216c5d5ce5SUlrich Weigand; Test an f64 addition that uses vector registers. 22*68b80521SKevin P. Nealdefine double @f6(<2 x double> %val1, <2 x double> %val2) strictfp { 236c5d5ce5SUlrich Weigand; CHECK-LABEL: f6: 246c5d5ce5SUlrich Weigand; CHECK: wfadb %f0, %v24, %v26 256c5d5ce5SUlrich Weigand; CHECK: br %r14 266c5d5ce5SUlrich Weigand %scalar1 = extractelement <2 x double> %val1, i32 0 276c5d5ce5SUlrich Weigand %scalar2 = extractelement <2 x double> %val2, i32 0 286c5d5ce5SUlrich Weigand %ret = call double @llvm.experimental.constrained.fadd.f64( 296c5d5ce5SUlrich Weigand double %scalar1, double %scalar2, 306c5d5ce5SUlrich Weigand metadata !"round.dynamic", 31*68b80521SKevin P. Neal metadata !"fpexcept.strict") strictfp 326c5d5ce5SUlrich Weigand ret double %ret 336c5d5ce5SUlrich Weigand} 34