Lines Matching refs:vsfa
15 ; CHECK-NEXT: vsfa %v0, %v0, %s0, %s1
17 …%4 = tail call fast <256 x double> @llvm.ve.vl.vsfa.vvssl(<256 x double> %0, i64 %1, i64 %2, i32 2…
22 declare <256 x double> @llvm.ve.vl.vsfa.vvssl(<256 x double>, i64, i64, i32)
30 ; CHECK-NEXT: vsfa %v1, %v0, %s0, %s1
35 …%5 = tail call fast <256 x double> @llvm.ve.vl.vsfa.vvssvl(<256 x double> %0, i64 %1, i64 %2, <256…
40 declare <256 x double> @llvm.ve.vl.vsfa.vvssvl(<256 x double>, i64, i64, <256 x double>, i32)
48 ; CHECK-NEXT: vsfa %v0, %v0, 8, %s0
50 …%3 = tail call fast <256 x double> @llvm.ve.vl.vsfa.vvssl(<256 x double> %0, i64 8, i64 %1, i32 25…
60 ; CHECK-NEXT: vsfa %v1, %v0, 8, %s0
65 …%4 = tail call fast <256 x double> @llvm.ve.vl.vsfa.vvssvl(<256 x double> %0, i64 8, i64 %1, <256 …
75 ; CHECK-NEXT: vsfa %v1, %v0, %s0, %s1, %vm1
80 …%6 = tail call fast <256 x double> @llvm.ve.vl.vsfa.vvssmvl(<256 x double> %0, i64 %1, i64 %2, <25…
85 declare <256 x double> @llvm.ve.vl.vsfa.vvssmvl(<256 x double>, i64, i64, <256 x i1>, <256 x double…
93 ; CHECK-NEXT: vsfa %v1, %v0, 8, %s0, %vm1
98 …%5 = tail call fast <256 x double> @llvm.ve.vl.vsfa.vvssmvl(<256 x double> %0, i64 8, i64 %1, <256…