Lines Matching full:wasm

20   // WEBASSEMBLY32: call {{i.*}} @llvm.wasm.memory.size.i32(i32 0)
21 // WEBASSEMBLY64: call {{i.*}} @llvm.wasm.memory.size.i64(i32 0)
26 // WEBASSEMBLY32: call i32 @llvm.wasm.memory.grow.i32(i32 0, i32 %{{.*}})
27 // WEBASSEMBLY64: call i64 @llvm.wasm.memory.grow.i64(i32 0, i64 %{{.*}})
32 // WEBASSEMBLY32: call i32 @llvm.wasm.tls.size.i32()
33 // WEBASSEMBLY64: call i64 @llvm.wasm.tls.size.i64()
38 // WEBASSEMBLY32: call i32 @llvm.wasm.tls.align.i32()
39 // WEBASSEMBLY64: call i64 @llvm.wasm.tls.align.i64()
44 // WEBASSEMBLY: call ptr @llvm.wasm.tls.base()
49 // WEBASSEMBLY: call void @llvm.wasm.throw(i32 0, ptr %{{.*}})
54 // WEBASSEMBLY32: call void @llvm.wasm.rethrow()
55 // WEBASSEMBLY64: call void @llvm.wasm.rethrow()
60 // WEBASSEMBLY: call i32 @llvm.wasm.memory.atomic.wait32(ptr %{{.*}}, i32 %{{.*}}, i64 %{{.*}})
65 // WEBASSEMBLY: call i32 @llvm.wasm.memory.atomic.wait64(ptr %{{.*}}, i64 %{{.*}}, i64 %{{.*}})
70 // WEBASSEMBLY: call i32 @llvm.wasm.memory.atomic.notify(ptr %{{.*}}, i32 %{{.*}})
75 // WEBASSEMBLY: call i32 @llvm.wasm.trunc.signed.i32.f32(float %f)
81 // WEBASSEMBLY: call i32 @llvm.wasm.trunc.unsigned.i32.f32(float %f)
87 // WEBASSEMBLY: call i32 @llvm.wasm.trunc.signed.i32.f64(double %f)
93 // WEBASSEMBLY: call i32 @llvm.wasm.trunc.unsigned.i32.f64(double %f)
99 // WEBASSEMBLY: call i64 @llvm.wasm.trunc.signed.i64.f32(float %f)
105 // WEBASSEMBLY: call i64 @llvm.wasm.trunc.unsigned.i64.f32(float %f)
111 // WEBASSEMBLY: call i64 @llvm.wasm.trunc.signed.i64.f64(double %f)
117 // WEBASSEMBLY: call i64 @llvm.wasm.trunc.unsigned.i64.f64(double %f)
220 // WEBASSEMBLY: call <16 x i8> @llvm.wasm.avgr.unsigned.v16i8(
227 // WEBASSEMBLY: call <8 x i16> @llvm.wasm.avgr.unsigned.v8i16(
234 // WEBASSEMBLY: call <8 x i16> @llvm.wasm.q15mulr.sat.signed(
241 // WEBASSEMBLY: call <8 x i16> @llvm.wasm.extadd.pairwise.signed.v8i16(
248 // WEBASSEMBLY: call <8 x i16> @llvm.wasm.extadd.pairwise.unsigned.v8i16(
255 // WEBASSEMBLY: call <4 x i32> @llvm.wasm.extadd.pairwise.signed.v4i32(
262 // WEBASSEMBLY: call <4 x i32> @llvm.wasm.extadd.pairwise.unsigned.v4i32(
269 // WEBASSEMBLY: call <4 x i32> @llvm.wasm.dot(<8 x i16> %x, <8 x i16> %y)
275 // WEBASSEMBLY: call <4 x i32> @llvm.wasm.bitselect.v4i32(
282 // WEBASSEMBLY: call i32 @llvm.wasm.anytrue.v16i8(<16 x i8> %x)
288 // WEBASSEMBLY: call i32 @llvm.wasm.alltrue.v16i8(<16 x i8> %x)
294 // WEBASSEMBLY: call i32 @llvm.wasm.alltrue.v8i16(<8 x i16> %x)
300 // WEBASSEMBLY: call i32 @llvm.wasm.alltrue.v4i32(<4 x i32> %x)
306 // WEBASSEMBLY: call i32 @llvm.wasm.alltrue.v2i64(<2 x i64> %x)
312 // WEBASSEMBLY: call i32 @llvm.wasm.bitmask.v16i8(<16 x i8> %x)
318 // WEBASSEMBLY: call i32 @llvm.wasm.bitmask.v8i16(<8 x i16> %x)
324 // WEBASSEMBLY: call i32 @llvm.wasm.bitmask.v4i32(<4 x i32> %x)
330 // WEBASSEMBLY: call i32 @llvm.wasm.bitmask.v2i64(<2 x i64> %x)
362 // WEBASSEMBLY: call <4 x float> @llvm.wasm.pmin.v4f32(
369 // WEBASSEMBLY: call <4 x float> @llvm.wasm.pmax.v4f32(
390 // WEBASSEMBLY: call <2 x double> @llvm.wasm.pmin.v2f64(
397 // WEBASSEMBLY: call <2 x double> @llvm.wasm.pmax.v2f64(
476 // WEBASSEMBLY: call <16 x i8> @llvm.wasm.narrow.signed.v16i8.v8i16(
483 // WEBASSEMBLY: call <16 x i8> @llvm.wasm.narrow.unsigned.v16i8.v8i16(
490 // WEBASSEMBLY: call <8 x i16> @llvm.wasm.narrow.signed.v8i16.v4i32(
497 // WEBASSEMBLY: call <8 x i16> @llvm.wasm.narrow.unsigned.v8i16.v4i32(
518 // WEBASSEMBLY: call <16 x i8> @llvm.wasm.swizzle(<16 x i8> %x, <16 x i8> %y)
524 // WEBASSEMBLY: call <16 x i8> @llvm.wasm.shuffle(<16 x i8> %x, <16 x i8> %y,
533 // WEBASSEMBLY: call <4 x float> @llvm.wasm.relaxed.madd.v4f32(
540 // WEBASSEMBLY: call <4 x float> @llvm.wasm.relaxed.nmadd.v4f32(
547 // WEBASSEMBLY: call <2 x double> @llvm.wasm.relaxed.madd.v2f64(
554 // WEBASSEMBLY: call <2 x double> @llvm.wasm.relaxed.nmadd.v2f64(
561 // WEBASSEMBLY: call <8 x half> @llvm.wasm.relaxed.madd.v8f16(
568 // WEBASSEMBLY: call <8 x half> @llvm.wasm.relaxed.nmadd.v8f16(
575 // WEBASSEMBLY: call <16 x i8> @llvm.wasm.relaxed.laneselect.v16i8(
582 // WEBASSEMBLY: call <8 x i16> @llvm.wasm.relaxed.laneselect.v8i16(
589 // WEBASSEMBLY: call <4 x i32> @llvm.wasm.relaxed.laneselect.v4i32(
596 // WEBASSEMBLY: call <2 x i64> @llvm.wasm.relaxed.laneselect.v2i64(
603 // WEBASSEMBLY: call <16 x i8> @llvm.wasm.relaxed.swizzle(<16 x i8> %x, <16 x i8> %y)
608 // WEBASSEMBLY: call <4 x float> @llvm.wasm.relaxed.min.v4f32(
615 // WEBASSEMBLY: call <4 x float> @llvm.wasm.relaxed.max.v4f32(
622 // WEBASSEMBLY: call <2 x double> @llvm.wasm.relaxed.min.v2f64(
629 // WEBASSEMBLY: call <2 x double> @llvm.wasm.relaxed.max.v2f64(
636 // WEBASSEMBLY: call <4 x i32> @llvm.wasm.relaxed.trunc.signed(<4 x float> %f)
642 // WEBASSEMBLY: call <4 x i32> @llvm.wasm.relaxed.trunc.unsigned(<4 x float> %f)
648 // WEBASSEMBLY: call <4 x i32> @llvm.wasm.relaxed.trunc.signed.zero(<2 x double> %x)
654 // WEBASSEMBLY: call <4 x i32> @llvm.wasm.relaxed.trunc.unsigned.zero(<2 x double> %x)
660 // WEBASSEMBLY: call <8 x i16> @llvm.wasm.relaxed.q15mulr.signed(
667 // WEBASSEMBLY: call <8 x i16> @llvm.wasm.relaxed.dot.i8x16.i7x16.signed(
674 // WEBASSEMBLY: call <4 x i32> @llvm.wasm.relaxed.dot.i8x16.i7x16.add.signed(
681 // WEBASSEMBLY: call <4 x float> @llvm.wasm.relaxed.dot.bf16x8.add.f32
688 // WEBASSEMBLY: call float @llvm.wasm.loadf16.f32(ptr %{{.*}})
693 // WEBASSEMBLY: tail call void @llvm.wasm.storef16.f32(float %val, ptr %{{.*}})
698 // WEBASSEMBLY: %0 = tail call <8 x half> @llvm.wasm.splat.f16x8(float %a)
704 // WEBASSEMBLY: %0 = tail call float @llvm.wasm.extract.lane.f16x8(<8 x half> %a, i32 7)
710 // WEBASSEMBLY: %0 = tail call <8 x half> @llvm.wasm.replace.lane.f16x8(<8 x half> %a, i32 7, float %v)
728 // WEBASSEMBLY: %0 = tail call <8 x half> @llvm.wasm.pmin.v8f16(<8 x half> %a, <8 x half> %b)
734 // WEBASSEMBLY: %0 = tail call <8 x half> @llvm.wasm.pmax.v8f16(<8 x half> %a, <8 x half> %b)
740 // WEBASSEMBLY: tail call ptr addrspace(10) @llvm.wasm.ref.null.extern()