xref: /llvm-project/llvm/test/CodeGen/ARM/neon_fpconv.ll (revision bed1c7f061aa12417aa081e334afdba45767b938)
1; RUN: llc -mtriple=arm-eabi -mattr=+neon %s -o - | FileCheck %s
2
3; PR12540: ARM backend lowering of FP_ROUND v2f64 to v2f32.
4define <2 x float> @vtrunc(<2 x double> %a) {
5; CHECK: vcvt.f32.f64 [[S0:s[0-9]+]], [[D0:d[0-9]+]]
6; CHECK: vcvt.f32.f64 [[S1:s[0-9]+]], [[D1:d[0-9]+]]
7  %vt = fptrunc <2 x double> %a to <2 x float>
8  ret <2 x float> %vt
9}
10
11define <2 x double> @vextend(<2 x float> %a) {
12; CHECK: vcvt.f64.f32 [[D0:d[0-9]+]], [[S0:s[0-9]+]]
13; CHECK: vcvt.f64.f32 [[D1:d[0-9]+]], [[S1:s[0-9]+]]
14  %ve = fpext <2 x float> %a to <2 x double>
15  ret <2 x double> %ve
16}
17
18; We used to generate vmovs between scalar and vfp/neon registers.
19; CHECK: vsitofp_double
20define void @vsitofp_double(ptr %loadaddr,
21                            ptr %storeaddr) {
22  %v0 = load <2 x i32>, ptr %loadaddr
23; CHECK:      vldr
24; CHECK-NEXT:	vcvt.f64.s32
25; CHECK-NEXT:	vcvt.f64.s32
26; CHECK-NEXT:	vst
27  %r = sitofp <2 x i32> %v0 to <2 x double>
28  store <2 x double> %r, ptr %storeaddr
29  ret void
30}
31; CHECK: vuitofp_double
32define void @vuitofp_double(ptr %loadaddr,
33                            ptr %storeaddr) {
34  %v0 = load <2 x i32>, ptr %loadaddr
35; CHECK:      vldr
36; CHECK-NEXT:	vcvt.f64.u32
37; CHECK-NEXT:	vcvt.f64.u32
38; CHECK-NEXT:	vst
39  %r = uitofp <2 x i32> %v0 to <2 x double>
40  store <2 x double> %r, ptr %storeaddr
41  ret void
42}
43