/llvm-project/lldb/include/lldb/Utility/ |
H A D | Scalar.h | 1 //===-- Scalar.h ------------------------------------------------*- C++ -*-===// 31 // Operators are defined and Scalar objects will correctly promote their types 34 class Scalar { 52 Scalar() : m_float(0.0f) {} in Scalar() function 53 Scalar(int v) : m_type(e_int), m_integer(MakeAPSInt(v)), m_float(0.0f) {} in Scalar() function 54 Scalar(unsigned int v) in Scalar() function 56 Scalar(long v) : m_type(e_int), m_integer(MakeAPSInt(v)), m_float(0.0f) {} in Scalar() function 57 Scalar(unsigned long v) in Scalar() function 59 Scalar(long long v) in Scalar() function 61 Scalar(unsigne in Scalar() function 63 Scalar(float v) : m_type(e_float), m_float(v) {} Scalar() function 64 Scalar(double v) : m_type(e_float), m_float(v) {} Scalar() function 65 Scalar(long double v) : m_type(e_float), m_float(double(v)) { Scalar() function 70 Scalar(llvm::APInt v) Scalar() function 72 Scalar(llvm::APSInt v) Scalar() function 74 Scalar(llvm::APFloat v) : m_type(e_float), m_integer(0), m_float(v) {} Scalar() function [all...] |
/llvm-project/lldb/source/Utility/ |
H A D | Scalar.cpp | 1 //===-- Scalar.cpp --------------------------------------------------------===// 9 #include "lldb/Utility/Scalar.h" 31 Scalar::PromotionKey Scalar::GetPromoKey() const { in GetPromoKey() 43 Scalar::PromotionKey Scalar::GetFloatPromoKey(const llvm::fltSemantics &sem) { in GetFloatPromoKey() 56 Scalar::Type Scalar::PromoteToMaxType(Scalar &lhs, Scalar 935 operator <<(llvm::raw_ostream & os,const Scalar & scalar) operator <<() argument [all...] |
/llvm-project/llvm/test/CodeGen/X86/ |
H A D | subvectorwise-store-of-vector-splat.ll | 2 ; RUN: llc < %s -mtriple=x86_64-unknown-linux-gnu -mattr=-sse2 | FileCheck %s --check-prefixes=ALL,SCALAR 14 ; SCALAR-LABEL: vec32_v2i8: 15 ; SCALAR: # %bb.0: 16 ; SCALAR-NEXT: movzbl (%rdi), %eax 17 ; SCALAR-NEXT: movzbl 1(%rdi), %ecx 18 ; SCALAR-NEXT: notb %al 19 ; SCALAR-NEXT: notb %cl 20 ; SCALAR-NEXT: movb %cl, 1(%rsi) 21 ; SCALAR-NEXT: movb %al, (%rsi) 22 ; SCALAR [all...] |
/llvm-project/polly/test/ScopInfo/NonAffine/ |
H A D | non-affine-loop-condition-dependent-access_1.ll | 1 …-detect>,print<polly-function-scops>' -disable-output < %s 2>&1 | FileCheck %s -check-prefix=SCALAR 4 ; SCALAR: Function: f 5 ; SCALAR-NEXT: Region: %bb1---%bb13 6 ; SCALAR-NEXT: Max Loop Depth: 1 7 ; SCALAR-NEXT: Invariant Accesses: { 8 ; SCALAR-NEXT: } 9 ; SCALAR-NEXT: Context: 10 ; SCALAR-NEXT: { : } 11 ; SCALAR-NEXT: Assumed Context: 12 ; SCALAR-NEXT: { : } [all …]
|
/llvm-project/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/non-policy/overloaded/ |
H A D | vredand.c | 10 // CHECK-RV64-SAME: (<vscale x 1 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 12 …4(<vscale x 8 x i8> poison, <vscale x 1 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 15 vint8m1_t test_vredand_vs_i8mf8_i8m1(vint8mf8_t vector, vint8m1_t scalar, size_t vl) { in test_vredand_vs_i8mf8_i8m1() argument 16 return __riscv_vredand(vector, scalar, vl); in test_vredand_vs_i8mf8_i8m1() 20 // CHECK-RV64-SAME: (<vscale x 2 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 22 …4(<vscale x 8 x i8> poison, <vscale x 2 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 25 vint8m1_t test_vredand_vs_i8mf4_i8m1(vint8mf4_t vector, vint8m1_t scalar, size_t vl) { in test_vredand_vs_i8mf4_i8m1() argument 26 return __riscv_vredand(vector, scalar, vl); in test_vredand_vs_i8mf4_i8m1() 30 // CHECK-RV64-SAME: (<vscale x 4 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 32 …4(<vscale x 8 x i8> poison, <vscale x 4 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) [all …]
|
H A D | vredor.c | 10 // CHECK-RV64-SAME: (<vscale x 1 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 12 …4(<vscale x 8 x i8> poison, <vscale x 1 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 15 vint8m1_t test_vredor_vs_i8mf8_i8m1(vint8mf8_t vector, vint8m1_t scalar, size_t vl) { in test_vredor_vs_i8mf8_i8m1() argument 16 return __riscv_vredor(vector, scalar, vl); in test_vredor_vs_i8mf8_i8m1() 20 // CHECK-RV64-SAME: (<vscale x 2 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 22 …4(<vscale x 8 x i8> poison, <vscale x 2 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 25 vint8m1_t test_vredor_vs_i8mf4_i8m1(vint8mf4_t vector, vint8m1_t scalar, size_t vl) { in test_vredor_vs_i8mf4_i8m1() argument 26 return __riscv_vredor(vector, scalar, vl); in test_vredor_vs_i8mf4_i8m1() 30 // CHECK-RV64-SAME: (<vscale x 4 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 32 …4(<vscale x 8 x i8> poison, <vscale x 4 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) [all …]
|
H A D | vredsum.c | 10 // CHECK-RV64-SAME: (<vscale x 1 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 12 …4(<vscale x 8 x i8> poison, <vscale x 1 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 15 vint8m1_t test_vredsum_vs_i8mf8_i8m1(vint8mf8_t vector, vint8m1_t scalar, size_t vl) { in test_vredsum_vs_i8mf8_i8m1() argument 16 return __riscv_vredsum(vector, scalar, vl); in test_vredsum_vs_i8mf8_i8m1() 20 // CHECK-RV64-SAME: (<vscale x 2 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 22 …4(<vscale x 8 x i8> poison, <vscale x 2 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 25 vint8m1_t test_vredsum_vs_i8mf4_i8m1(vint8mf4_t vector, vint8m1_t scalar, size_t vl) { in test_vredsum_vs_i8mf4_i8m1() argument 26 return __riscv_vredsum(vector, scalar, vl); in test_vredsum_vs_i8mf4_i8m1() 30 // CHECK-RV64-SAME: (<vscale x 4 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 32 …4(<vscale x 8 x i8> poison, <vscale x 4 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) [all …]
|
H A D | vredxor.c | 10 // CHECK-RV64-SAME: (<vscale x 1 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 12 …4(<vscale x 8 x i8> poison, <vscale x 1 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 15 vint8m1_t test_vredxor_vs_i8mf8_i8m1(vint8mf8_t vector, vint8m1_t scalar, size_t vl) { in test_vredxor_vs_i8mf8_i8m1() argument 16 return __riscv_vredxor(vector, scalar, vl); in test_vredxor_vs_i8mf8_i8m1() 20 // CHECK-RV64-SAME: (<vscale x 2 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 22 …4(<vscale x 8 x i8> poison, <vscale x 2 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 25 vint8m1_t test_vredxor_vs_i8mf4_i8m1(vint8mf4_t vector, vint8m1_t scalar, size_t vl) { in test_vredxor_vs_i8mf4_i8m1() argument 26 return __riscv_vredxor(vector, scalar, vl); in test_vredxor_vs_i8mf4_i8m1() 30 // CHECK-RV64-SAME: (<vscale x 4 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 32 …4(<vscale x 8 x i8> poison, <vscale x 4 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) [all …]
|
H A D | vfredosum.c | 11 // CHECK-RV64-SAME: (<vscale x 1 x half> [[VECTOR:%.*]], <vscale x 4 x half> [[SCALAR:%.*]], i64 no… 13 … 4 x half> poison, <vscale x 1 x half> [[VECTOR]], <vscale x 4 x half> [[SCALAR]], i64 7, i64 [[VL… 16 vfloat16m1_t test_vfredosum_vs_f16mf4_f16m1(vfloat16mf4_t vector, vfloat16m1_t scalar, size_t vl) { in test_vfredosum_vs_f16mf4_f16m1() argument 17 return __riscv_vfredosum(vector, scalar, vl); in test_vfredosum_vs_f16mf4_f16m1() 21 // CHECK-RV64-SAME: (<vscale x 2 x half> [[VECTOR:%.*]], <vscale x 4 x half> [[SCALAR:%.*]], i64 no… 23 … 4 x half> poison, <vscale x 2 x half> [[VECTOR]], <vscale x 4 x half> [[SCALAR]], i64 7, i64 [[VL… 26 vfloat16m1_t test_vfredosum_vs_f16mf2_f16m1(vfloat16mf2_t vector, vfloat16m1_t scalar, size_t vl) { in test_vfredosum_vs_f16mf2_f16m1() argument 27 return __riscv_vfredosum(vector, scalar, vl); in test_vfredosum_vs_f16mf2_f16m1() 31 // CHECK-RV64-SAME: (<vscale x 4 x half> [[VECTOR:%.*]], <vscale x 4 x half> [[SCALAR:%.*]], i64 no… 33 … 4 x half> poison, <vscale x 4 x half> [[VECTOR]], <vscale x 4 x half> [[SCALAR]], i64 7, i64 [[VL… [all …]
|
H A D | vfredusum.c | 11 // CHECK-RV64-SAME: (<vscale x 1 x half> [[VECTOR:%.*]], <vscale x 4 x half> [[SCALAR:%.*]], i64 no… 13 … 4 x half> poison, <vscale x 1 x half> [[VECTOR]], <vscale x 4 x half> [[SCALAR]], i64 7, i64 [[VL… 16 vfloat16m1_t test_vfredusum_vs_f16mf4_f16m1(vfloat16mf4_t vector, vfloat16m1_t scalar, size_t vl) { in test_vfredusum_vs_f16mf4_f16m1() argument 17 return __riscv_vfredusum(vector, scalar, vl); in test_vfredusum_vs_f16mf4_f16m1() 21 // CHECK-RV64-SAME: (<vscale x 2 x half> [[VECTOR:%.*]], <vscale x 4 x half> [[SCALAR:%.*]], i64 no… 23 … 4 x half> poison, <vscale x 2 x half> [[VECTOR]], <vscale x 4 x half> [[SCALAR]], i64 7, i64 [[VL… 26 vfloat16m1_t test_vfredusum_vs_f16mf2_f16m1(vfloat16mf2_t vector, vfloat16m1_t scalar, size_t vl) { in test_vfredusum_vs_f16mf2_f16m1() argument 27 return __riscv_vfredusum(vector, scalar, vl); in test_vfredusum_vs_f16mf2_f16m1() 31 // CHECK-RV64-SAME: (<vscale x 4 x half> [[VECTOR:%.*]], <vscale x 4 x half> [[SCALAR:%.*]], i64 no… 33 … 4 x half> poison, <vscale x 4 x half> [[VECTOR]], <vscale x 4 x half> [[SCALAR]], i64 7, i64 [[VL… [all …]
|
H A D | vredmax.c | 10 // CHECK-RV64-SAME: (<vscale x 1 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 12 …4(<vscale x 8 x i8> poison, <vscale x 1 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 15 vint8m1_t test_vredmax_vs_i8mf8_i8m1(vint8mf8_t vector, vint8m1_t scalar, size_t vl) { in test_vredmax_vs_i8mf8_i8m1() argument 16 return __riscv_vredmax(vector, scalar, vl); in test_vredmax_vs_i8mf8_i8m1() 20 // CHECK-RV64-SAME: (<vscale x 2 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 22 …4(<vscale x 8 x i8> poison, <vscale x 2 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 25 vint8m1_t test_vredmax_vs_i8mf4_i8m1(vint8mf4_t vector, vint8m1_t scalar, size_t vl) { in test_vredmax_vs_i8mf4_i8m1() argument 26 return __riscv_vredmax(vector, scalar, vl); in test_vredmax_vs_i8mf4_i8m1() 30 // CHECK-RV64-SAME: (<vscale x 4 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 32 …4(<vscale x 8 x i8> poison, <vscale x 4 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) [all …]
|
H A D | vfwredusum.c | 11 // CHECK-RV64-SAME: (<vscale x 1 x half> [[VECTOR:%.*]], <vscale x 2 x float> [[SCALAR:%.*]], i64 n… 13 … x float> poison, <vscale x 1 x half> [[VECTOR]], <vscale x 2 x float> [[SCALAR]], i64 7, i64 [[VL… 16 vfloat32m1_t test_vfwredusum_vs_f16mf4_f32m1(vfloat16mf4_t vector, vfloat32m1_t scalar, size_t vl) { in test_vfwredusum_vs_f16mf4_f32m1() argument 17 return __riscv_vfwredusum(vector, scalar, vl); in test_vfwredusum_vs_f16mf4_f32m1() 21 // CHECK-RV64-SAME: (<vscale x 2 x half> [[VECTOR:%.*]], <vscale x 2 x float> [[SCALAR:%.*]], i64 n… 23 … x float> poison, <vscale x 2 x half> [[VECTOR]], <vscale x 2 x float> [[SCALAR]], i64 7, i64 [[VL… 26 vfloat32m1_t test_vfwredusum_vs_f16mf2_f32m1(vfloat16mf2_t vector, vfloat32m1_t scalar, size_t vl) { in test_vfwredusum_vs_f16mf2_f32m1() argument 27 return __riscv_vfwredusum(vector, scalar, vl); in test_vfwredusum_vs_f16mf2_f32m1() 31 // CHECK-RV64-SAME: (<vscale x 4 x half> [[VECTOR:%.*]], <vscale x 2 x float> [[SCALAR:%.*]], i64 n… 33 … x float> poison, <vscale x 4 x half> [[VECTOR]], <vscale x 2 x float> [[SCALAR]], i64 7, i64 [[VL… [all …]
|
H A D | vfwredosum.c | 11 // CHECK-RV64-SAME: (<vscale x 1 x half> [[VECTOR:%.*]], <vscale x 2 x float> [[SCALAR:%.*]], i64 n… 13 … x float> poison, <vscale x 1 x half> [[VECTOR]], <vscale x 2 x float> [[SCALAR]], i64 7, i64 [[VL… 16 vfloat32m1_t test_vfwredosum_vs_f16mf4_f32m1(vfloat16mf4_t vector, vfloat32m1_t scalar, size_t vl) { in test_vfwredosum_vs_f16mf4_f32m1() argument 17 return __riscv_vfwredosum(vector, scalar, vl); in test_vfwredosum_vs_f16mf4_f32m1() 21 // CHECK-RV64-SAME: (<vscale x 2 x half> [[VECTOR:%.*]], <vscale x 2 x float> [[SCALAR:%.*]], i64 n… 23 … x float> poison, <vscale x 2 x half> [[VECTOR]], <vscale x 2 x float> [[SCALAR]], i64 7, i64 [[VL… 26 vfloat32m1_t test_vfwredosum_vs_f16mf2_f32m1(vfloat16mf2_t vector, vfloat32m1_t scalar, size_t vl) { in test_vfwredosum_vs_f16mf2_f32m1() argument 27 return __riscv_vfwredosum(vector, scalar, vl); in test_vfwredosum_vs_f16mf2_f32m1() 31 // CHECK-RV64-SAME: (<vscale x 4 x half> [[VECTOR:%.*]], <vscale x 2 x float> [[SCALAR:%.*]], i64 n… 33 … x float> poison, <vscale x 4 x half> [[VECTOR]], <vscale x 2 x float> [[SCALAR]], i64 7, i64 [[VL… [all …]
|
/llvm-project/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/non-policy/non-overloaded/ |
H A D | vredand.c | 10 // CHECK-RV64-SAME: (<vscale x 1 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 12 …4(<vscale x 8 x i8> poison, <vscale x 1 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 15 vint8m1_t test_vredand_vs_i8mf8_i8m1(vint8mf8_t vector, vint8m1_t scalar, size_t vl) { in test_vredand_vs_i8mf8_i8m1() argument 16 return __riscv_vredand_vs_i8mf8_i8m1(vector, scalar, vl); in test_vredand_vs_i8mf8_i8m1() 20 // CHECK-RV64-SAME: (<vscale x 2 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 22 …4(<vscale x 8 x i8> poison, <vscale x 2 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 25 vint8m1_t test_vredand_vs_i8mf4_i8m1(vint8mf4_t vector, vint8m1_t scalar, size_t vl) { in test_vredand_vs_i8mf4_i8m1() argument 26 return __riscv_vredand_vs_i8mf4_i8m1(vector, scalar, vl); in test_vredand_vs_i8mf4_i8m1() 30 // CHECK-RV64-SAME: (<vscale x 4 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 32 …4(<vscale x 8 x i8> poison, <vscale x 4 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) [all …]
|
H A D | vredxor.c | 10 // CHECK-RV64-SAME: (<vscale x 1 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 12 …4(<vscale x 8 x i8> poison, <vscale x 1 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 15 vint8m1_t test_vredxor_vs_i8mf8_i8m1(vint8mf8_t vector, vint8m1_t scalar, size_t vl) { in test_vredxor_vs_i8mf8_i8m1() argument 16 return __riscv_vredxor_vs_i8mf8_i8m1(vector, scalar, vl); in test_vredxor_vs_i8mf8_i8m1() 20 // CHECK-RV64-SAME: (<vscale x 2 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 22 …4(<vscale x 8 x i8> poison, <vscale x 2 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 25 vint8m1_t test_vredxor_vs_i8mf4_i8m1(vint8mf4_t vector, vint8m1_t scalar, size_t vl) { in test_vredxor_vs_i8mf4_i8m1() argument 26 return __riscv_vredxor_vs_i8mf4_i8m1(vector, scalar, vl); in test_vredxor_vs_i8mf4_i8m1() 30 // CHECK-RV64-SAME: (<vscale x 4 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 32 …4(<vscale x 8 x i8> poison, <vscale x 4 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) [all …]
|
H A D | vredor.c | 10 // CHECK-RV64-SAME: (<vscale x 1 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 12 …4(<vscale x 8 x i8> poison, <vscale x 1 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 15 vint8m1_t test_vredor_vs_i8mf8_i8m1(vint8mf8_t vector, vint8m1_t scalar, size_t vl) { in test_vredor_vs_i8mf8_i8m1() argument 16 return __riscv_vredor_vs_i8mf8_i8m1(vector, scalar, vl); in test_vredor_vs_i8mf8_i8m1() 20 // CHECK-RV64-SAME: (<vscale x 2 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 22 …4(<vscale x 8 x i8> poison, <vscale x 2 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 25 vint8m1_t test_vredor_vs_i8mf4_i8m1(vint8mf4_t vector, vint8m1_t scalar, size_t vl) { in test_vredor_vs_i8mf4_i8m1() argument 26 return __riscv_vredor_vs_i8mf4_i8m1(vector, scalar, vl); in test_vredor_vs_i8mf4_i8m1() 30 // CHECK-RV64-SAME: (<vscale x 4 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 32 …4(<vscale x 8 x i8> poison, <vscale x 4 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) [all …]
|
H A D | vredsum.c | 10 // CHECK-RV64-SAME: (<vscale x 1 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 12 …4(<vscale x 8 x i8> poison, <vscale x 1 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 15 vint8m1_t test_vredsum_vs_i8mf8_i8m1(vint8mf8_t vector, vint8m1_t scalar, size_t vl) { in test_vredsum_vs_i8mf8_i8m1() argument 16 return __riscv_vredsum_vs_i8mf8_i8m1(vector, scalar, vl); in test_vredsum_vs_i8mf8_i8m1() 20 // CHECK-RV64-SAME: (<vscale x 2 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 22 …4(<vscale x 8 x i8> poison, <vscale x 2 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 25 vint8m1_t test_vredsum_vs_i8mf4_i8m1(vint8mf4_t vector, vint8m1_t scalar, size_t vl) { in test_vredsum_vs_i8mf4_i8m1() argument 26 return __riscv_vredsum_vs_i8mf4_i8m1(vector, scalar, vl); in test_vredsum_vs_i8mf4_i8m1() 30 // CHECK-RV64-SAME: (<vscale x 4 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 32 …4(<vscale x 8 x i8> poison, <vscale x 4 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) [all …]
|
H A D | vfredusum.c | 11 // CHECK-RV64-SAME: (<vscale x 1 x half> [[VECTOR:%.*]], <vscale x 4 x half> [[SCALAR:%.*]], i64 no… 13 … 4 x half> poison, <vscale x 1 x half> [[VECTOR]], <vscale x 4 x half> [[SCALAR]], i64 7, i64 [[VL… 16 vfloat16m1_t test_vfredusum_vs_f16mf4_f16m1(vfloat16mf4_t vector, vfloat16m1_t scalar, size_t vl) { in test_vfredusum_vs_f16mf4_f16m1() argument 17 return __riscv_vfredusum_vs_f16mf4_f16m1(vector, scalar, vl); in test_vfredusum_vs_f16mf4_f16m1() 21 // CHECK-RV64-SAME: (<vscale x 2 x half> [[VECTOR:%.*]], <vscale x 4 x half> [[SCALAR:%.*]], i64 no… 23 … 4 x half> poison, <vscale x 2 x half> [[VECTOR]], <vscale x 4 x half> [[SCALAR]], i64 7, i64 [[VL… 26 vfloat16m1_t test_vfredusum_vs_f16mf2_f16m1(vfloat16mf2_t vector, vfloat16m1_t scalar, size_t vl) { in test_vfredusum_vs_f16mf2_f16m1() argument 27 return __riscv_vfredusum_vs_f16mf2_f16m1(vector, scalar, vl); in test_vfredusum_vs_f16mf2_f16m1() 31 // CHECK-RV64-SAME: (<vscale x 4 x half> [[VECTOR:%.*]], <vscale x 4 x half> [[SCALAR:%.*]], i64 no… 33 … 4 x half> poison, <vscale x 4 x half> [[VECTOR]], <vscale x 4 x half> [[SCALAR]], i64 7, i64 [[VL… [all …]
|
H A D | vfredosum.c | 11 // CHECK-RV64-SAME: (<vscale x 1 x half> [[VECTOR:%.*]], <vscale x 4 x half> [[SCALAR:%.*]], i64 no… 13 … 4 x half> poison, <vscale x 1 x half> [[VECTOR]], <vscale x 4 x half> [[SCALAR]], i64 7, i64 [[VL… 16 vfloat16m1_t test_vfredosum_vs_f16mf4_f16m1(vfloat16mf4_t vector, vfloat16m1_t scalar, size_t vl) { in test_vfredosum_vs_f16mf4_f16m1() argument 17 return __riscv_vfredosum_vs_f16mf4_f16m1(vector, scalar, vl); in test_vfredosum_vs_f16mf4_f16m1() 21 // CHECK-RV64-SAME: (<vscale x 2 x half> [[VECTOR:%.*]], <vscale x 4 x half> [[SCALAR:%.*]], i64 no… 23 … 4 x half> poison, <vscale x 2 x half> [[VECTOR]], <vscale x 4 x half> [[SCALAR]], i64 7, i64 [[VL… 26 vfloat16m1_t test_vfredosum_vs_f16mf2_f16m1(vfloat16mf2_t vector, vfloat16m1_t scalar, size_t vl) { in test_vfredosum_vs_f16mf2_f16m1() argument 27 return __riscv_vfredosum_vs_f16mf2_f16m1(vector, scalar, vl); in test_vfredosum_vs_f16mf2_f16m1() 31 // CHECK-RV64-SAME: (<vscale x 4 x half> [[VECTOR:%.*]], <vscale x 4 x half> [[SCALAR:%.*]], i64 no… 33 … 4 x half> poison, <vscale x 4 x half> [[VECTOR]], <vscale x 4 x half> [[SCALAR]], i64 7, i64 [[VL… [all …]
|
H A D | vredminu.c | 10 // CHECK-RV64-SAME: (<vscale x 1 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 12 …4(<vscale x 8 x i8> poison, <vscale x 1 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 15 vuint8m1_t test_vredminu_vs_u8mf8_u8m1(vuint8mf8_t vector, vuint8m1_t scalar, size_t vl) { in test_vredminu_vs_u8mf8_u8m1() argument 16 return __riscv_vredminu_vs_u8mf8_u8m1(vector, scalar, vl); in test_vredminu_vs_u8mf8_u8m1() 20 // CHECK-RV64-SAME: (<vscale x 2 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 22 …4(<vscale x 8 x i8> poison, <vscale x 2 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 25 vuint8m1_t test_vredminu_vs_u8mf4_u8m1(vuint8mf4_t vector, vuint8m1_t scalar, size_t vl) { in test_vredminu_vs_u8mf4_u8m1() argument 26 return __riscv_vredminu_vs_u8mf4_u8m1(vector, scalar, vl); in test_vredminu_vs_u8mf4_u8m1() 30 // CHECK-RV64-SAME: (<vscale x 4 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 32 …4(<vscale x 8 x i8> poison, <vscale x 4 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) [all …]
|
H A D | vredmin.c | 10 // CHECK-RV64-SAME: (<vscale x 1 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 12 …4(<vscale x 8 x i8> poison, <vscale x 1 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 15 vint8m1_t test_vredmin_vs_i8mf8_i8m1(vint8mf8_t vector, vint8m1_t scalar, size_t vl) { in test_vredmin_vs_i8mf8_i8m1() argument 16 return __riscv_vredmin_vs_i8mf8_i8m1(vector, scalar, vl); in test_vredmin_vs_i8mf8_i8m1() 20 // CHECK-RV64-SAME: (<vscale x 2 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 22 …4(<vscale x 8 x i8> poison, <vscale x 2 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 25 vint8m1_t test_vredmin_vs_i8mf4_i8m1(vint8mf4_t vector, vint8m1_t scalar, size_t vl) { in test_vredmin_vs_i8mf4_i8m1() argument 26 return __riscv_vredmin_vs_i8mf4_i8m1(vector, scalar, vl); in test_vredmin_vs_i8mf4_i8m1() 30 // CHECK-RV64-SAME: (<vscale x 4 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 32 …4(<vscale x 8 x i8> poison, <vscale x 4 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) [all …]
|
H A D | vredmax.c | 10 // CHECK-RV64-SAME: (<vscale x 1 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 12 …4(<vscale x 8 x i8> poison, <vscale x 1 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 15 vint8m1_t test_vredmax_vs_i8mf8_i8m1(vint8mf8_t vector, vint8m1_t scalar, size_t vl) { in test_vredmax_vs_i8mf8_i8m1() argument 16 return __riscv_vredmax_vs_i8mf8_i8m1(vector, scalar, vl); in test_vredmax_vs_i8mf8_i8m1() 20 // CHECK-RV64-SAME: (<vscale x 2 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 22 …4(<vscale x 8 x i8> poison, <vscale x 2 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 25 vint8m1_t test_vredmax_vs_i8mf4_i8m1(vint8mf4_t vector, vint8m1_t scalar, size_t vl) { in test_vredmax_vs_i8mf4_i8m1() argument 26 return __riscv_vredmax_vs_i8mf4_i8m1(vector, scalar, vl); in test_vredmax_vs_i8mf4_i8m1() 30 // CHECK-RV64-SAME: (<vscale x 4 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 32 …4(<vscale x 8 x i8> poison, <vscale x 4 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) [all …]
|
H A D | vredmaxu.c | 10 // CHECK-RV64-SAME: (<vscale x 1 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 12 …4(<vscale x 8 x i8> poison, <vscale x 1 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 15 vuint8m1_t test_vredmaxu_vs_u8mf8_u8m1(vuint8mf8_t vector, vuint8m1_t scalar, size_t vl) { in test_vredmaxu_vs_u8mf8_u8m1() argument 16 return __riscv_vredmaxu_vs_u8mf8_u8m1(vector, scalar, vl); in test_vredmaxu_vs_u8mf8_u8m1() 20 // CHECK-RV64-SAME: (<vscale x 2 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 22 …4(<vscale x 8 x i8> poison, <vscale x 2 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) 25 vuint8m1_t test_vredmaxu_vs_u8mf4_u8m1(vuint8mf4_t vector, vuint8m1_t scalar, size_t vl) { in test_vredmaxu_vs_u8mf4_u8m1() argument 26 return __riscv_vredmaxu_vs_u8mf4_u8m1(vector, scalar, vl); in test_vredmaxu_vs_u8mf4_u8m1() 30 // CHECK-RV64-SAME: (<vscale x 4 x i8> [[VECTOR:%.*]], <vscale x 8 x i8> [[SCALAR:%.*]], i64 nounde… 32 …4(<vscale x 8 x i8> poison, <vscale x 4 x i8> [[VECTOR]], <vscale x 8 x i8> [[SCALAR]], i64 [[VL]]) [all …]
|
/llvm-project/llvm/test/CodeGen/SPIRV/instructions/ |
H A D | scalar-floating-point-arithmetic.ll | 14 ; CHECK-DAG: [[SCALAR:%.+]] = OpTypeFloat 32 15 ; CHECK-DAG: [[SCALAR_FN:%.+]] = OpTypeFunction [[SCALAR]] [[SCALAR]] [[SCALAR]] 20 ;; Test fneg on scalar: 28 ; DISABLED-CHECK: [[SCALAR_FNEG]] = OpFunction [[SCALAR]] None [[SCALAR_FN]] 29 ; DISABLED-CHECK-NEXT: [[A:%.+]] = OpFunctionParameter [[SCALAR]] 30 ; DISABLED-CHECK-NEXT: [[B:%.+]] = OpFunctionParameter [[SCALAR]] 32 ; DISABLED-CHECK: [[C:%.+]] = OpFNegate [[SCALAR]] [[A]] 37 ;; Test fadd on scalar [all...] |
H A D | scalar-bitwise-operations.ll | 12 ; CHECK-DAG: [[SCALAR:%.+]] = OpTypeInt 32 13 ; CHECK-DAG: [[SCALAR_FN:%.+]] = OpTypeFunction [[SCALAR]] [[SCALAR]] [[SCALAR]] 18 ;; Test shl on scalar: 24 ; CHECK: [[SCALAR_SHL]] = OpFunction [[SCALAR]] None [[SCALAR_FN]] 25 ; CHECK-NEXT: [[A:%.+]] = OpFunctionParameter [[SCALAR]] 26 ; CHECK-NEXT: [[B:%.+]] = OpFunctionParameter [[SCALAR]] 28 ; CHECK: [[C:%.+]] = OpShiftLeftLogical [[SCALAR]] [[A]] [[B]] 33 ;; Test lshr on scalar [all...] |