/dports/devel/wasi-libcxx/llvm-project-13.0.1.src/llvm/test/CodeGen/AArch64/ |
H A D | sve-split-fp-reduce.ll | 13 %res = call double @llvm.vector.reduce.fadd.nxv8f64(double %init, <vscale x 8 x double> %a) 40 %res = call double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double> %a) 56 declare double @llvm.vector.reduce.fadd.nxv8f64(double, <vscale x 8 x double>) 59 declare double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double>)
|
H A D | sve-calling-convention-mixed.ll | 33 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 34 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 74 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 75 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 104 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 196 declare <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1>, double*) 200 declare <vscale x 2 x double> @llvm.aarch64.sve.tuple.get.nxv2f64.nxv8f64(<vscale x 8 x double>, i3…
|
/dports/graphics/llvm-mesa/llvm-13.0.1.src/test/CodeGen/AArch64/ |
H A D | sve-split-fp-reduce.ll | 13 %res = call double @llvm.vector.reduce.fadd.nxv8f64(double %init, <vscale x 8 x double> %a) 40 %res = call double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double> %a) 56 declare double @llvm.vector.reduce.fadd.nxv8f64(double, <vscale x 8 x double>) 59 declare double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double>)
|
H A D | sve-calling-convention-mixed.ll | 33 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 34 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 74 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 75 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 104 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 196 declare <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1>, double*) 200 declare <vscale x 2 x double> @llvm.aarch64.sve.tuple.get.nxv2f64.nxv8f64(<vscale x 8 x double>, i3…
|
/dports/devel/wasi-compiler-rt13/llvm-project-13.0.1.src/llvm/test/CodeGen/AArch64/ |
H A D | sve-split-fp-reduce.ll | 13 %res = call double @llvm.vector.reduce.fadd.nxv8f64(double %init, <vscale x 8 x double> %a) 40 %res = call double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double> %a) 56 declare double @llvm.vector.reduce.fadd.nxv8f64(double, <vscale x 8 x double>) 59 declare double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double>)
|
H A D | sve-calling-convention-mixed.ll | 33 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 34 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 74 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 75 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 104 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 196 declare <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1>, double*) 200 declare <vscale x 2 x double> @llvm.aarch64.sve.tuple.get.nxv2f64.nxv8f64(<vscale x 8 x double>, i3…
|
/dports/devel/llvm13/llvm-project-13.0.1.src/llvm/test/CodeGen/AArch64/ |
H A D | sve-split-fp-reduce.ll | 13 %res = call double @llvm.vector.reduce.fadd.nxv8f64(double %init, <vscale x 8 x double> %a) 40 %res = call double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double> %a) 56 declare double @llvm.vector.reduce.fadd.nxv8f64(double, <vscale x 8 x double>) 59 declare double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double>)
|
H A D | sve-calling-convention-mixed.ll | 33 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 34 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 74 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 75 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 104 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 196 declare <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1>, double*) 200 declare <vscale x 2 x double> @llvm.aarch64.sve.tuple.get.nxv2f64.nxv8f64(<vscale x 8 x double>, i3…
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/llvm/llvm/test/CodeGen/AArch64/ |
H A D | sve-calling-convention-tuples-broken.ll | 13 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 14 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 23 declare <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1>, double*)
|
/dports/devel/llvm11/llvm-11.0.1.src/test/CodeGen/AArch64/ |
H A D | sve-calling-convention-tuples-broken.ll | 13 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 14 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 23 declare <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1>, double*)
|
/dports/devel/llvm12/llvm-project-12.0.1.src/llvm/test/CodeGen/AArch64/ |
H A D | sve-split-fp-reduce.ll | 17 %res = call double @llvm.vector.reduce.fadd.nxv8f64(double %init, <vscale x 8 x double> %a) 44 %res = call double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double> %a) 60 declare double @llvm.vector.reduce.fadd.nxv8f64(double, <vscale x 8 x double>) 63 declare double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double>)
|
H A D | sve-calling-convention-mixed.ll | 33 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 34 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 74 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 75 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 104 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 196 declare <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1>, double*) 200 declare <vscale x 2 x double> @llvm.aarch64.sve.tuple.get.nxv2f64.nxv8f64(<vscale x 8 x double>, i3…
|
/dports/devel/llvm-devel/llvm-project-f05c95f10fc1d8171071735af8ad3a9e87633120/llvm/test/CodeGen/AArch64/ |
H A D | sve-split-fp-reduce.ll | 17 %res = call double @llvm.vector.reduce.fadd.nxv8f64(double %init, <vscale x 8 x double> %a) 47 %res = call double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double> %a) 65 declare double @llvm.vector.reduce.fadd.nxv8f64(double, <vscale x 8 x double>) 68 declare double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double>)
|
H A D | sve-calling-convention-mixed.ll | 33 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 34 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 74 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 75 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 104 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 196 declare <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1>, double*) 200 declare <vscale x 2 x double> @llvm.aarch64.sve.tuple.get.nxv2f64.nxv8f64(<vscale x 8 x double>, i3…
|
/dports/devel/wasi-compiler-rt12/llvm-project-12.0.1.src/llvm/test/CodeGen/AArch64/ |
H A D | sve-split-fp-reduce.ll | 17 %res = call double @llvm.vector.reduce.fadd.nxv8f64(double %init, <vscale x 8 x double> %a) 44 %res = call double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double> %a) 60 declare double @llvm.vector.reduce.fadd.nxv8f64(double, <vscale x 8 x double>) 63 declare double @llvm.vector.reduce.fmax.nxv8f64(<vscale x 8 x double>)
|
H A D | sve-calling-convention-mixed.ll | 33 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 34 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 74 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 75 …%3 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 104 …%2 = call <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1> %1, double*… 196 declare <vscale x 8 x double> @llvm.aarch64.sve.ld4.nxv8f64.nxv2i1(<vscale x 2 x i1>, double*) 200 declare <vscale x 2 x double> @llvm.aarch64.sve.tuple.get.nxv2f64.nxv8f64(<vscale x 8 x double>, i3…
|
/dports/devel/llvm80/llvm-8.0.1.src/include/llvm/Support/ |
H A D | MachineValueType.h | 173 nxv8f64 = 108, // n x 8 x f64 enumerator 176 LAST_FP_VECTOR_VALUETYPE = nxv8f64, 179 LAST_FP_SCALABLE_VALUETYPE = nxv8f64, 182 LAST_VECTOR_VALUETYPE = nxv8f64, 514 case nxv8f64: return f64; in getVectorElementType() 569 case nxv8f64: return 8; in getVectorNumElements() 738 case nxv8f64: return 512; in getSizeInBits() 966 if (NumElements == 8) return MVT::nxv8f64; in getScalableVectorVT()
|
/dports/devel/llvm70/llvm-7.0.1.src/include/llvm/Support/ |
H A D | MachineValueType.h | 173 nxv8f64 = 108, // n x 8 x f64 enumerator 176 LAST_FP_VECTOR_VALUETYPE = nxv8f64, 179 LAST_FP_SCALABLE_VALUETYPE = nxv8f64, 182 LAST_VECTOR_VALUETYPE = nxv8f64, 514 case nxv8f64: return f64; in getVectorElementType() 569 case nxv8f64: return 8; in getVectorNumElements() 738 case nxv8f64: return 512; in getSizeInBits() 966 if (NumElements == 8) return MVT::nxv8f64; in getScalableVectorVT()
|
/dports/devel/llvm90/llvm-9.0.1.src/include/llvm/Support/ |
H A D | MachineValueType.h | 188 nxv8f64 = 124, // n x 8 x f64 enumerator 191 LAST_FP_VECTOR_VALUETYPE = nxv8f64, 194 LAST_FP_SCALABLE_VALUETYPE = nxv8f64, 197 LAST_VECTOR_VALUETYPE = nxv8f64, 545 case nxv8f64: return f64; in getVectorElementType() 612 case nxv8f64: return 8; in getVectorNumElements() 789 case nxv8f64: return 512; in getSizeInBits() 1045 if (NumElements == 8) return MVT::nxv8f64; in getScalableVectorVT()
|
/dports/devel/llvm10/llvm-10.0.1.src/include/llvm/Support/ |
H A D | MachineValueType.h | 199 nxv8f64 = 129, // n x 8 x f64 enumerator 202 LAST_FP_SCALABLE_VECTOR_VALUETYPE = nxv8f64, 205 LAST_SCALABLE_VECTOR_VALUETYPE = nxv8f64, 208 LAST_VECTOR_VALUETYPE = nxv8f64, 543 case nxv8f64: return f64; in getVectorElementType() 613 case nxv8f64: return 8; in getVectorNumElements() 802 case nxv8f64: return TypeSize::Scalable(512); in getSizeInBits() 1078 if (NumElements == 8) return MVT::nxv8f64; in getScalableVectorVT()
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/swiftshader/third_party/llvm-10.0/llvm/include/llvm/Support/ |
H A D | MachineValueType.h | 199 nxv8f64 = 129, // n x 8 x f64 enumerator 202 LAST_FP_SCALABLE_VECTOR_VALUETYPE = nxv8f64, 205 LAST_SCALABLE_VECTOR_VALUETYPE = nxv8f64, 208 LAST_VECTOR_VALUETYPE = nxv8f64, 543 case nxv8f64: return f64; in getVectorElementType() 613 case nxv8f64: return 8; in getVectorNumElements() 802 case nxv8f64: return TypeSize::Scalable(512); in getSizeInBits() 1078 if (NumElements == 8) return MVT::nxv8f64; in getScalableVectorVT()
|
/dports/devel/tinygo/tinygo-0.14.1/llvm-project/llvm/include/llvm/Support/ |
H A D | MachineValueType.h | 199 nxv8f64 = 129, // n x 8 x f64 enumerator 202 LAST_FP_SCALABLE_VECTOR_VALUETYPE = nxv8f64, 205 LAST_SCALABLE_VECTOR_VALUETYPE = nxv8f64, 208 LAST_VECTOR_VALUETYPE = nxv8f64, 543 case nxv8f64: return f64; in getVectorElementType() 613 case nxv8f64: return 8; in getVectorNumElements() 802 case nxv8f64: return TypeSize::Scalable(512); in getSizeInBits() 1078 if (NumElements == 8) return MVT::nxv8f64; in getScalableVectorVT()
|
/dports/devel/llvm-cheri/llvm-project-37c49ff00e3eadce5d8703fdc4497f28458c64a8/llvm/include/llvm/Support/ |
H A D | MachineValueType.h | 221 nxv8f64 = 150, // n x 8 x f64 enumerator 224 LAST_FP_SCALABLE_VECTOR_VALUETYPE = nxv8f64, 227 LAST_SCALABLE_VECTOR_VALUETYPE = nxv8f64, 230 LAST_VECTOR_VALUETYPE = nxv8f64, 608 case nxv8f64: return f64; in getVectorElementType() 692 case nxv8f64: return 8; in getVectorNumElements() 912 case nxv8f64: return TypeSize::Scalable(512); in getSizeInBits() 1233 if (NumElements == 8) return MVT::nxv8f64; in getScalableVectorVT()
|
/dports/devel/llvm11/llvm-11.0.1.src/include/llvm/Support/ |
H A D | MachineValueType.h | 220 nxv8f64 = 150, // n x 8 x f64 enumerator 223 LAST_FP_SCALABLE_VECTOR_VALUETYPE = nxv8f64, 226 LAST_SCALABLE_VECTOR_VALUETYPE = nxv8f64, 229 LAST_VECTOR_VALUETYPE = nxv8f64, 592 case nxv8f64: return f64; in getVectorElementType() 676 case nxv8f64: return 8; in getVectorNumElements() 891 case nxv8f64: return TypeSize::Scalable(512); in getSizeInBits() 1197 if (NumElements == 8) return MVT::nxv8f64; in getScalableVectorVT()
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/llvm/llvm/include/llvm/Support/ |
H A D | MachineValueType.h | 226 nxv8f64 = 156, // n x 8 x f64 enumerator 229 LAST_FP_SCALABLE_VECTOR_VALUETYPE = nxv8f64, 232 LAST_SCALABLE_VECTOR_VALUETYPE = nxv8f64, 235 LAST_VECTOR_VALUETYPE = nxv8f64, 636 case nxv8f64: return f64; in getVectorElementType() 726 case nxv8f64: return 8; in getVectorNumElements() 941 case nxv8f64: return TypeSize::Scalable(512); in getSizeInBits() 1295 if (NumElements == 8) return MVT::nxv8f64; in getScalableVectorVT()
|