/dports/graphics/art/ART-1.9.3/rtengine/ |
H A D | sleefsseavx.h | 59 static INLINE vdouble vadd5(vdouble v0, vdouble v1, vdouble v2, vdouble v3, vdouble v4) { in vadd5() function 64 return vadd5(vadd(v0, v1), v2, v3, v4, v5); in vadd6() 179 vdouble u = vadd5(vsub(vmul(nhh, th), q.x), vmul(nhh, tl), vmul(nhl, th), vmul(nhl, tl), in div_dd() 193 r.y = vadd5(vmul(xh, yh), vneg(r.x), vmul(xl, yh), vmul(xh, yl), vmul(xl, yl)); in mul_ss() 225 …r.y = vadd5(vmul(xh, xh), vneg(r.x), vmul(vadd(xh, xh), xl), vmul(xl, xl), vmul(x.x, vadd(x.y, x.y… in squ_d()
|
/dports/graphics/librtprocess/librtprocess-0.11.0/src/include/ |
H A D | sleefsseavx.c | 58 static INLINE vdouble vadd5(vdouble v0, vdouble v1, vdouble v2, vdouble v3, vdouble v4) { in vadd5() function 63 return vadd5(vadd(v0, v1), v2, v3, v4, v5); in vadd6() 178 vdouble u = vadd5(vsub(vmul(nhh, th), q.x), vmul(nhh, tl), vmul(nhl, th), vmul(nhl, tl), in div_dd() 192 r.y = vadd5(vmul(xh, yh), vneg(r.x), vmul(xl, yh), vmul(xh, yl), vmul(xl, yl)); in mul_ss() 224 …r.y = vadd5(vmul(xh, xh), vneg(r.x), vmul(vadd(xh, xh), xl), vmul(xl, xl), vmul(x.x, vadd(x.y, x.y… in squ_d()
|
/dports/astro/siril/siril/subprojects/librtprocess/src/include/ |
H A D | sleefsseavx.h | 63 static INLINE vdouble vadd5(vdouble v0, vdouble v1, vdouble v2, vdouble v3, vdouble v4) { in vadd5() function 68 return vadd5(vadd(v0, v1), v2, v3, v4, v5); in vadd6() 183 vdouble u = vadd5(vsub(vmul(nhh, th), q.x), vmul(nhh, tl), vmul(nhl, th), vmul(nhl, tl), in div_dd() 197 r.y = vadd5(vmul(xh, yh), vneg(r.x), vmul(xl, yh), vmul(xh, yl), vmul(xl, yl)); in mul_ss() 229 …r.y = vadd5(vmul(xh, xh), vneg(r.x), vmul(vadd(xh, xh), xl), vmul(xl, xl), vmul(x.x, vadd(x.y, x.y… in squ_d()
|
/dports/astro/siril/siril/src/rt/ |
H A D | sleefsseavx.c | 61 static INLINE vdouble vadd5(vdouble v0, vdouble v1, vdouble v2, vdouble v3, vdouble v4) { in vadd5() function 66 return vadd5(vadd(v0, v1), v2, v3, v4, v5); in vadd6() 181 vdouble u = vadd5(vsub(vmul(nhh, th), q.x), vmul(nhh, tl), vmul(nhl, th), vmul(nhl, tl), in div_dd() 195 r.y = vadd5(vmul(xh, yh), vneg(r.x), vmul(xl, yh), vmul(xh, yl), vmul(xl, yl)); in mul_ss() 227 …r.y = vadd5(vmul(xh, xh), vneg(r.x), vmul(vadd(xh, xh), xl), vmul(xl, xl), vmul(x.x, vadd(x.y, x.y… in squ_d()
|
/dports/graphics/luminance-qt5/luminance-hdr-2.6.1.1/src/ |
H A D | sleefsseavx.c | 68 static INLINE vdouble vadd5(vdouble v0, vdouble v1, vdouble v2, vdouble v3, vdouble v4) { in vadd5() function 73 return vadd5(vadd(v0, v1), v2, v3, v4, v5); in vadd6() 188 vdouble u = vadd5(vsub(vmul(nhh, th), q.x), vmul(nhh, tl), vmul(nhl, th), vmul(nhl, tl), in div_dd() 202 r.y = vadd5(vmul(xh, yh), vneg(r.x), vmul(xl, yh), vmul(xh, yl), vmul(xl, yl)); in mul_ss() 234 …r.y = vadd5(vmul(xh, xh), vneg(r.x), vmul(vadd(xh, xh), xl), vmul(xl, xl), vmul(x.x, vadd(x.y, x.y… in squ_d()
|
/dports/graphics/rawtherapee/rawtherapee-5.8/rtengine/ |
H A D | sleefsseavx.c | 58 static INLINE vdouble vadd5(vdouble v0, vdouble v1, vdouble v2, vdouble v3, vdouble v4) { in vadd5() function 63 return vadd5(vadd(v0, v1), v2, v3, v4, v5); in vadd6() 178 vdouble u = vadd5(vsub(vmul(nhh, th), q.x), vmul(nhh, tl), vmul(nhl, th), vmul(nhl, tl), in div_dd() 192 r.y = vadd5(vmul(xh, yh), vneg(r.x), vmul(xl, yh), vmul(xh, yl), vmul(xl, yl)); in mul_ss() 224 …r.y = vadd5(vmul(xh, xh), vneg(r.x), vmul(vadd(xh, xh), xl), vmul(xl, xl), vmul(x.x, vadd(x.y, x.y… in squ_d()
|
/dports/graphics/photoflow/PhotoFlow-8472024f/src/rt/rtengine/ |
H A D | sleefsseavx.c | 67 static INLINE vdouble vadd5(vdouble v0, vdouble v1, vdouble v2, vdouble v3, vdouble v4) { in vadd5() function 72 return vadd5(vadd(v0, v1), v2, v3, v4, v5); in vadd6() 187 vdouble u = vadd5(vsub(vmul(nhh, th), q.x), vmul(nhh, tl), vmul(nhl, th), vmul(nhl, tl), in div_dd() 201 r.y = vadd5(vmul(xh, yh), vneg(r.x), vmul(xl, yh), vmul(xh, yl), vmul(xl, yl)); in mul_ss() 233 …r.y = vadd5(vmul(xh, xh), vneg(r.x), vmul(vadd(xh, xh), xl), vmul(xl, xl), vmul(x.x, vadd(x.y, x.y… in squ_d()
|
/dports/graphics/photoflow/PhotoFlow-8472024f/src/external/librtprocess/src/include/ |
H A D | sleefsseavx.c | 58 static INLINE vdouble vadd5(vdouble v0, vdouble v1, vdouble v2, vdouble v3, vdouble v4) { in vadd5() function 63 return vadd5(vadd(v0, v1), v2, v3, v4, v5); in vadd6() 178 vdouble u = vadd5(vsub(vmul(nhh, th), q.x), vmul(nhh, tl), vmul(nhl, th), vmul(nhl, tl), in div_dd() 192 r.y = vadd5(vmul(xh, yh), vneg(r.x), vmul(xl, yh), vmul(xh, yl), vmul(xl, yl)); in mul_ss() 224 …r.y = vadd5(vmul(xh, xh), vneg(r.x), vmul(vadd(xh, xh), xl), vmul(xl, xl), vmul(x.x, vadd(x.y, x.y… in squ_d()
|
/dports/devel/llvm-cheri/llvm-project-37c49ff00e3eadce5d8703fdc4497f28458c64a8/llvm/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 345 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 351 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/devel/llvm11/llvm-11.0.1.src/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 345 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 351 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/devel/llvm-devel/llvm-project-f05c95f10fc1d8171071735af8ad3a9e87633120/llvm/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 345 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 351 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/devel/llvm12/llvm-project-12.0.1.src/llvm/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 345 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 351 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/devel/llvm10/llvm-10.0.1.src/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 345 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 351 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/devel/tinygo/tinygo-0.14.1/llvm-project/llvm/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 345 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 351 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/devel/wasi-compiler-rt13/llvm-project-13.0.1.src/llvm/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 345 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 351 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/devel/llvm90/llvm-9.0.1.src/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 345 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 351 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/llvm/llvm/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 345 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 351 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/devel/wasi-compiler-rt12/llvm-project-12.0.1.src/llvm/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 345 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 351 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/devel/wasi-libcxx/llvm-project-13.0.1.src/llvm/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 345 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 351 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/devel/llvm80/llvm-8.0.1.src/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 346 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 352 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/graphics/llvm-mesa/llvm-13.0.1.src/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 345 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 351 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/devel/llvm70/llvm-7.0.1.src/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 346 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 352 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|
/dports/devel/llvm13/llvm-project-13.0.1.src/llvm/test/Transforms/LoopStrengthReduce/ARM/ |
H A D | ivchain-ARM.ll | 345 %vadd5 = tail call <8 x i8> @llvm.arm.neon.vhaddu.v8i8(<8 x i8> %vld5, <8 x i8> %vld6) nounwind 351 tail call void @llvm.arm.neon.vst1.p0i8.v8i8(i8* %add.ptr17, <8 x i8> %vadd5, i32 1)
|