/dports/science/elk/elk-7.2.42/src/ |
H A D | plotpt1d.f90 | 62 vpl(:,i)=vvl(:,1) 71 vl(:)=vvl(:,i+1)-vvl(:,i) 91 vpl(:,k)=vvl(:,i)*(1.d0-f)+vvl(:,i+1)*f
|
/dports/print/ghostscript7-base/ghostscript-7.07/gimp-print-4.2.7/scripts/ |
H A D | mkchlog | 79 my (@vvl) = split /\./, $vl; 80 if ($#vvl > $#vvw) { 84 for ($i = 0; $i < $#vvl; $i += 2) { 85 my ($bl) = $vvl[$i]; 86 my ($vl) = $vvl[$i + 1];
|
/dports/devel/llvm-devel/llvm-project-f05c95f10fc1d8171071735af8ad3a9e87633120/llvm/test/CodeGen/VE/VELIntrinsics/ |
H A D | vseq.ll | 31 %2 = tail call fast <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double> %0, i32 256) 36 declare <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double>, i32) 61 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double> %0, i32 256) 66 declare <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double>, i32) 91 %2 = tail call fast <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double> %0, i32 256) 96 declare <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double>, i32) 121 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double> %0, i32 256) 126 declare <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double>, i32)
|
H A D | vrmin.ll | 16 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswfstsx.vvl(<256 x double> %0, i32 256) 21 declare <256 x double> @llvm.ve.vl.vrminswfstsx.vvl(<256 x double>, i32) 49 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswlstsx.vvl(<256 x double> %0, i32 256) 54 declare <256 x double> @llvm.ve.vl.vrminswlstsx.vvl(<256 x double>, i32) 82 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswfstzx.vvl(<256 x double> %0, i32 256) 87 declare <256 x double> @llvm.ve.vl.vrminswfstzx.vvl(<256 x double>, i32) 120 declare <256 x double> @llvm.ve.vl.vrminswlstzx.vvl(<256 x double>, i32) 148 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminslfst.vvl(<256 x double> %0, i32 256) 153 declare <256 x double> @llvm.ve.vl.vrminslfst.vvl(<256 x double>, i32) 181 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminsllst.vvl(<256 x double> %0, i32 256) [all …]
|
H A D | vrsqrt.ll | 17 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrtd.vvl(<256 x double> %0, i32 256) 22 declare <256 x double> @llvm.ve.vl.vrsqrtd.vvl(<256 x double>, i32) 50 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrts.vvl(<256 x double> %0, i32 256) 55 declare <256 x double> @llvm.ve.vl.vrsqrts.vvl(<256 x double>, i32) 83 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrtdnex.vvl(<256 x double> %0, i32 256) 88 declare <256 x double> @llvm.ve.vl.vrsqrtdnex.vvl(<256 x double>, i32) 116 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrtsnex.vvl(<256 x double> %0, i32 256) 121 declare <256 x double> @llvm.ve.vl.vrsqrtsnex.vvl(<256 x double>, i32) 149 %2 = tail call fast <256 x double> @llvm.ve.vl.pvrsqrt.vvl(<256 x double> %0, i32 256) 154 declare <256 x double> @llvm.ve.vl.pvrsqrt.vvl(<256 x double>, i32) [all …]
|
H A D | vrmax.ll | 16 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswfstsx.vvl(<256 x double> %0, i32 256) 21 declare <256 x double> @llvm.ve.vl.vrmaxswfstsx.vvl(<256 x double>, i32) 49 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswlstsx.vvl(<256 x double> %0, i32 256) 54 declare <256 x double> @llvm.ve.vl.vrmaxswlstsx.vvl(<256 x double>, i32) 82 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswfstzx.vvl(<256 x double> %0, i32 256) 87 declare <256 x double> @llvm.ve.vl.vrmaxswfstzx.vvl(<256 x double>, i32) 120 declare <256 x double> @llvm.ve.vl.vrmaxswlstzx.vvl(<256 x double>, i32) 148 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxslfst.vvl(<256 x double> %0, i32 256) 153 declare <256 x double> @llvm.ve.vl.vrmaxslfst.vvl(<256 x double>, i32) 181 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxsllst.vvl(<256 x double> %0, i32 256) [all …]
|
/dports/devel/wasi-libcxx/llvm-project-13.0.1.src/llvm/test/CodeGen/VE/VELIntrinsics/ |
H A D | vseq.ll | 31 %2 = tail call fast <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double> %0, i32 256) 36 declare <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double>, i32) 61 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double> %0, i32 256) 66 declare <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double>, i32) 91 %2 = tail call fast <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double> %0, i32 256) 96 declare <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double>, i32) 121 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double> %0, i32 256) 126 declare <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double>, i32)
|
H A D | vrmin.ll | 16 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswfstsx.vvl(<256 x double> %0, i32 256) 21 declare <256 x double> @llvm.ve.vl.vrminswfstsx.vvl(<256 x double>, i32) 49 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswlstsx.vvl(<256 x double> %0, i32 256) 54 declare <256 x double> @llvm.ve.vl.vrminswlstsx.vvl(<256 x double>, i32) 82 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswfstzx.vvl(<256 x double> %0, i32 256) 87 declare <256 x double> @llvm.ve.vl.vrminswfstzx.vvl(<256 x double>, i32) 120 declare <256 x double> @llvm.ve.vl.vrminswlstzx.vvl(<256 x double>, i32) 148 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminslfst.vvl(<256 x double> %0, i32 256) 153 declare <256 x double> @llvm.ve.vl.vrminslfst.vvl(<256 x double>, i32) 181 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminsllst.vvl(<256 x double> %0, i32 256) [all …]
|
H A D | vrmax.ll | 16 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswfstsx.vvl(<256 x double> %0, i32 256) 21 declare <256 x double> @llvm.ve.vl.vrmaxswfstsx.vvl(<256 x double>, i32) 49 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswlstsx.vvl(<256 x double> %0, i32 256) 54 declare <256 x double> @llvm.ve.vl.vrmaxswlstsx.vvl(<256 x double>, i32) 82 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswfstzx.vvl(<256 x double> %0, i32 256) 87 declare <256 x double> @llvm.ve.vl.vrmaxswfstzx.vvl(<256 x double>, i32) 120 declare <256 x double> @llvm.ve.vl.vrmaxswlstzx.vvl(<256 x double>, i32) 148 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxslfst.vvl(<256 x double> %0, i32 256) 153 declare <256 x double> @llvm.ve.vl.vrmaxslfst.vvl(<256 x double>, i32) 181 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxsllst.vvl(<256 x double> %0, i32 256) [all …]
|
/dports/graphics/llvm-mesa/llvm-13.0.1.src/test/CodeGen/VE/VELIntrinsics/ |
H A D | vseq.ll | 31 %2 = tail call fast <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double> %0, i32 256) 36 declare <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double>, i32) 61 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double> %0, i32 256) 66 declare <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double>, i32) 91 %2 = tail call fast <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double> %0, i32 256) 96 declare <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double>, i32) 121 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double> %0, i32 256) 126 declare <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double>, i32)
|
H A D | vrmax.ll | 16 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswfstsx.vvl(<256 x double> %0, i32 256) 21 declare <256 x double> @llvm.ve.vl.vrmaxswfstsx.vvl(<256 x double>, i32) 49 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswlstsx.vvl(<256 x double> %0, i32 256) 54 declare <256 x double> @llvm.ve.vl.vrmaxswlstsx.vvl(<256 x double>, i32) 82 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswfstzx.vvl(<256 x double> %0, i32 256) 87 declare <256 x double> @llvm.ve.vl.vrmaxswfstzx.vvl(<256 x double>, i32) 120 declare <256 x double> @llvm.ve.vl.vrmaxswlstzx.vvl(<256 x double>, i32) 148 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxslfst.vvl(<256 x double> %0, i32 256) 153 declare <256 x double> @llvm.ve.vl.vrmaxslfst.vvl(<256 x double>, i32) 181 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxsllst.vvl(<256 x double> %0, i32 256) [all …]
|
H A D | vrmin.ll | 16 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswfstsx.vvl(<256 x double> %0, i32 256) 21 declare <256 x double> @llvm.ve.vl.vrminswfstsx.vvl(<256 x double>, i32) 49 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswlstsx.vvl(<256 x double> %0, i32 256) 54 declare <256 x double> @llvm.ve.vl.vrminswlstsx.vvl(<256 x double>, i32) 82 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswfstzx.vvl(<256 x double> %0, i32 256) 87 declare <256 x double> @llvm.ve.vl.vrminswfstzx.vvl(<256 x double>, i32) 120 declare <256 x double> @llvm.ve.vl.vrminswlstzx.vvl(<256 x double>, i32) 148 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminslfst.vvl(<256 x double> %0, i32 256) 153 declare <256 x double> @llvm.ve.vl.vrminslfst.vvl(<256 x double>, i32) 181 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminsllst.vvl(<256 x double> %0, i32 256) [all …]
|
/dports/devel/llvm12/llvm-project-12.0.1.src/llvm/test/CodeGen/VE/VELIntrinsics/ |
H A D | vseq.ll | 31 %2 = tail call fast <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double> %0, i32 256) 36 declare <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double>, i32) 61 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double> %0, i32 256) 66 declare <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double>, i32) 91 %2 = tail call fast <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double> %0, i32 256) 96 declare <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double>, i32) 121 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double> %0, i32 256) 126 declare <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double>, i32)
|
H A D | vrmax.ll | 16 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswfstsx.vvl(<256 x double> %0, i32 256) 21 declare <256 x double> @llvm.ve.vl.vrmaxswfstsx.vvl(<256 x double>, i32) 49 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswlstsx.vvl(<256 x double> %0, i32 256) 54 declare <256 x double> @llvm.ve.vl.vrmaxswlstsx.vvl(<256 x double>, i32) 82 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswfstzx.vvl(<256 x double> %0, i32 256) 87 declare <256 x double> @llvm.ve.vl.vrmaxswfstzx.vvl(<256 x double>, i32) 120 declare <256 x double> @llvm.ve.vl.vrmaxswlstzx.vvl(<256 x double>, i32) 148 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxslfst.vvl(<256 x double> %0, i32 256) 153 declare <256 x double> @llvm.ve.vl.vrmaxslfst.vvl(<256 x double>, i32) 181 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxsllst.vvl(<256 x double> %0, i32 256) [all …]
|
H A D | vrmin.ll | 16 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswfstsx.vvl(<256 x double> %0, i32 256) 21 declare <256 x double> @llvm.ve.vl.vrminswfstsx.vvl(<256 x double>, i32) 49 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswlstsx.vvl(<256 x double> %0, i32 256) 54 declare <256 x double> @llvm.ve.vl.vrminswlstsx.vvl(<256 x double>, i32) 82 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswfstzx.vvl(<256 x double> %0, i32 256) 87 declare <256 x double> @llvm.ve.vl.vrminswfstzx.vvl(<256 x double>, i32) 120 declare <256 x double> @llvm.ve.vl.vrminswlstzx.vvl(<256 x double>, i32) 148 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminslfst.vvl(<256 x double> %0, i32 256) 153 declare <256 x double> @llvm.ve.vl.vrminslfst.vvl(<256 x double>, i32) 181 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminsllst.vvl(<256 x double> %0, i32 256) [all …]
|
H A D | vrsqrt.ll | 17 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrtd.vvl(<256 x double> %0, i32 256) 22 declare <256 x double> @llvm.ve.vl.vrsqrtd.vvl(<256 x double>, i32) 50 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrts.vvl(<256 x double> %0, i32 256) 55 declare <256 x double> @llvm.ve.vl.vrsqrts.vvl(<256 x double>, i32) 83 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrtdnex.vvl(<256 x double> %0, i32 256) 88 declare <256 x double> @llvm.ve.vl.vrsqrtdnex.vvl(<256 x double>, i32) 116 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrtsnex.vvl(<256 x double> %0, i32 256) 121 declare <256 x double> @llvm.ve.vl.vrsqrtsnex.vvl(<256 x double>, i32) 149 %2 = tail call fast <256 x double> @llvm.ve.vl.pvrsqrt.vvl(<256 x double> %0, i32 256) 154 declare <256 x double> @llvm.ve.vl.pvrsqrt.vvl(<256 x double>, i32) [all …]
|
/dports/devel/wasi-compiler-rt13/llvm-project-13.0.1.src/llvm/test/CodeGen/VE/VELIntrinsics/ |
H A D | vseq.ll | 31 %2 = tail call fast <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double> %0, i32 256) 36 declare <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double>, i32) 61 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double> %0, i32 256) 66 declare <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double>, i32) 91 %2 = tail call fast <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double> %0, i32 256) 96 declare <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double>, i32) 121 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double> %0, i32 256) 126 declare <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double>, i32)
|
H A D | vrmax.ll | 16 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswfstsx.vvl(<256 x double> %0, i32 256) 21 declare <256 x double> @llvm.ve.vl.vrmaxswfstsx.vvl(<256 x double>, i32) 49 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswlstsx.vvl(<256 x double> %0, i32 256) 54 declare <256 x double> @llvm.ve.vl.vrmaxswlstsx.vvl(<256 x double>, i32) 82 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswfstzx.vvl(<256 x double> %0, i32 256) 87 declare <256 x double> @llvm.ve.vl.vrmaxswfstzx.vvl(<256 x double>, i32) 120 declare <256 x double> @llvm.ve.vl.vrmaxswlstzx.vvl(<256 x double>, i32) 148 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxslfst.vvl(<256 x double> %0, i32 256) 153 declare <256 x double> @llvm.ve.vl.vrmaxslfst.vvl(<256 x double>, i32) 181 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxsllst.vvl(<256 x double> %0, i32 256) [all …]
|
H A D | vrmin.ll | 16 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswfstsx.vvl(<256 x double> %0, i32 256) 21 declare <256 x double> @llvm.ve.vl.vrminswfstsx.vvl(<256 x double>, i32) 49 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswlstsx.vvl(<256 x double> %0, i32 256) 54 declare <256 x double> @llvm.ve.vl.vrminswlstsx.vvl(<256 x double>, i32) 82 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswfstzx.vvl(<256 x double> %0, i32 256) 87 declare <256 x double> @llvm.ve.vl.vrminswfstzx.vvl(<256 x double>, i32) 120 declare <256 x double> @llvm.ve.vl.vrminswlstzx.vvl(<256 x double>, i32) 148 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminslfst.vvl(<256 x double> %0, i32 256) 153 declare <256 x double> @llvm.ve.vl.vrminslfst.vvl(<256 x double>, i32) 181 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminsllst.vvl(<256 x double> %0, i32 256) [all …]
|
H A D | vrsqrt.ll | 17 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrtd.vvl(<256 x double> %0, i32 256) 22 declare <256 x double> @llvm.ve.vl.vrsqrtd.vvl(<256 x double>, i32) 50 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrts.vvl(<256 x double> %0, i32 256) 55 declare <256 x double> @llvm.ve.vl.vrsqrts.vvl(<256 x double>, i32) 83 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrtdnex.vvl(<256 x double> %0, i32 256) 88 declare <256 x double> @llvm.ve.vl.vrsqrtdnex.vvl(<256 x double>, i32) 116 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrtsnex.vvl(<256 x double> %0, i32 256) 121 declare <256 x double> @llvm.ve.vl.vrsqrtsnex.vvl(<256 x double>, i32) 149 %2 = tail call fast <256 x double> @llvm.ve.vl.pvrsqrt.vvl(<256 x double> %0, i32 256) 154 declare <256 x double> @llvm.ve.vl.pvrsqrt.vvl(<256 x double>, i32) [all …]
|
/dports/devel/wasi-compiler-rt12/llvm-project-12.0.1.src/llvm/test/CodeGen/VE/VELIntrinsics/ |
H A D | vseq.ll | 31 %2 = tail call fast <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double> %0, i32 256) 36 declare <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double>, i32) 61 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double> %0, i32 256) 66 declare <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double>, i32) 91 %2 = tail call fast <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double> %0, i32 256) 96 declare <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double>, i32) 121 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double> %0, i32 256) 126 declare <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double>, i32)
|
H A D | vrmax.ll | 16 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswfstsx.vvl(<256 x double> %0, i32 256) 21 declare <256 x double> @llvm.ve.vl.vrmaxswfstsx.vvl(<256 x double>, i32) 49 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswlstsx.vvl(<256 x double> %0, i32 256) 54 declare <256 x double> @llvm.ve.vl.vrmaxswlstsx.vvl(<256 x double>, i32) 82 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxswfstzx.vvl(<256 x double> %0, i32 256) 87 declare <256 x double> @llvm.ve.vl.vrmaxswfstzx.vvl(<256 x double>, i32) 120 declare <256 x double> @llvm.ve.vl.vrmaxswlstzx.vvl(<256 x double>, i32) 148 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxslfst.vvl(<256 x double> %0, i32 256) 153 declare <256 x double> @llvm.ve.vl.vrmaxslfst.vvl(<256 x double>, i32) 181 %2 = tail call fast <256 x double> @llvm.ve.vl.vrmaxsllst.vvl(<256 x double> %0, i32 256) [all …]
|
H A D | vrmin.ll | 16 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswfstsx.vvl(<256 x double> %0, i32 256) 21 declare <256 x double> @llvm.ve.vl.vrminswfstsx.vvl(<256 x double>, i32) 49 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswlstsx.vvl(<256 x double> %0, i32 256) 54 declare <256 x double> @llvm.ve.vl.vrminswlstsx.vvl(<256 x double>, i32) 82 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminswfstzx.vvl(<256 x double> %0, i32 256) 87 declare <256 x double> @llvm.ve.vl.vrminswfstzx.vvl(<256 x double>, i32) 120 declare <256 x double> @llvm.ve.vl.vrminswlstzx.vvl(<256 x double>, i32) 148 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminslfst.vvl(<256 x double> %0, i32 256) 153 declare <256 x double> @llvm.ve.vl.vrminslfst.vvl(<256 x double>, i32) 181 %2 = tail call fast <256 x double> @llvm.ve.vl.vrminsllst.vvl(<256 x double> %0, i32 256) [all …]
|
H A D | vrsqrt.ll | 17 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrtd.vvl(<256 x double> %0, i32 256) 22 declare <256 x double> @llvm.ve.vl.vrsqrtd.vvl(<256 x double>, i32) 50 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrts.vvl(<256 x double> %0, i32 256) 55 declare <256 x double> @llvm.ve.vl.vrsqrts.vvl(<256 x double>, i32) 83 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrtdnex.vvl(<256 x double> %0, i32 256) 88 declare <256 x double> @llvm.ve.vl.vrsqrtdnex.vvl(<256 x double>, i32) 116 %2 = tail call fast <256 x double> @llvm.ve.vl.vrsqrtsnex.vvl(<256 x double> %0, i32 256) 121 declare <256 x double> @llvm.ve.vl.vrsqrtsnex.vvl(<256 x double>, i32) 149 %2 = tail call fast <256 x double> @llvm.ve.vl.pvrsqrt.vvl(<256 x double> %0, i32 256) 154 declare <256 x double> @llvm.ve.vl.pvrsqrt.vvl(<256 x double>, i32) [all …]
|
/dports/devel/llvm13/llvm-project-13.0.1.src/llvm/test/CodeGen/VE/VELIntrinsics/ |
H A D | vseq.ll | 31 %2 = tail call fast <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double> %0, i32 256) 36 declare <256 x double> @llvm.ve.vl.vseq.vvl(<256 x double>, i32) 61 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double> %0, i32 256) 66 declare <256 x double> @llvm.ve.vl.pvseqlo.vvl(<256 x double>, i32) 91 %2 = tail call fast <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double> %0, i32 256) 96 declare <256 x double> @llvm.ve.vl.pvsequp.vvl(<256 x double>, i32) 121 %2 = tail call fast <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double> %0, i32 256) 126 declare <256 x double> @llvm.ve.vl.pvseq.vvl(<256 x double>, i32)
|