/dports/misc/mnn/MNN-1.2.0/source/backend/cpu/ |
H A D | CPUDepthwiseConvInt8.cpp | 102 const int dst_depth_quad = UP_DIV(output->channel(), UNIT); in onResize() local 110 mThreadNumber = std::min(threadNumber, dst_depth_quad * input->batch()); in onResize() 140 const int dst_depth_quad = UP_DIV(output->channel(), UNIT); in onExecute() local 150 auto totalCount = batch * dst_depth_quad; in onExecute() 163 int dz = index % dst_depth_quad; in onExecute()
|
H A D | CPUConvolutionDepthwise.cpp | 133 int dst_depth_quad = UP_DIV(layer->outputCount(), unit); in onResize() local 180 int numberThread = std::min(((CPUBackend*)backend())->threadNumber(), dst_depth_quad); in onResize() 202 int total = inputs[0]->batch() * dst_depth_quad; in onResize() 205 int dz = index % dst_depth_quad; in onResize()
|
H A D | CPUDeconvolutionDepthwise.cpp | 103 int dst_depth_quad = UP_DIV(layer->outputCount(), core->pack); in onResize() local 155 int totalSize = batch * dst_depth_quad; in onResize() 160 auto zPos = dz % dst_depth_quad; in onResize()
|
H A D | CPUTFQuantizedConv2D.cpp | 47 size_t dst_step, size_t dst_depth_quad); 53 size_t dst_step, size_t dst_depth_quad) { in MNNGemmint8to32_8x4_Unit() argument 54 for (int dz = 0; dz < dst_depth_quad; ++dz) { in MNNGemmint8to32_8x4_Unit()
|
/dports/misc/mnn/MNN-1.2.0/source/backend/cpu/x86_x64/avx/ |
H A D | FunctionSummary.hpp | 39 … size_t dst_step, size_t dst_depth_quad, size_t width, size_t weight_depth_offset); 41 size_t dst_step, size_t dst_depth_quad, size_t weight_depth_offset); 58 …const int8_t* weight, size_t src_depth_quad, size_t dst_step, size_t dst_depth_quad, const QuanPos… 59 …const int8_t* weight, size_t src_depth_quad, size_t dst_step, size_t dst_depth_quad, const QuanPos…
|
H A D | GemmInt8.cpp | 26 …const int8_t* weight, size_t src_depth_quad, size_t dst_step, size_t dst_depth_quad, const QuanPos… in _AVX_MNNGemmInt8AddBiasScale_16x4_Unit() argument 31 strides[2] = dst_depth_quad; in _AVX_MNNGemmInt8AddBiasScale_16x4_Unit() 55 for (int dz = 0; dz < dst_depth_quad; ++dz) { in _AVX_MNNGemmInt8AddBiasScale_16x4_Unit() 209 for (int dz = 0; dz < dst_depth_quad; ++dz) { in _AVX_MNNGemmInt8AddBiasScale_16x4_Unit() 359 for (int dz = 0; dz < dst_depth_quad; ++dz) { in _AVX_MNNGemmInt8AddBiasScale_16x4_Unit() 501 for (int dz = 0; dz < dst_depth_quad; ++dz) { in _AVX_MNNGemmInt8AddBiasScale_16x4_Unit() 635 …const int8_t* weight, size_t src_depth_quad, size_t dst_step, size_t dst_depth_quad, const QuanPos… in _AVX_MNNGemmInt8AddBiasScale_16x4_Unit_Fast() argument 644 for (int dz = 0; dz < dst_depth_quad; ++dz) { in _AVX_MNNGemmInt8AddBiasScale_16x4_Unit_Fast() 739 for (int dz = 0; dz < dst_depth_quad; ++dz) { in _AVX_MNNGemmInt8AddBiasScale_16x4_Unit_Fast() 835 for (int dz = 0; dz < dst_depth_quad; ++dz) { in _AVX_MNNGemmInt8AddBiasScale_16x4_Unit_Fast() [all …]
|
/dports/misc/mnn/MNN-1.2.0/source/backend/cpu/x86_x64/avxfma/ |
H A D | FunctionSummary.hpp | 37 … size_t dst_step, size_t dst_depth_quad, size_t width, size_t weight_depth_offset); 40 size_t dst_step, size_t dst_depth_quad, size_t weight_depth_offset);
|
/dports/misc/mnn/MNN-1.2.0/source/backend/cpu/x86_x64/sse/ |
H A D | FunctionSummary.hpp | 40 size_t dst_step, size_t dst_depth_quad, size_t weight_depth_offset); 45 … size_t dst_step, size_t dst_depth_quad, size_t width, size_t weight_depth_offset); 75 … size_t dst_depth_quad, const QuanPostTreatParameters* post, size_t realDst);
|
H A D | GemmCommon.cpp | 174 … size_t dst_depth_quad, const QuanPostTreatParameters* post, size_t realDst) { in _SSE_MNNGemmInt8AddBiasScale_16x4_Unit() argument 182 for (int dz = 0; dz < dst_depth_quad; ++dz) { in _SSE_MNNGemmInt8AddBiasScale_16x4_Unit()
|
/dports/misc/mnn/MNN-1.2.0/source/backend/cpu/compute/ |
H A D | Int8FunctionsOpt.h | 58 …const int8_t* weight, size_t src_depth_quad, size_t dst_step, size_t dst_depth_quad, const QuanPos… 59 …const int8_t* weight, size_t src_depth_quad, size_t dst_step, size_t dst_depth_quad, const QuanPos…
|
H A D | Int8FunctionsOpt.cpp | 19 …t int8_t* src, const int8_t* weight, size_t src_depth_quad, size_t dst_step, size_t dst_depth_quad, 21 …t int8_t* src, const int8_t* weight, size_t src_depth_quad, size_t dst_step, size_t dst_depth_quad, 26 …t int8_t* src, const int8_t* weight, size_t src_depth_quad, size_t dst_step, size_t dst_depth_quad, 42 … size_t dst_depth_quad, const QuanPostTreatParameters* post, size_t realCount) { in MNNGemmInt8AddBiasScale_16x4_Unit() argument 44 for (int dz = 0; dz < dst_depth_quad; ++dz) { in MNNGemmInt8AddBiasScale_16x4_Unit() 77 …const int8_t* weight, size_t src_depth_quad, size_t dst_step, size_t dst_depth_quad, const QuanPos… in MNNGemmInt8AddBiasScale_16x4_Unit_FAST() argument 78 …t8AddBiasScale_16x4_Unit(dst, src, weight, src_depth_quad, dst_step, dst_depth_quad, post, realCou… in MNNGemmInt8AddBiasScale_16x4_Unit_FAST()
|
/dports/misc/mnn/MNN-1.2.0/source/backend/opengl/ |
H A D | GLConvolutionDepthwise.cpp | 119 int dst_depth_quad = UP_DIV(output->channel(), 4); in onExecute() local 152 UP_DIV(dst_depth_quad, gZLocal)); in onExecute()
|
/dports/misc/mnn/MNN-1.2.0/source/backend/cpu/x86_x64/avx512/ |
H A D | GemmCommon.cpp | 19 …const int8_t* weight, size_t src_depth_quad, size_t dst_step, size_t dst_depth_quad, const QuanPos… in _AVX512_MNNGemmInt8AddBiasScale_16x4_Unit() argument 27 for (int dz = 0; dz < dst_depth_quad; ++dz) { in _AVX512_MNNGemmInt8AddBiasScale_16x4_Unit() 151 for (int dz = 0; dz < dst_depth_quad; ++dz) { in _AVX512_MNNGemmInt8AddBiasScale_16x4_Unit() 258 for (int dz = 0; dz < dst_depth_quad; ++dz) { in _AVX512_MNNGemmInt8AddBiasScale_16x4_Unit() 340 for (int dz = 0; dz < dst_depth_quad; ++dz) { in _AVX512_MNNGemmInt8AddBiasScale_16x4_Unit()
|
H A D | FunctionSummary.hpp | 39 …const int8_t* weight, size_t src_depth_quad, size_t dst_step, size_t dst_depth_quad, const QuanPos…
|