/dports/misc/mnn/MNN-1.2.0/source/backend/cpu/compute/ |
H A D | Int8FunctionsOpt.cpp | 145 static void _fastIm2Col(int8_t* colAddr, const int8_t* inputOrigin, int8_t inputZeroPoint, in _fastIm2Col() argument 149 …::memset(colAddr, inputZeroPoint, col_buffer_size); // the padding process, since per-channel is r… in _fastIm2Col() 171 static void _im2colCommonZ1(int8_t* colAddr, const int8_t* inputOrigin, int8_t inputZeroPoint, in _im2colCommonZ1() argument 175 …::memset(colAddr, inputZeroPoint, col_buffer_size); // the padding process, since per-channel is r… in _im2colCommonZ1() 217 static void _im2colCommon(int8_t* colAddr, const int8_t* inputOrigin, int8_t inputZeroPoint, in _im2colCommon() argument 221 …::memset(colAddr, inputZeroPoint, col_buffer_size); // the padding process, since per-channel is r… in _im2colCommon() 304 static void _im2colCommonSdot(int8_t* colAddr, const int8_t* src, int8_t inputZeroPoint, in _im2colCommonSdot() argument 308 memset(colAddr, inputZeroPoint, colBufferSize); in _im2colCommonSdot() 353 static void _fastIm2ColSdot(int8_t* colAddr, const int8_t* inputOrigin, int8_t inputZeroPoint, in _fastIm2ColSdot() argument 357 ::memset(colAddr, inputZeroPoint, col_buffer_size); in _fastIm2ColSdot()
|
H A D | OptimizedComputer.cpp | 126 void Logistic(const uint8_t* input_data, const std::vector<int>& input_dims, int32_t inputZeroPoint, in Logistic() argument 142 …vsubq_s16(vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(input_val_u8))), vdupq_n_s16(inputZeroPoint)); in Logistic() 144 …subq_s16(vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(input_val_u8))), vdupq_n_s16(inputZeroPoint)); in Logistic() 213 const int32_t input_val_centered = static_cast<int32_t>(input_val_u8) - inputZeroPoint; in Logistic()
|
H A D | OptimizedComputer.hpp | 29 void Logistic(const uint8_t* input_data, const std::vector<int>& input_dims, int32_t inputZeroPoint,
|
H A D | Int8FunctionsOpt.h | 62 typedef void(*Im2ColFunc)(int8_t* colAddr, const int8_t* inputOrigin, int8_t inputZeroPoint,
|
/dports/misc/mnn/MNN-1.2.0/pymnn/pip_package/MNN/tools/mnn_fb/ |
H A D | QuantizedConcat.py | 95 …eroPoint(builder, inputZeroPoint): builder.PrependUOffsetTRelativeSlot(3, flatbuffers.number_types… argument
|
/dports/misc/mnn/MNN-1.2.0/tools/converter/source/optimizer/merge/ |
H A D | ConvBNReluFuseToConvInt8.cpp | 83 float inputZeroPoint = int8ToFloatOp->main.AsQuantizedFloatParam()->zeroPoint; in __anon801be1270102() local 157 …auto remains = _ReduceSum(_Scalar<int32_t>(inputZeroPoint) * _Cast<int32_t>(quanWeight), {1, 2, 3}… in __anon801be1270102() 207 conv->symmetricQuan->zeroPoint = std::move(int8_t(inputZeroPoint)); in __anon801be1270102()
|
/dports/misc/mnn/MNN-1.2.0/source/backend/cpu/ |
H A D | CPUConvolution.cpp | 55 float inputZeroPoint = inputQuantInfo[1]; in updateInputOutputScale() local 75 int temp = (int(inputZeroPoint) - mInputZeroPoint) * mInt8WeightKernelSum[i]; in updateInputOutputScale() 105 mInputZeroPoint = int8_t(inputZeroPoint); in updateInputOutputScale()
|
/dports/misc/mnn/MNN-1.2.0/source/core/ |
H A D | ConvolutionCommon.hpp | 26 …at*& scale, int32_t*& bias, float inputScale, float outputScale, int inputZeroPoint, int outputZer…
|
H A D | ConvolutionCommon.cpp | 482 … float inputScale, float outputScale, int inputZeroPoint, int outputZeroPoint) { in getConvInt8Parameters() argument 521 temp += inputZeroPoint * weight[offset + j]; in getConvInt8Parameters()
|
/dports/misc/mnn/MNN-1.2.0/tools/converter/source/tflite/ |
H A D | ConcatTflite.cpp | 37 concatParamQuan->inputZeroPoint.push_back(inputTensor->quantization->zero_point[0]); in run()
|
/dports/misc/mnn/MNN-1.2.0/schema/default/ |
H A D | TFQuantizeOp.fbs | 80 inputZeroPoint: [int];
|
/dports/misc/mnn/MNN-1.2.0/schema/current/ |
H A D | TFQuantizeOp_generated.h | 743 std::vector<int32_t> inputZeroPoint; member 765 const flatbuffers::Vector<int32_t> *inputZeroPoint() const { in inputZeroPoint() function 778 verifier.VerifyVector(inputZeroPoint()) && in Verify() 800 void add_inputZeroPoint(flatbuffers::Offset<flatbuffers::Vector<int32_t>> inputZeroPoint) { in add_inputZeroPoint() 801 fbb_.AddOffset(10, inputZeroPoint); in add_inputZeroPoint() 823 flatbuffers::Offset<flatbuffers::Vector<int32_t>> inputZeroPoint = 0, 827 builder_.add_inputZeroPoint(inputZeroPoint); 1943 … = inputZeroPoint(); if (_e) { _o->inputZeroPoint.resize(_e->size()); for (flatbuffers::uoffset_t … in UnPackTo() 1958 auto _inputZeroPoint = _o->inputZeroPoint.size() ? _fbb.CreateVector(_o->inputZeroPoint) : 0; in CreateQuantizedConcat()
|
/dports/misc/mnn/MNN-1.2.0/include/MNN/expr/ |
H A D | NeuralNetWorkOp.hpp | 136 int8_t inputZeroPoint, int8_t outputZeroPoint, 142 int8_t inputZeroPoint, int8_t outputZeroPoint,
|
/dports/misc/mnn/MNN-1.2.0/tools/train/source/nn/ |
H A D | NN.cpp | 770 int8_t inputZeroPoint, outputZeroPoint; in onForward() local 781 inputZeroPoint = zeroPoint->readMap<int8_t>()[0]; in onForward() 783 …utClampValue->readMap<float>()[0]), int8_t(mInputClampValue->readMap<float>()[0]), inputZeroPoint); in onForward() 838 inputZeroPoint, outputZeroPoint, in onForward()
|
/dports/misc/mnn/MNN-1.2.0/express/ |
H A D | NeuralNetWorkOp.cpp | 1521 int8_t inputZeroPoint, int8_t outputZeroPoint, in _Conv() argument 1557 conv2D->symmetricQuan->zeroPoint = std::move(inputZeroPoint); in _Conv() 1575 int8_t inputZeroPoint, int8_t outputZeroPoint, in _Conv() argument 1613 conv2D->symmetricQuan->zeroPoint = std::move(inputZeroPoint); in _Conv()
|