Searched refs:onValueTensor (Results 1 – 5 of 5) sorted by relevance
/dports/misc/mnn/MNN-1.2.0/source/backend/cpu/ |
H A D | CPUOneHot.cpp | 15 …neHotImpl(int depth, int outerSize, int innerSize, const int* indices, const Tensor* onValueTensor, in OneHotImpl() argument 17 const T onValue = onValueTensor->host<T>()[0]; in OneHotImpl() 39 auto onValueTensor = inputs[2]; in onExecute() local 54 auto dataType = onValueTensor->getType(); in onExecute() 58 …OneHotImpl<float>(depth, outerSize, innerSize, indicesPtr, onValueTensor, offValueTensor, outputs[… in onExecute() 60 …OneHotImpl<int>(depth, outerSize, innerSize, indicesPtr, onValueTensor, offValueTensor, outputs[0]… in onExecute()
|
/dports/misc/mnn/MNN-1.2.0/source/backend/tensorrt/execution/plugin/ |
H A D | OneHotPlugin.cu | 25 …nst int count, const float* depth, int innerSize, const float* indices, const float* onValueTensor, in OneHotExecute() argument 30 …_BLOCKS(count), CUDA_NUM_THREADS>>>(count, depth, innerSize, indices, onValueTensor, offValueTenso… in OneHotExecute() 32 …DS>>>(count, depth, innerSize, (const __half*)indices, (const __half*)onValueTensor, (const __half… in OneHotExecute()
|
H A D | OneHotPlugin.cpp | 26 auto onValueTensor = reinterpret_cast<const float*>(inputs[2]); in onEnqueue() local 28 …return OneHotExecute(dataType, mOuterSize, depthTensor, mInnerSize, indices, onValueTensor, offVal… in onEnqueue()
|
H A D | OneHotPlugin.hpp | 20 …nst int count, const float* depth, int innerSize, const float* indices, const float* onValueTensor,
|
/dports/misc/mnn/MNN-1.2.0/source/backend/tensorrt/execution/ |
H A D | TRTOneHot.cpp | 32 auto onValueTensor = mInputs[2]; in onEncode() local 46 auto dataType = onValueTensor->getType(); in onEncode()
|