/dports/math/gemmlowp/gemmlowp-dc69acd/meta/ |
H A D | legacy_operations_common.h | 20 Quantized8BitOperation(std::int32_t lhs_offset, std::int32_t rhs_offset, in Quantized8BitOperation() argument 24 rhs_offset(rhs_offset), in Quantized8BitOperation() 31 std::int32_t rhs_offset; variable 39 FloatOperation(std::int32_t lhs_offset, std::int32_t rhs_offset, in FloatOperation() argument 42 rhs_offset(rhs_offset), in FloatOperation() 47 std::int32_t rhs_offset; variable 53 Int32Operation(std::int32_t lhs_offset, std::int32_t rhs_offset) in Int32Operation() argument 54 : lhs_offset(lhs_offset), rhs_offset(rhs_offset) {} in Int32Operation() 58 std::int32_t rhs_offset; variable
|
H A D | legacy_multi_thread_gemv.h | 33 GemvQuantized8BitOperation(std::int32_t lhs_offset, std::int32_t rhs_offset, in GemvQuantized8BitOperation() argument 43 gemv_q8(scratch, lhs, rhs, n, k, lhs_offset, rhs_offset, sum_offset, in ExecuteMatrixMatrix() 55 GemvFloatOperation(std::int32_t lhs_offset, std::int32_t rhs_offset, in GemvFloatOperation() argument 57 : FloatOperation(lhs_offset, rhs_offset, result_offset) {} in GemvFloatOperation() 63 gemv_f(scratch, lhs, rhs, n, k, lhs_offset, rhs_offset, result_offset, in ExecuteMatrixMatrix() 75 GemvInt32Operation(std::int32_t lhs_offset, std::int32_t rhs_offset) in GemvInt32Operation() argument 76 : Int32Operation(lhs_offset, rhs_offset) {} in GemvInt32Operation() 82 gemv_i32(scratch, lhs, rhs, n, k, lhs_offset, rhs_offset, result); in ExecuteMatrixMatrix() 107 internal::GemvQuantized8BitOperation operation(lhs_offset, rhs_offset, in multi_thread_gemv_q8() 127 std::int32_t rhs_offset, float result_offset, in multi_thread_gemv_f() argument [all …]
|
H A D | legacy_multi_thread_gemm.h | 81 gemm_q8_strided(scratch, lhs, rhs, m, n, k, lhs_offset, rhs_offset, in ExecuteCacheFriendlyMatrixMatrix() 93 GemmFloatOperation(std::int32_t lhs_offset, std::int32_t rhs_offset, in GemmFloatOperation() argument 95 : FloatOperation(lhs_offset, rhs_offset, result_offset) {} in GemmFloatOperation() 111 gemm_f_strided(scratch, lhs, rhs, m, n, k, lhs_offset, rhs_offset, in ExecuteCacheFriendlyMatrixMatrix() 123 GemmInt32Operation(std::int32_t lhs_offset, std::int32_t rhs_offset) in GemmInt32Operation() argument 124 : Int32Operation(lhs_offset, rhs_offset) {} in GemmInt32Operation() 176 internal::GemmQuantized8BitOperation operation(lhs_offset, rhs_offset, in multi_thread_gemm_q8() 201 rhs_offset, result_offset, result); in multi_thread_gemm_f() 234 lhs_offset, rhs_offset, result); in multi_thread_gemm_i32() 238 rhs_offset, lhs_offset, result); in multi_thread_gemm_i32() [all …]
|
H A D | legacy_single_thread_gemm.h | 56 params.left_stream.multiplicative_sum_offset = rhs_offset; in gemm_q8_strided() 58 result_offset + k * lhs_offset * rhs_offset; in gemm_q8_strided() 76 std::int32_t lhs_offset, std::int32_t rhs_offset, in gemv_q8() argument 100 params.left_stream.multiplicative_sum_offset = rhs_offset; in gemv_q8() 102 result_offset + k * lhs_offset * rhs_offset; in gemv_q8() 149 params.left_stream.multiplicative_sum_offset = rhs_offset; in gemm_i32_strided() 165 std::int32_t lhs_offset, std::int32_t rhs_offset, in gemv_i32() argument 189 params.left_stream.multiplicative_sum_offset = rhs_offset; in gemv_i32() 233 params.left_stream.multiplicative_sum_offset = rhs_offset; in gemm_f_strided() 250 std::int32_t lhs_offset, std::int32_t rhs_offset, in gemv_f() argument [all …]
|
/dports/math/gemmlowp/gemmlowp-dc69acd/public/ |
H A D | gemmlowp.h | 39 const RhsOffset& rhs_offset, in GemmWithOutputPipelinePC() argument 42 context, lhs, rhs, result, lhs_offset, rhs_offset, output_pipeline); in GemmWithOutputPipelinePC() 57 int lhs_offset, int rhs_offset, in GemmWithOutputPipeline() argument 62 const OffsetRowDup rhs_offset_vector(rhs_offset, rhs.cols()); in GemmWithOutputPipeline() 78 int rhs_offset, int result_offset, int result_mult_int, in Gemm() argument 81 context, lhs, rhs, result, lhs_offset, rhs_offset, in Gemm()
|
/dports/math/gemmlowp/gemmlowp-dc69acd/internal/ |
H A D | unpack.h | 117 LoadForBroadcasting<RegisterBlockType>(rhs_offset, src_col); in UnpackResultBlock() 136 const LhsOffset& lhs_offset, const RhsOffset& rhs_offset, in UnpackResult() argument 190 rhs_offset, depth, r, c, global_row, global_col, 0, cx); in UnpackResult() 203 rhs_offset, depth, r, c, global_row, global_col, global_row, in UnpackResult() 215 rhs_offset, depth, r, c, global_row, global_col, global_row, in UnpackResult() 231 rhs_sums_of_each_slice, lhs_offset, rhs_offset, depth, r, c, in UnpackResult() 238 rhs_sums_of_each_slice, lhs_offset, rhs_offset, depth, r, c, in UnpackResult() 245 rhs_sums_of_each_slice, lhs_offset, rhs_offset, depth, r, c, in UnpackResult() 258 rhs_sums_of_each_slice, lhs_offset, rhs_offset, depth, r, c, in UnpackResult() 265 rhs_sums_of_each_slice, lhs_offset, rhs_offset, depth, r, c, in UnpackResult() [all …]
|
H A D | multi_thread_gemm.h | 485 rhs_offset(_rhs_offset), in GemmWithPackedRhsTask() 519 rhs_offset.block(curr_result_block.start_col, cs), output_pipeline); in Run() 533 const RhsOffset& rhs_offset; member 640 const LhsOffset& lhs_offset, const RhsOffset& rhs_offset, in MultiThreadGemm() argument 663 lhs_offset, rhs_offset, in MultiThreadGemm() 710 lhs_offset, rhs_offset, block_params, output_pipeline)); in MultiThreadGemm()
|
H A D | dispatch_gemm_shape.h | 175 const LhsOffset& lhs_offset, const RhsOffset& rhs_offset, 193 Transpose(rhs_offset), Transpose(lhs_offset), 200 lhs_offset, rhs_offset, output_pipeline);
|
H A D | single_thread_gemm.h | 71 const LhsOffset& lhs_offset, const RhsOffset& rhs_offset, in SingleThreadGemm() argument 148 lhs_offset.block(r, rs), rhs_offset.block(c, cs), output_pipeline); in SingleThreadGemm()
|
/dports/math/gemmlowp/gemmlowp-dc69acd/doc/ |
H A D | low-precision.md | 41 - A int32 rhs_offset, that will be added to each entry of the rhs matrix. 47 2. Cast rhs entries from uint8 to int32 and add rhs_offset to each of them. 64 be interpreted during the matrix computation: - lhs_offset - rhs_offset - 71 2. Cast rhs entries from uint8 to int32 and add rhs_offset to each of them. 105 One may then consider adding the lhs_offset and rhs_offset once and for all to 121 Adding rhs_offset to each entry of `rhs`, means adding `rhs_offset * Q` to 128 (lhs + lhs_offset * P) * (rhs + rhs_offset * Q) 137 + lhs * rhs_offset * Q 138 + lhs_offset * rhs_offset * P * Q 158 single column-vector, and later multiplying these sums by rhs_offset. [all …]
|
H A D | public.md | 21 `lhs_offset`, `rhs_offset` to them, is explained in 39 int lhs_offset, int rhs_offset, 49 &uint8_result_matrix, lhs_offset, rhs_offset, output_pipeline); 94 * `lhs_offset`, `rhs_offset` are constants added to each matrix entry in the 142 This is a variant where `lhs_offset` and `rhs_offset` may be vectors instead of
|
H A D | quantization_example.cc | 333 const int rhs_offset = -rhs_qparams.zero_point; in main() local 364 &actual_uint8_result_map, lhs_offset, rhs_offset, output_pipeline); in main()
|
H A D | quantization.md | 283 mathematical trick to handle `lhs_offset`, `rhs_offset` that we alluded to 287 `lhs_offset`, `rhs_offset` were 0.
|
/dports/math/gemmlowp/gemmlowp-dc69acd/test/ |
H A D | test.cc | 145 const OffsetRowDup rhs_offset_vector(rhs_offset, cols); in Gemm() 186 const OffsetRowDup rhs_offset_vector(rhs_offset, cols); in Gemm() 760 const OffsetRowDup rhs_offset(b_offset, rhs.cols()); in TestWithSmallDataPerChannelQuantization() local 911 const OffsetRowDup rhs_offset(b_offset, rhs.cols()); in TestWithLargeDataPerChannelQuantization() local 1053 const OffsetRowDup rhs_offset(b_offset, rhs.cols()); in TestMultithreadedPerChannelQuantization() local 1209 const int rhs_offset = -34; in TestOutputStages() local 1216 rhs_offset, empty_pipeline); in TestOutputStages() 1242 lhs_offset, rhs_offset, quantize_down_pipeline); in TestOutputStages() 1385 rhs_offset, clamp_pipeline); in TestOutputStages() 1405 rhs_offset, tanh_pipeline); in TestOutputStages() [all …]
|
H A D | correctness_meta_gemm.cc | 64 std::int32_t lhs_offset, std::int32_t rhs_offset, in check_result() argument 75 (static_cast<std::int32_t>(right[depth * j + k]) + rhs_offset); in check_result() 107 std::int32_t lhs_offset, std::int32_t rhs_offset, in check_result_f() argument 116 (static_cast<std::int32_t>(right[depth * j + k]) + rhs_offset); in check_result_f() 141 std::int32_t lhs_offset, std::int32_t rhs_offset) { in check_result_i32() argument 149 (static_cast<std::int32_t>(right[depth * j + k]) + rhs_offset); in check_result_i32()
|
/dports/math/gemmlowp/gemmlowp-dc69acd/eight_bit_int_gemm/ |
H A D | eight_bit_int_gemm.cc | 76 const int rhs_offset = b_offset; in EightBitIntGemmImpl() local 96 context, lhs, rhs, &result, lhs_offset, rhs_offset, result_offset, \ in EightBitIntGemmImpl() 114 const int rhs_offset = b_offset; in EightBitIntGemmInt32Impl() local 133 context, lhs, rhs, &result, lhs_offset, rhs_offset, empty_pipeline); \ in EightBitIntGemmInt32Impl() 251 std::int32_t lhs_offset, std::int32_t rhs_offset, in MetaGemmQuantized8Bit() argument 262 rhs_offset, sum_offset, multiplicative_offset, in MetaGemmQuantized8Bit() 267 scratch->buffer(), rhs, lhs, n, m, k, rhs_offset, in MetaGemmQuantized8Bit() 277 std::int32_t lhs_offset, std::int32_t rhs_offset, in MetaGemmFloat() argument 286 rhs_offset, result_offset, result); in MetaGemmFloat() 290 scratch->buffer(), rhs, lhs, n, m, k, rhs_offset, in MetaGemmFloat()
|
/dports/lang/rust/rustc-1.58.1-src/src/tools/rustfmt/src/ |
H A D | pairs.rs | 97 let rhs_offset = shape.rhs_overhead(context.config); in rewrite_pairs_multiline() localVariable 103 .sub_width(rhs_offset)?; in rewrite_pairs_multiline() 267 let rhs_offset = shape.rhs_overhead(context.config); in flatten() localVariable 273 .sub_width(rhs_offset)?; in flatten()
|
/dports/math/py-or-tools/or-tools-9.2/ortools/sat/ |
H A D | cp_model_lns.cc | 936 int64_t* rhs_offset) { in AddLinearExpressionToConstraint() argument 937 *rhs_offset -= coeff * expr.offset(); in AddLinearExpressionToConstraint() 991 int64_t rhs_offset = 0; in AddPrecedenceConstraints() local 995 AddLinearExpressionToConstraint(1, before_end, linear, &rhs_offset); in AddPrecedenceConstraints() 999 rhs_offset = GetLinearExpressionValue(before_start, initial_solution) - in AddPrecedenceConstraints() 1001 AddLinearExpressionToConstraint(1, before_start, linear, &rhs_offset); in AddPrecedenceConstraints() 1004 AddLinearExpressionToConstraint(-1, after_start, linear, &rhs_offset); in AddPrecedenceConstraints() 1005 linear->add_domain(rhs_offset); in AddPrecedenceConstraints()
|
/dports/math/py-z3-solver/z3-z3-4.8.10/src/ast/rewriter/ |
H A D | poly_rewriter_def.h | 878 const unsigned rhs_offset = insert_c_rhs ? 0 : 1; in cancel_monomials() local 882 …rhs_result = mk_add_app(new_rhs_monomials.size() - rhs_offset, new_rhs_monomials.c_ptr() + rhs_off… in cancel_monomials()
|
/dports/math/z3/z3-z3-4.8.13/src/ast/rewriter/ |
H A D | poly_rewriter_def.h | 878 const unsigned rhs_offset = insert_c_rhs ? 0 : 1; in cancel_monomials() local 882 …rhs_result = mk_add_app(new_rhs_monomials.size() - rhs_offset, new_rhs_monomials.data() + rhs_offs… in cancel_monomials()
|
/dports/devel/arm-none-eabi-gcc492/gcc-4.9.2/gcc/ |
H A D | tree-ssa-sccvn.c | 1857 HOST_WIDE_INT rhs_offset, copy_size, lhs_offset; in vn_reference_lookup_3() local 1896 rhs_offset = 0; in vn_reference_lookup_3() 1902 &rhs_offset); in vn_reference_lookup_3() 1909 rhs_offset += tree_to_uhwi (TREE_OPERAND (tem, 1)); in vn_reference_lookup_3() 1957 op.op0 = build_int_cst (ptr_type_node, at - rhs_offset); in vn_reference_lookup_3() 1958 op.off = at - lhs_offset + rhs_offset; in vn_reference_lookup_3()
|
/dports/lang/gcc48/gcc-4.8.5/gcc/ |
H A D | tree-ssa-sccvn.c | 1767 HOST_WIDE_INT rhs_offset, copy_size, lhs_offset; in vn_reference_lookup_3() local 1806 rhs_offset = 0; in vn_reference_lookup_3() 1812 &rhs_offset); in vn_reference_lookup_3() 1819 rhs_offset += TREE_INT_CST_LOW (TREE_OPERAND (tem, 1)); in vn_reference_lookup_3() 1867 op.op0 = build_int_cst (ptr_type_node, at - rhs_offset); in vn_reference_lookup_3() 1868 op.off = at - lhs_offset + rhs_offset; in vn_reference_lookup_3()
|
/dports/devel/mingw32-gcc/gcc-4.8.1/gcc/ |
H A D | tree-ssa-sccvn.c | 1759 HOST_WIDE_INT rhs_offset, copy_size, lhs_offset; in vn_reference_lookup_3() local 1798 rhs_offset = 0; in vn_reference_lookup_3() 1804 &rhs_offset); in vn_reference_lookup_3() 1811 rhs_offset += TREE_INT_CST_LOW (TREE_OPERAND (tem, 1)); in vn_reference_lookup_3() 1859 op.op0 = build_int_cst (ptr_type_node, at - rhs_offset); in vn_reference_lookup_3() 1860 op.off = at - lhs_offset + rhs_offset; in vn_reference_lookup_3()
|
/dports/lang/gcc6-aux/gcc-6-20180516/gcc/ |
H A D | tree-ssa-sccvn.c | 2058 HOST_WIDE_INT rhs_offset, copy_size, lhs_offset; in vn_reference_lookup_3() local 2107 rhs_offset = 0; in vn_reference_lookup_3() 2113 &rhs_offset); in vn_reference_lookup_3() 2120 rhs_offset += tree_to_uhwi (TREE_OPERAND (tem, 1)); in vn_reference_lookup_3() 2172 op.op0 = build_int_cst (ptr_type_node, at - lhs_offset + rhs_offset); in vn_reference_lookup_3() 2173 op.off = at - lhs_offset + rhs_offset; in vn_reference_lookup_3()
|
/dports/lang/gnat_util/gcc-6-20180516/gcc/ |
H A D | tree-ssa-sccvn.c | 2058 HOST_WIDE_INT rhs_offset, copy_size, lhs_offset; in vn_reference_lookup_3() local 2107 rhs_offset = 0; in vn_reference_lookup_3() 2113 &rhs_offset); in vn_reference_lookup_3() 2120 rhs_offset += tree_to_uhwi (TREE_OPERAND (tem, 1)); in vn_reference_lookup_3() 2172 op.op0 = build_int_cst (ptr_type_node, at - lhs_offset + rhs_offset); in vn_reference_lookup_3() 2173 op.off = at - lhs_offset + rhs_offset; in vn_reference_lookup_3()
|