/dports/math/onednn/oneDNN-2.5.1/tests/benchdnn/rnn/ |
H A D | cfg.cpp | 73 CASE(SRC_LAYER, BF16_ENTRY_BF16); in CFG() 84 CASE(SRC_LAYER, BF16_ENTRY_BF16); in CFG() 143 CASE(SRC_LAYER, U8_ENTRY_U8); in CFG() 159 CASE(SRC_LAYER, U8_ENTRY_U8); in CFG() 175 CASE(SRC_LAYER, U8_ENTRY_U8); in CFG() 191 CASE(SRC_LAYER, U8_ENTRY_U8); in CFG() 207 CASE(SRC_LAYER, S8_ENTRY_S8); in CFG() 223 CASE(SRC_LAYER, S8_ENTRY_S8); in CFG() 239 CASE(SRC_LAYER, S8_ENTRY_S8); in CFG() 255 CASE(SRC_LAYER, S8_ENTRY_S8); in CFG()
|
H A D | rnn.hpp | 52 SRC_LAYER, enumerator 195 return operator[](SRC_LAYER).dt == dnnl_u8 in is_int8() 196 || operator[](SRC_LAYER).dt == dnnl_s8; in is_int8() 198 bool is_s8() const { return operator[](SRC_LAYER).dt == dnnl_s8; } in is_s8() 287 set_tparams(cfg[SRC_LAYER].f_min, cfg[SRC_LAYER].f_max); in prb_t() 369 return cfg[SRC_LAYER].dt == dnnl_u8 || cfg[SRC_LAYER].dt == dnnl_s8; in is_int8() 371 bool is_u8() const { return cfg[SRC_LAYER].dt == dnnl_u8; } in is_u8() 372 bool is_s8() const { return cfg[SRC_LAYER].dt == dnnl_s8; } in is_s8()
|
H A D | rnn_aux.cpp | 111 CASE(SRC_LAYER); in data_kind2str() 402 if (fp > prb.cfg[SRC_LAYER].max) fp = prb.cfg[SRC_LAYER].max; in maybe_q() 403 if (fp < prb.cfg[SRC_LAYER].min) fp = prb.cfg[SRC_LAYER].min; in maybe_q() 439 if (kind == SRC_LAYER || kind == DIFF_SRC_LAYER) C = prb.slc; in inv_tnc_off_f() 509 case SRC_LAYER: in print_value() 738 float int8_src_range = cfg[SRC_LAYER].f_max - cfg[SRC_LAYER].f_min, in set_qparams() 742 data_shift = cfg.is_s8() ? 0 : cfg[SRC_LAYER].f_mean; in set_qparams()
|
H A D | rnn.cpp | 219 case SRC_LAYER: in check_s8s8_reorder() 293 assert(kind == SRC_LAYER || kind == SRC_ITER); in fill_memory() 294 auto ld = kind == SRC_LAYER ? prb.slc : prb.sic; in fill_memory() 313 case SRC_LAYER: in fill_memory() 388 && (kind == SRC_LAYER || kind == SRC_ITER); in fill_activation() 445 float expect_gemm_output = (1.f / prb.n_gates()) * prb.cfg[SRC_LAYER].f_mean in fill_src_iter_c() 594 prb.cfg[SRC_LAYER].dt, dnnl_tnc), in init_pd() 763 check_known_skipped_case_common({prb.cfg[SRC_LAYER].dt}, dir, res); in check_known_skipped_case() 824 if (prb.is_lstm_projection() && prb.cfg[SRC_LAYER].dt == dnnl_bf16) { in check_known_skipped_case() 943 SAFE(fill_activation(prb, SRC_LAYER, src_layer_dt, src_layer_fp, rnn_attr), in doit()
|
/dports/misc/mxnet/incubator-mxnet-1.9.0/3rdparty/mkldnn/tests/benchdnn/rnn/ |
H A D | cfg.cpp | 69 CASE(SRC_LAYER, BF16_ENTRY_BF16); in CFG() 121 CASE(SRC_LAYER, U8_ENTRY_U8); in CFG() 136 CASE(SRC_LAYER, U8_ENTRY_U8); in CFG() 151 CASE(SRC_LAYER, U8_ENTRY_U8); in CFG() 166 CASE(SRC_LAYER, U8_ENTRY_U8); in CFG() 181 CASE(SRC_LAYER, S8_ENTRY_S8); in CFG() 196 CASE(SRC_LAYER, S8_ENTRY_S8); in CFG() 211 CASE(SRC_LAYER, S8_ENTRY_S8); in CFG() 226 CASE(SRC_LAYER, S8_ENTRY_S8); in CFG()
|
H A D | rnn.hpp | 52 SRC_LAYER, enumerator 195 return operator[](SRC_LAYER).dt == dnnl_u8 in is_int8() 196 || operator[](SRC_LAYER).dt == dnnl_s8; in is_int8() 198 bool is_s8() const { return operator[](SRC_LAYER).dt == dnnl_s8; } in is_s8() 287 set_tparams(cfg[SRC_LAYER].f_min, cfg[SRC_LAYER].f_max); in prb_t() 369 return cfg[SRC_LAYER].dt == dnnl_u8 || cfg[SRC_LAYER].dt == dnnl_s8; in is_int8() 371 bool is_u8() const { return cfg[SRC_LAYER].dt == dnnl_u8; } in is_u8() 372 bool is_s8() const { return cfg[SRC_LAYER].dt == dnnl_s8; } in is_s8()
|
H A D | rnn_aux.cpp | 111 CASE(SRC_LAYER); in data_kind2str() 402 if (fp > prb.cfg[SRC_LAYER].max) fp = prb.cfg[SRC_LAYER].max; in maybe_q() 403 if (fp < prb.cfg[SRC_LAYER].min) fp = prb.cfg[SRC_LAYER].min; in maybe_q() 439 if (kind == SRC_LAYER || kind == DIFF_SRC_LAYER) C = prb.slc; in inv_tnc_off_f() 509 case SRC_LAYER: in print_value() 738 float int8_src_range = cfg[SRC_LAYER].f_max - cfg[SRC_LAYER].f_min, in set_qparams() 742 data_shift = cfg.is_s8() ? 0 : cfg[SRC_LAYER].f_mean; in set_qparams()
|
H A D | rnn.cpp | 219 case SRC_LAYER: in check_s8s8_reorder() 293 assert(kind == SRC_LAYER || kind == SRC_ITER); in fill_memory() 294 auto ld = kind == SRC_LAYER ? prb.slc : prb.sic; in fill_memory() 313 case SRC_LAYER: in fill_memory() 388 && (kind == SRC_LAYER || kind == SRC_ITER); in fill_activation() 445 float expect_gemm_output = (1.f / prb.n_gates()) * prb.cfg[SRC_LAYER].f_mean in fill_src_iter_c() 594 prb.cfg[SRC_LAYER].dt, dnnl_tnc), in init_pd() 763 check_known_skipped_case_common({prb.cfg[SRC_LAYER].dt}, dir, res); in check_known_skipped_case() 824 if (prb.is_lstm_projection() && prb.cfg[SRC_LAYER].dt == dnnl_bf16) { in check_known_skipped_case() 936 SAFE(fill_activation(prb, SRC_LAYER, src_layer_dt, src_layer_fp, rnn_attr), in doit()
|
/dports/graphics/mtpaint/mtPaint-4ea607e/src/ |
H A D | layer.c | 178 if ((blend_src == SRC_LAYER + layer_selected) || (blend_src == SRC_LAYER + lv)) in shift_layer() 179 blend_src ^= (SRC_LAYER + layer_selected) ^ (SRC_LAYER + lv); in shift_layer() 314 if (blend_src == SRC_LAYER + layer_selected) blend_src = SRC_NORMAL; in layer_press_delete() 395 if (blend_src > SRC_LAYER + 0) blend_src = SRC_NORMAL; in layers_free_all()
|
H A D | toolbar.c | 327 char *names[SRC_LAYER + MAX_LAYERS + 2]; in blend_settings() 346 names[SRC_LAYER + i] = ns[i]; in blend_settings() 348 names[SRC_LAYER + i] = NULL; in blend_settings()
|
H A D | memory.h | 316 SRC_LAYER enumerator
|
H A D | memory.c | 6960 (blend_src == SRC_LAYER + layer_selected)) in put_pixel_def() 6968 layer_node *t = layer_table + blend_src - SRC_LAYER; in put_pixel_def() 7148 (blend_src == SRC_LAYER + layer_selected)) mode = PP_OFS; in put_pixel_row_def() 7151 layer_node *t = layer_table + blend_src - SRC_LAYER; in put_pixel_row_def()
|