/dports/security/gcr/gcr-3.40.0/egg/ |
H A D | egg-padding.h | 31 gsize n_block, 38 gsize n_block, 45 gsize n_block, 52 gsize n_block, 59 gsize n_block, 66 gsize n_block, 73 gsize n_block, 80 gsize n_block,
|
/dports/security/gnome-keyring/gnome-keyring-40.0/egg/ |
H A D | egg-padding.h | 32 gsize n_block, 39 gsize n_block, 46 gsize n_block, 53 gsize n_block, 60 gsize n_block, 67 gsize n_block, 74 gsize n_block, 81 gsize n_block,
|
/dports/misc/mxnet/incubator-mxnet-1.9.0/3rdparty/mkldnn/src/cpu/x64/rnn/ |
H A D | rnn_brgemm_utils.cpp | 278 rnn.n_block = 32; in configure_brgemm() 280 rnn.n_tail = rnn.N % rnn.n_block; in configure_brgemm() 320 rnn.LDB1 = rnn.n_block; in configure_brgemm() 321 rnn.LDB2 = rnn.n_block; in configure_brgemm() 327 dim_t n_block = nstl::min(rnn.N, rnn.n_block); in configure_brgemm() local 372 rnn.LDBproj = rnn.n_block; in configure_brgemm() 382 dim_t n_block = nstl::min(rnn.Nproj, rnn.n_block); in configure_brgemm() local 600 diff_src_conf.n_block = 32; in configure_brgemm() 649 const dim_t n_block = nstl::min(diff_src_conf.N, diff_src_conf.n_block); in configure_brgemm() local 668 diff_wei_conf.n_block = 32; in configure_brgemm() [all …]
|
H A D | jit_gates_reduction.cpp | 32 : rnn_.diff_wei_brgemm.n_block) in jit_gates_reduction_t() 108 const int n_block_off = rnn_.diff_wei_brgemm.n_block * sizeof(float); in compute() 114 for (int n_block = first_reversed_block; n_block >= 0; --n_block) { in compute() local 116 && n_block == first_reversed_block; in compute() 117 const auto &acc_zmm = acc_regs_[n_block]; in compute() 118 const int nk_offset = k_offset + n_block * simd_w_ * sizeof(float); in compute() 129 const int n_block_off = rnn_.diff_wei_brgemm.n_block in compute_loop()
|
H A D | brgemm_cell_common_fwd.cpp | 57 , Bl_n_offset_(rnn_.K1padded * rnn_.n_block) in brgemm_dst_layer_iter_t() 58 , Bi_n_offset_(rnn_.K2padded * rnn_.n_block) in brgemm_dst_layer_iter_t() 63 , Bl_kb_offset_(rnn_.k1_block * rnn_.n_block) in brgemm_dst_layer_iter_t() 64 , Bi_kb_offset_(rnn_.k2_block * rnn_.n_block) in brgemm_dst_layer_iter_t() 126 ? amx_scratchpad_ + rnn_.m_block * rnn_.n_block * ithr in kernel() 146 const auto n = nb * rnn_.n_block; in kernel() 171 const bool do_n_tail = (n + rnn_.n_block) > rnn_.N; in kernel() 271 , B_n_offset_(rnn_.Kprojpadded * rnn_.n_block) in brgemm_dst_proj_t() 272 , Bp_kb_offset_(rnn_.kproj_block * rnn_.n_block) in brgemm_dst_proj_t() 312 const int n = nb * rnn_.n_block; in kernel() [all …]
|
H A D | brgemm_cell_common_bwd.cpp | 53 , B_k_tail_offset_(A_k_tail_offset_ * rnn.diff_src_brgemm.n_block) in brgemm_diff_src_layer_iter_t() 55 , B_kb_offset_(k_block_ * rnn.diff_src_brgemm.n_block) in brgemm_diff_src_layer_iter_t() 59 * rnn.diff_src_brgemm.n_block in brgemm_diff_src_layer_iter_t() 126 const int n = n_block_id * rnn_.diff_src_brgemm.n_block; in kernel_amx() 267 const int n = n_block_id * rnn_.diff_src_brgemm.n_block; in kernel() 411 , B_kb_offset_(k_block_ * rnn.diff_wei_brgemm.n_block) in brgemm_diff_weights_layer_iter_t() 414 k_blocks_ * k_block_ * rnn.diff_wei_brgemm.n_block) in brgemm_diff_weights_layer_iter_t() 463 * rnn_.diff_wei_brgemm.n_block; in kernel() 519 const int n = n_block_id * rnn_.diff_wei_brgemm.n_block; in kernel() 619 * rnn_.diff_wei_brgemm.n_block; in kernel_amx() [all …]
|
/dports/graphics/geomview/geomview-1.9.5/src/lib/geometry/cmodel/ |
H A D | cmodel_data.c | 87 struct vertex_block *n_block; in new_vertex_block() local 89 curr_vertex_block->next = n_block = in new_vertex_block() 91 n_block->next = NULL; in new_vertex_block() 92 return curr_vertex_block = n_block; in new_vertex_block() 197 struct edge_block *n_block; in new_edge_block() local 198 curr_edge_block->next = n_block = in new_edge_block() 200 n_block->next = NULL; in new_edge_block() 201 return curr_edge_block = n_block; in new_edge_block() 284 struct triangle_block *n_block; in new_triangle_block() local 288 n_block->next = NULL; in new_triangle_block() [all …]
|
/dports/math/onednn/oneDNN-2.5.1/src/cpu/x64/rnn/ |
H A D | rnn_brgemm_utils.cpp | 302 rnn.n_block = 32; in configure_brgemm() 304 rnn.n_tail = rnn.N % rnn.n_block; in configure_brgemm() 353 rnn.LDB1 = rnn.n_block; in configure_brgemm() 354 rnn.LDB2 = rnn.n_block; in configure_brgemm() 361 dim_t n_block = nstl::min(rnn.N, rnn.n_block); in configure_brgemm() local 406 rnn.LDBproj = rnn.n_block; in configure_brgemm() 416 dim_t n_block = nstl::min(rnn.Nproj, rnn.n_block); in configure_brgemm() local 665 diff_src_conf.n_block = 32; in configure_brgemm() 719 const dim_t n_block = nstl::min(diff_src_conf.N, diff_src_conf.n_block); in configure_brgemm() local 738 diff_wei_conf.n_block = 32; in configure_brgemm() [all …]
|
H A D | jit_gates_reduction.cpp | 32 : rnn_.diff_wei_brgemm.n_block) in jit_gates_reduction_t() 108 const int n_block_off = rnn_.diff_wei_brgemm.n_block * sizeof(float); in compute() 114 for (int n_block = first_reversed_block; n_block >= 0; --n_block) { in compute() local 116 && n_block == first_reversed_block; in compute() 117 const auto &acc_zmm = acc_regs_[n_block]; in compute() 118 const int nk_offset = k_offset + n_block * simd_w_ * sizeof(float); in compute() 129 const int n_block_off = rnn_.diff_wei_brgemm.n_block in compute_loop()
|
H A D | brgemm_cell_common_fwd.cpp | 57 , Bl_n_offset_(rnn_.K1padded * rnn_.n_block) in brgemm_dst_layer_iter_t() 58 , Bi_n_offset_(rnn_.K2padded * rnn_.n_block) in brgemm_dst_layer_iter_t() 63 , Bl_kb_offset_(rnn_.k1_block * rnn_.n_block) in brgemm_dst_layer_iter_t() 64 , Bi_kb_offset_(rnn_.k2_block * rnn_.n_block) in brgemm_dst_layer_iter_t() 153 const auto n = nb * rnn_.n_block; in kernel() 178 const bool do_n_tail = (n + rnn_.n_block) > rnn_.N; in kernel() 298 const auto n = nb * rnn_.n_block; in kernel_fused_iter_layer() 317 const bool do_n_tail = (n + rnn_.n_block) > rnn_.N; in kernel_fused_iter_layer() 408 , B_n_offset_(rnn_.Kprojpadded * rnn_.n_block) in brgemm_dst_proj_t() 409 , Bp_kb_offset_(rnn_.kproj_block * rnn_.n_block) in brgemm_dst_proj_t() [all …]
|
H A D | brgemm_cell_common_bwd.cpp | 53 , B_k_tail_offset_(A_k_tail_offset_ * rnn.diff_src_brgemm.n_block) in brgemm_diff_src_layer_iter_t() 55 , B_kb_offset_(k_block_ * rnn.diff_src_brgemm.n_block) in brgemm_diff_src_layer_iter_t() 59 * rnn.diff_src_brgemm.n_block in brgemm_diff_src_layer_iter_t() 126 const int n = n_block_id * rnn_.diff_src_brgemm.n_block; in kernel_amx() 267 const int n = n_block_id * rnn_.diff_src_brgemm.n_block; in kernel() 411 , B_kb_offset_(k_block_ * rnn.diff_wei_brgemm.n_block) in brgemm_diff_weights_layer_iter_t() 414 k_blocks_ * k_block_ * rnn.diff_wei_brgemm.n_block) in brgemm_diff_weights_layer_iter_t() 463 * rnn_.diff_wei_brgemm.n_block; in kernel() 513 const int n = n_block_id * rnn_.diff_wei_brgemm.n_block; in kernel() 619 * rnn_.diff_wei_brgemm.n_block; in kernel_amx() [all …]
|
/dports/devel/concurrencykit/ck-0.6.0/include/ |
H A D | ck_bitmap.h | 125 unsigned int n_block; member 471 i->n_block = 0; in ck_bitmap_iterator_init() 490 unsigned int n_block = i->n_block; in ck_bitmap_next() local 494 if (n_block >= n_limit) in ck_bitmap_next() 497 for (n_block++; n_block < n_limit; n_block++) { in ck_bitmap_next() 498 cache = ck_pr_load_uint(&bitmap->map[n_block]); in ck_bitmap_next() 504 i->n_block = n_block; in ck_bitmap_next() 509 *bit = CK_BITMAP_BLOCK * n_block + ck_cc_ctz(cache); in ck_bitmap_next() 511 i->n_block = n_block; in ck_bitmap_next()
|
/dports/benchmarks/sysbench/sysbench-1.0.20/third_party/concurrency_kit/ck/include/ |
H A D | ck_bitmap.h | 125 unsigned int n_block; member 471 i->n_block = 0; in ck_bitmap_iterator_init() 490 unsigned int n_block = i->n_block; in ck_bitmap_next() local 494 if (n_block >= n_limit) in ck_bitmap_next() 497 for (n_block++; n_block < n_limit; n_block++) { in ck_bitmap_next() 498 cache = ck_pr_load_uint(&bitmap->map[n_block]); in ck_bitmap_next() 504 i->n_block = n_block; in ck_bitmap_next() 509 *bit = CK_BITMAP_BLOCK * n_block + ck_cc_ctz(cache); in ck_bitmap_next() 511 i->n_block = n_block; in ck_bitmap_next()
|
/dports/devel/ga/ga-5.8/global/testing/unit-tests/ |
H A D | ga_create_irreg.c | 25 int n_block[DIM], block_dims[DIM], i; in irregular_array1() local 55 int n_block[DIM], block_dims[DIM], i; in irregular_array2() local 81 int n_block[DIM], block_dims[DIM], b_temp, i; in auto_number1() local 117 GA_Get_block_info(g_B, n_block, block_dims); in auto_number1() 119 printf(" %d: %d --- %d ... %d\n", rank, n_block[i], block_dims[i], b_temp); in auto_number1() 129 int n_block[DIM], block_dims[DIM], b_temp, i; in auto_number2() local 166 GA_Get_block_info(g_B, n_block, block_dims); in auto_number2() 168 printf(" %d: %d --- %d ... %d\n", rank, n_block[i], block_dims[i], b_temp); in auto_number2()
|
H A D | ga_create_irreg2.c | 24 int n_block[DIM], block_dims[DIM], i; in irregular_array1() local 54 int n_block[DIM], block_dims[DIM], i; in irregular_array2() local 84 int n_block[DIM], block_dims[DIM], b_temp, i; in auto_number1() local 120 GA_Get_block_info(g_B, n_block, block_dims); in auto_number1() 122 printf(" %d: %d --- %d ... %d\n", rank, n_block[i], block_dims[i], b_temp); in auto_number1() 132 int n_block[DIM], block_dims[DIM], b_temp, i; in auto_number2() local 169 GA_Get_block_info(g_B, n_block, block_dims); in auto_number2() 171 printf(" %d: %d --- %d ... %d\n", rank, n_block[i], block_dims[i], b_temp); in auto_number2()
|
H A D | ga_get_blockinfo.c | 23 int n_block[DIM], block_dims[DIM]; in main() local 44 GA_Get_block_info(g_B, n_block, block_dims); in main() 46 printf(" %d: %d ___ %d --- \n", rank, n_block[i], block_dims[i]); in main()
|
/dports/science/py-nilearn/nilearn-0.8.1/nilearn/tests/ |
H A D | test_testing.py | 47 for shape, length, kind, n_block, bsize, btype in itertools.product( 50 if n_block is None: 53 n_blocks=n_block, block_size=bsize, 59 n_blocks=n_block, block_size=bsize, 66 if n_block is not None:
|
/dports/misc/mxnet/incubator-mxnet-1.9.0/3rdparty/mkldnn/src/cpu/x64/ |
H A D | jit_brgemm_post_ops.hpp | 401 Xbyak::Zmm vector(int m, int n, int n_block) { in vector() 402 return Xbyak::Zmm(m * n_block + n); in vector() 416 for (int n = 0; n < n_block; n++) { in inject_attr_postops() 439 for (int n = 0; n < n_block; n++) { in inject_attr_postops() 451 0, m_block * n_block, rhs_arg_params); in inject_attr_postops() 465 for (int n = 0; n < n_block; n++) { in apply_post_ops() 480 for (int n = 0; n < n_block; n++) { in apply_post_ops() 492 for (int n = 0; n < n_block; n++) { in apply_post_ops() 516 for (int n = 0; n < n_block; n++) { in apply_post_ops() 612 int m_max_regs = 28 / n_block; in generate() [all …]
|
/dports/math/onednn/oneDNN-2.5.1/src/cpu/x64/ |
H A D | jit_brgemm_post_ops.hpp | 403 Xbyak::Zmm vector(int m, int n, int n_block) { in vector() 404 return Xbyak::Zmm(m * n_block + n); in vector() 425 for (int n = 0; n < n_block; n++) { in inject_attr_postops() 449 for (int n = 0; n < n_block; n++) { in inject_attr_postops() 462 0, m_block * n_block, rhs_arg_params); in inject_attr_postops() 476 for (int n = 0; n < n_block; n++) { in apply_post_ops() 491 for (int n = 0; n < n_block; n++) { in apply_post_ops() 503 for (int n = 0; n < n_block; n++) { in apply_post_ops() 527 for (int n = 0; n < n_block; n++) { in apply_post_ops() 623 int m_max_regs = 28 / n_block; in generate() [all …]
|
/dports/science/octopus/octopus-10.5/src/hamiltonian/ |
H A D | vxc_inc.F90 | 36 integer :: n_block 247 do ib = 1, n_block 251 do ib = 1, n_block 263 SAFE_ALLOCATE(unp_dens(1:n_block)) 266 do ib = 1, n_block 280 do ib = 1, n_block 292 do ib = 1, n_block 430 integer, intent(in) :: n_block 440 do ib = 1, n_block 452 integer, intent(in) :: n_block [all …]
|
/dports/math/faiss/faiss-1.7.1/faiss/invlists/ |
H A D | BlockInvertedLists.cpp | 48 size_t n_block = (n_entry + n_per_block - 1) / n_per_block; in add_entries() local 49 codes[list_no].resize(n_block * block_size); in add_entries() 50 memcpy(&codes[list_no][o * code_size], code, n_block * block_size); in add_entries() 72 size_t n_block = (new_size + n_per_block - 1) / n_per_block; in resize() local 73 size_t new_nbytes = n_block * block_size; in resize()
|
/dports/security/gnome-keyring/gnome-keyring-40.0/pkcs11/gkm/ |
H A D | gkm-crypto.c | 88 gsize n_block, offset, len; in gkm_crypto_sexp_to_data() local 115 n_block = (bits + 7) / 8; in gkm_crypto_sexp_to_data() 118 g_return_val_if_fail (len <= n_block, CKR_GENERAL_ERROR); in gkm_crypto_sexp_to_data() 119 offset = n_block - len; in gkm_crypto_sexp_to_data() 120 block = g_malloc0 (n_block); in gkm_crypto_sexp_to_data() 124 g_return_val_if_fail (len == n_block - offset, CKR_GENERAL_ERROR); in gkm_crypto_sexp_to_data() 129 ret = (padding) (g_realloc, n_block, block, n_block, &padded, &n_block); in gkm_crypto_sexp_to_data() 137 if (n_block > *n_data) in gkm_crypto_sexp_to_data() 140 memcpy (data, block, n_block); in gkm_crypto_sexp_to_data() 141 *n_data = n_block; in gkm_crypto_sexp_to_data()
|
/dports/devel/hpx/hpx-1.2.1/examples/jacobi_smp/ |
H A D | jacobi_hpx.cpp | 50 std::size_t n_block = static_cast<std::size_t>(std::ceil(double(n)/block_size)); in jacobi() local 54 new deps_vector(n_block, hpx::make_ready_future())); in jacobi() 56 new deps_vector(n_block, hpx::make_ready_future())); in jacobi() 68 if(j + 1 < n_block) trigger.push_back((*deps_old)[j+1]); in jacobi()
|
/dports/math/adept/adept-2.0.8/adept/ |
H A D | jacobian.cpp | 134 int n_block = (n_independent() + MULTIPASS_SIZE - 1) in jacobian_forward_openmp() local 147 for (int iblock = 0; iblock < n_block; iblock++) { in jacobian_forward_openmp() 158 if (iblock == n_block-1 && n_extra > 0) { in jacobian_forward_openmp() 236 uIndex n_block = n_independent() / MULTIPASS_SIZE; in jacobian_forward() local 245 for (uIndex iblock = 0; iblock < n_block; iblock++) { in jacobian_forward() 285 uIndex i_independent = MULTIPASS_SIZE * n_block; in jacobian_forward() 339 int n_block = (n_dependent() + MULTIPASS_SIZE - 1) in jacobian_reverse_openmp() local 358 for (int iblock = 0; iblock < n_block; iblock++) { in jacobian_reverse_openmp() 369 if (iblock == n_block-1 && n_extra > 0) { in jacobian_reverse_openmp() 500 uIndex n_block = n_dependent() / MULTIPASS_SIZE; in jacobian_reverse() local [all …]
|
/dports/math/openturns/openturns-1.18/python/doc/examples/reliability_sensitivity/reliability/ |
H A D | plot_proba_system_event.py | 107 def computeCrossingProbability_MonteCarlo(b, t, mu_S, covariance, R, delta_t, n_block, n_iter, CoV): argument 116 algo.setBlockSize(n_block) 127 def computeCrossingProbability_QMC(b, t, mu_S, covariance, R, delta_t, n_block, n_iter, CoV): argument 136 SobolSequence(X.getDimension()), n_block, False)) 137 algo.setBlockSize(n_block)
|