/dports/misc/py-mxnet/incubator-mxnet-1.9.0/src/operator/ |
H A D | svm_output.cu | 44 for (index_t n_index = n; n_index < nmax; n_index += n_size) { in L1_SVMKernel() local 45 if (n_index == k) { in L1_SVMKernel() 48 dst[y][n_index] = DType(margin > -src[y][n_index]) * reg_coef; in L1_SVMKernel() 79 for (index_t n_index = n; n_index < nmax; n_index += n_size) { in L2_SVMKernel() local 80 if (n_index == k) { in L2_SVMKernel() 84 dst[y][n_index] = margin > -src[y][n_index] ? (-2)*(margin + src[y][n_index]) : DType(0.0f); in L2_SVMKernel() 85 dst[y][n_index] *= -reg_coef; in L2_SVMKernel()
|
/dports/misc/mxnet/incubator-mxnet-1.9.0/src/operator/ |
H A D | svm_output.cu | 44 for (index_t n_index = n; n_index < nmax; n_index += n_size) { in L1_SVMKernel() local 45 if (n_index == k) { in L1_SVMKernel() 48 dst[y][n_index] = DType(margin > -src[y][n_index]) * reg_coef; in L1_SVMKernel() 79 for (index_t n_index = n; n_index < nmax; n_index += n_size) { in L2_SVMKernel() local 80 if (n_index == k) { in L2_SVMKernel() 84 dst[y][n_index] = margin > -src[y][n_index] ? (-2)*(margin + src[y][n_index]) : DType(0.0f); in L2_SVMKernel() 85 dst[y][n_index] *= -reg_coef; in L2_SVMKernel()
|
/dports/emulators/mess/mame-mame0226/src/devices/cpu/psx/ |
H A D | mdec.cpp | 425 int n_index; in dma_write() local 439 n_index = 0; in dma_write() 442 if( n_index < DCTSIZE2 ) in dma_write() 444 p_n_quantize_y[ n_index + 0 ] = ( p_n_psxram[ n_address / 4 ] >> 0 ) & 0xff; in dma_write() 445 p_n_quantize_y[ n_index + 1 ] = ( p_n_psxram[ n_address / 4 ] >> 8 ) & 0xff; in dma_write() 446 p_n_quantize_y[ n_index + 2 ] = ( p_n_psxram[ n_address / 4 ] >> 16 ) & 0xff; in dma_write() 447 p_n_quantize_y[ n_index + 3 ] = ( p_n_psxram[ n_address / 4 ] >> 24 ) & 0xff; in dma_write() 449 else if( n_index < DCTSIZE2 * 2 ) in dma_write() 456 n_index += 4; in dma_write() 463 n_index = 0; in dma_write() [all …]
|
/dports/emulators/mame/mame-mame0226/src/devices/cpu/psx/ |
H A D | mdec.cpp | 425 int n_index; in dma_write() local 439 n_index = 0; in dma_write() 442 if( n_index < DCTSIZE2 ) in dma_write() 444 p_n_quantize_y[ n_index + 0 ] = ( p_n_psxram[ n_address / 4 ] >> 0 ) & 0xff; in dma_write() 445 p_n_quantize_y[ n_index + 1 ] = ( p_n_psxram[ n_address / 4 ] >> 8 ) & 0xff; in dma_write() 446 p_n_quantize_y[ n_index + 2 ] = ( p_n_psxram[ n_address / 4 ] >> 16 ) & 0xff; in dma_write() 447 p_n_quantize_y[ n_index + 3 ] = ( p_n_psxram[ n_address / 4 ] >> 24 ) & 0xff; in dma_write() 449 else if( n_index < DCTSIZE2 * 2 ) in dma_write() 456 n_index += 4; in dma_write() 463 n_index = 0; in dma_write() [all …]
|
/dports/misc/py-xgboost/xgboost-1.5.1/src/data/ |
H A D | gradient_index.cc | 71 const size_t n_index = row_ptr[rbegin + batch.Size()]; in PushBatch() local 72 ResizeIndex(n_index, isDense_); in PushBatch() 88 common::Span<uint8_t> index_data_span = {index.data<uint8_t>(), n_index}; in PushBatch() 96 n_index}; in PushBatch() 104 n_index}; in PushBatch() 115 common::Span<uint32_t> index_data_span = {index.data<uint32_t>(), n_index}; in PushBatch() 182 void GHistIndexMatrix::ResizeIndex(const size_t n_index, in ResizeIndex() argument 186 index.Resize((sizeof(uint8_t)) * n_index); in ResizeIndex() 190 index.Resize((sizeof(uint16_t)) * n_index); in ResizeIndex() 193 index.Resize((sizeof(uint32_t)) * n_index); in ResizeIndex()
|
/dports/misc/xgboost/xgboost-1.5.1/src/data/ |
H A D | gradient_index.cc | 71 const size_t n_index = row_ptr[rbegin + batch.Size()]; in PushBatch() local 72 ResizeIndex(n_index, isDense_); in PushBatch() 88 common::Span<uint8_t> index_data_span = {index.data<uint8_t>(), n_index}; in PushBatch() 96 n_index}; in PushBatch() 104 n_index}; in PushBatch() 115 common::Span<uint32_t> index_data_span = {index.data<uint32_t>(), n_index}; in PushBatch() 182 void GHistIndexMatrix::ResizeIndex(const size_t n_index, in ResizeIndex() argument 186 index.Resize((sizeof(uint8_t)) * n_index); in ResizeIndex() 190 index.Resize((sizeof(uint16_t)) * n_index); in ResizeIndex() 193 index.Resize((sizeof(uint32_t)) * n_index); in ResizeIndex()
|
/dports/science/elmerfem/elmerfem-release-9.0/elmerice/Solvers/MeshAdaptation_2D/ |
H A D | Compute2DNodalGradient.F90 | 83 TYPE(lbuff_t), ALLOCATABLE :: n_index(:) local 204 ALLOCATE( n_count(ParEnv% PEs),n_index(ParEnv % PEs) ) 220 ALLOCATE( n_index(i) % buff(n_count(i)), & 221 n_index(i) % values((STDOFs+1)*n_count(i)) ) 234 n_index(k) % buff(n_count(k)) = Solver%Mesh % Parallelinfo % & 236 n_index(k) % values((STDOFs+1)*(n_count(k)-1)+1)=grad(STDOFs*(i-1)+1) 238 n_index(k) % Values((STDOFs+1)*(n_count(k)-1)+2)=grad(STDOFs*(i-1)+2) 239 n_index(k) % values((STDOFs+1)*(n_count(k)-1)+STDOFs+1)=weight(i) 248 CALL MPI_BSEND( n_index(i) % buff, n_count(i), MPI_INTEGER, i-1, & 256 IF ( n_count(i)>0 ) DEALLOCATE( n_index(i) % Buff, n_index(i) % values) [all …]
|
/dports/math/z3/z3-z3-4.8.13/src/muz/spacer/ |
H A D | spacer_manager.h | 81 unsigned n_index() const { return 0; } in n_index() function 96 {return m_mux.is_homogenous_formula(f, n_index());} in is_n_formula() 99 {return m_mux.shift_decl(p, o_index(o_idx), n_index());} in o2n() 103 {return m_mux.shift_decl(p, n_index(), o_index(o_idx));} in n2o() 107 {m_mux.shift_expr(f, o_index(o_idx), n_index(), result, homogenous);} 111 {m_mux.shift_expr(f, n_index(), o_index(o_idx), result, homogenous);} 114 {m_mux.shift_expr(result.get(), n_index(), o_index(o_idx), in formula_n2o()
|
/dports/math/py-brial/BRiAl-1.2.10/groebner/include/polybori/groebner/ |
H A D | fixed_path_divisors.h | 47 int n_index=*n_nav; in do_fixed_path_divisors() local 50 PBORI_ASSERT(m_index<=n_index); in do_fixed_path_divisors() 57 n_index=*n_nav; in do_fixed_path_divisors() 58 PBORI_ASSERT(n_index>=m_index); in do_fixed_path_divisors() 59 if (m_index==n_index){ in do_fixed_path_divisors() 68 n_index=*n_nav; in do_fixed_path_divisors() 106 if (m_index==n_index){ in do_fixed_path_divisors()
|
/dports/math/brial/BRiAl-1.2.10/groebner/include/polybori/groebner/ |
H A D | fixed_path_divisors.h | 47 int n_index=*n_nav; in do_fixed_path_divisors() local 50 PBORI_ASSERT(m_index<=n_index); in do_fixed_path_divisors() 57 n_index=*n_nav; in do_fixed_path_divisors() 58 PBORI_ASSERT(n_index>=m_index); in do_fixed_path_divisors() 59 if (m_index==n_index){ in do_fixed_path_divisors() 68 n_index=*n_nav; in do_fixed_path_divisors() 106 if (m_index==n_index){ in do_fixed_path_divisors()
|
/dports/math/py-z3-solver/z3-z3-4.8.10/src/muz/spacer/ |
H A D | spacer_manager.h | 81 unsigned n_index() const { return 0; } in n_index() function 96 {return m_mux.is_homogenous_formula(f, n_index());} in is_n_formula() 99 {return m_mux.shift_decl(p, o_index(o_idx), n_index());} in o2n() 103 {return m_mux.shift_decl(p, n_index(), o_index(o_idx));} in n2o() 107 {m_mux.shift_expr(f, o_index(o_idx), n_index(), result, homogenous);} 111 {m_mux.shift_expr(f, n_index(), o_index(o_idx), result, homogenous);} 114 {m_mux.shift_expr(result.get(), n_index(), o_index(o_idx), in formula_n2o()
|
/dports/biology/gemma/GEMMA-0.98.3/src/ |
H A D | lmm.cpp | 284 assert(Uab->size2 == n_index); in CalcPab() 287 assert(Pab->size2 == n_index); in CalcPab() 605 assert(Pab->size2 == n_index); in LogL_dev1() 607 assert(ab->size == n_index); in LogL_dev1() 1383 gsl_vector *ab = gsl_vector_safe_alloc(n_index); in AnalyzeGene() 1485 gsl_vector *ab = gsl_vector_safe_alloc(n_index); in Analyze() 1725 gsl_vector *ab = gsl_vector_alloc(n_index); in AnalyzePlink() 2109 gsl_vector *ab = gsl_vector_safe_alloc(n_index); in CalcLambda() 2136 gsl_vector *ab = gsl_vector_safe_alloc(n_index); in CalcPve() 2163 gsl_vector *ab = gsl_vector_safe_alloc(n_index); in CalcLmmVgVeBeta() [all …]
|
/dports/devel/llvm-cheri/llvm-project-37c49ff00e3eadce5d8703fdc4497f28458c64a8/polly/lib/External/ppcg/ |
H A D | gpu_array_tile.c | 28 struct gpu_array_tile *gpu_array_tile_create(isl_ctx *ctx, int n_index) in gpu_array_tile_create() argument 38 tile->bound = isl_alloc_array(ctx, struct gpu_array_bound, n_index); in gpu_array_tile_create() 42 tile->n = n_index; in gpu_array_tile_create() 44 for (i = 0; i < n_index; ++i) { in gpu_array_tile_create()
|
/dports/devel/llvm-devel/llvm-project-f05c95f10fc1d8171071735af8ad3a9e87633120/polly/lib/External/ppcg/ |
H A D | gpu_array_tile.c | 28 struct gpu_array_tile *gpu_array_tile_create(isl_ctx *ctx, int n_index) in gpu_array_tile_create() argument 38 tile->bound = isl_alloc_array(ctx, struct gpu_array_bound, n_index); in gpu_array_tile_create() 42 tile->n = n_index; in gpu_array_tile_create() 44 for (i = 0; i < n_index; ++i) { in gpu_array_tile_create()
|
/dports/devel/llvm12/llvm-project-12.0.1.src/polly/lib/External/ppcg/ |
H A D | gpu_array_tile.c | 28 struct gpu_array_tile *gpu_array_tile_create(isl_ctx *ctx, int n_index) in gpu_array_tile_create() argument 38 tile->bound = isl_alloc_array(ctx, struct gpu_array_bound, n_index); in gpu_array_tile_create() 42 tile->n = n_index; in gpu_array_tile_create() 44 for (i = 0; i < n_index; ++i) { in gpu_array_tile_create()
|
/dports/devel/wasi-compiler-rt12/llvm-project-12.0.1.src/polly/lib/External/ppcg/ |
H A D | gpu_array_tile.c | 28 struct gpu_array_tile *gpu_array_tile_create(isl_ctx *ctx, int n_index) in gpu_array_tile_create() argument 38 tile->bound = isl_alloc_array(ctx, struct gpu_array_bound, n_index); in gpu_array_tile_create() 42 tile->n = n_index; in gpu_array_tile_create() 44 for (i = 0; i < n_index; ++i) { in gpu_array_tile_create()
|
/dports/devel/wasi-compiler-rt13/llvm-project-13.0.1.src/polly/lib/External/ppcg/ |
H A D | gpu_array_tile.c | 28 struct gpu_array_tile *gpu_array_tile_create(isl_ctx *ctx, int n_index) in gpu_array_tile_create() argument 38 tile->bound = isl_alloc_array(ctx, struct gpu_array_bound, n_index); in gpu_array_tile_create() 42 tile->n = n_index; in gpu_array_tile_create() 44 for (i = 0; i < n_index; ++i) { in gpu_array_tile_create()
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/llvm/polly/lib/External/ppcg/ |
H A D | gpu_array_tile.c | 28 struct gpu_array_tile *gpu_array_tile_create(isl_ctx *ctx, int n_index) in gpu_array_tile_create() argument 38 tile->bound = isl_alloc_array(ctx, struct gpu_array_bound, n_index); in gpu_array_tile_create() 42 tile->n = n_index; in gpu_array_tile_create() 44 for (i = 0; i < n_index; ++i) { in gpu_array_tile_create()
|
/dports/devel/tinygo/tinygo-0.14.1/llvm-project/polly/lib/External/ppcg/ |
H A D | gpu_array_tile.c | 28 struct gpu_array_tile *gpu_array_tile_create(isl_ctx *ctx, int n_index) in gpu_array_tile_create() argument 38 tile->bound = isl_alloc_array(ctx, struct gpu_array_bound, n_index); in gpu_array_tile_create() 42 tile->n = n_index; in gpu_array_tile_create() 44 for (i = 0; i < n_index; ++i) { in gpu_array_tile_create()
|
/dports/devel/wasi-libcxx/llvm-project-13.0.1.src/polly/lib/External/ppcg/ |
H A D | gpu_array_tile.c | 28 struct gpu_array_tile *gpu_array_tile_create(isl_ctx *ctx, int n_index) in gpu_array_tile_create() argument 38 tile->bound = isl_alloc_array(ctx, struct gpu_array_bound, n_index); in gpu_array_tile_create() 42 tile->n = n_index; in gpu_array_tile_create() 44 for (i = 0; i < n_index; ++i) { in gpu_array_tile_create()
|
/dports/devel/llvm13/llvm-project-13.0.1.src/polly/lib/External/ppcg/ |
H A D | gpu_array_tile.c | 28 struct gpu_array_tile *gpu_array_tile_create(isl_ctx *ctx, int n_index) in gpu_array_tile_create() argument 38 tile->bound = isl_alloc_array(ctx, struct gpu_array_bound, n_index); in gpu_array_tile_create() 42 tile->n = n_index; in gpu_array_tile_create() 44 for (i = 0; i < n_index; ++i) { in gpu_array_tile_create()
|
/dports/misc/mxnet/incubator-mxnet-1.9.0/example/ssd/dataset/ |
H A D | concat_db.py | 108 n_db, n_index = self._locate_index(index) 109 return self.imdbs[n_db].image_path_from_index(n_index) 126 n_db, n_index = self._locate_index(index) 127 return self.imdbs[n_db].label_from_index(n_index)
|
/dports/misc/py-mxnet/incubator-mxnet-1.9.0/example/ssd/dataset/ |
H A D | concat_db.py | 108 n_db, n_index = self._locate_index(index) 109 return self.imdbs[n_db].image_path_from_index(n_index) 126 n_db, n_index = self._locate_index(index) 127 return self.imdbs[n_db].label_from_index(n_index)
|
/dports/devel/vitables/ViTables-3.0.2/vitables/vttables/ |
H A D | df_model.py | 218 n_columns, n_index = self._nheaders 220 and section < n_index) 250 n_columns, n_index = self._nheaders 256 is_index = col < n_index 265 if n_index > 1: 276 val = df.columns[col - n_index] 287 return str(self._chunk.iat[row - n_columns, col - n_index])
|
/dports/misc/mxnet/incubator-mxnet-1.9.0/3rdparty/mshadow/mshadow/cuda/ |
H A D | tensor_gpu-inl.cuh | 507 for (index_t n_index = n; n_index < nmax; n_index += n_size) { in Softmax3DGradKernel() local 511 dst[y][i][n_index] = src[y][i][n_index] - 1.0f; in Softmax3DGradKernel() 513 dst[y][i][n_index] = src[y][i][n_index]; in Softmax3DGradKernel() 529 for (index_t n_index = n; n_index < nmax; n_index += n_size) { in Softmax3DGradKernel() local 533 dst[y][i][n_index] = 0.0f; in Softmax3DGradKernel() 538 dst[y][i][n_index] = src[y][i][n_index] - 1.0f; in Softmax3DGradKernel() 540 dst[y][i][n_index] = src[y][i][n_index]; in Softmax3DGradKernel() 556 for (index_t n_index = n; n_index < nmax; n_index += n_size) { in Softmax3DKernel() local 557 DType smax = src[y][0][n_index]; in Softmax3DKernel() 565 dst[y][i][n_index] = p; in Softmax3DKernel() [all …]
|