/dports/science/lammps/lammps-stable_29Sep2021/lib/gpu/ |
H A D | lal_balance.h | 65 inline int get_gpu_count(const int ago, const int inum_full); 104 inline int balance(const int ago,const int inum_full,const double cpu_time) { in balance() argument 106 return get_gpu_count(ago,inum_full); in balance() 149 int BalanceT::get_gpu_count(const int ago, const int inum_full) { in get_gpu_count() argument 154 _inum_full=inum_full; in get_gpu_count() 161 _inum=static_cast<int>(floor(_actual_split*inum_full)); in get_gpu_count()
|
H A D | lal_re_squared_ext.cpp | 102 int** compute(const int ago, const int inum_full, const int nall, 110 int** re_gpu_compute_n(const int ago, const int inum_full, const int nall, in re_gpu_compute_n() argument 117 return REMF.compute(ago, inum_full, nall, host_x, host_type, sublo, subhi, in re_gpu_compute_n() 122 int * re_gpu_compute(const int ago, const int inum_full, const int nall, in re_gpu_compute() argument 127 return REMF.compute(ago, inum_full, nall, host_x, host_type, ilist, in re_gpu_compute()
|
H A D | lal_gayberne_ext.cpp | 105 int** compute(const int ago, const int inum_full, const int nall, 113 int** gb_gpu_compute_n(const int ago, const int inum_full, const int nall, in gb_gpu_compute_n() argument 120 return GBMF.compute(ago, inum_full, nall, host_x, host_type, sublo, subhi, in gb_gpu_compute_n() 125 int * gb_gpu_compute(const int ago, const int inum_full, const int nall, in gb_gpu_compute() argument 130 return GBMF.compute(ago, inum_full, nall, host_x, host_type, ilist, in gb_gpu_compute()
|
H A D | lal_yukawa_colloid.cpp | 134 void YukawaColloidT::compute(const int f_ago, const int inum_full, in compute() argument 168 if (inum_full==0) { in compute() 177 int inum=this->hd_balancer.balance(ago,inum_full,cpu_time); in compute() 205 int** YukawaColloidT::compute(const int ago, const int inum_full, in compute() argument 240 if (inum_full==0) { in compute() 250 int inum=this->hd_balancer.get_gpu_count(ago,inum_full); in compute() 256 this->build_nbor_list(inum, inum_full-inum, nall, host_x, host_type, in compute()
|
H A D | lal_base_dipole.cpp | 190 void BaseDipoleT::compute(const int f_ago, const int inum_full, in compute() argument 213 if (inum_full==0) { in compute() 222 int inum=hd_balancer.balance(ago,inum_full,cpu_time); in compute() 253 int** BaseDipoleT::compute(const int ago, const int inum_full, in compute() argument 278 if (inum_full==0) { in compute() 287 int inum=hd_balancer.get_gpu_count(ago,inum_full); in compute() 293 build_nbor_list(inum, inum_full-inum, nall, host_x, host_type, in compute() 312 device->precompute(ago,inum_full,nall,host_x,host_type,success,host_q, in compute()
|
H A D | lal_base_charge.cpp | 189 void BaseChargeT::compute(const int f_ago, const int inum_full, in compute() argument 212 if (inum_full==0) { in compute() 221 int inum=hd_balancer.balance(ago,inum_full,cpu_time); in compute() 250 int** BaseChargeT::compute(const int ago, const int inum_full, in compute() argument 274 if (inum_full==0) { in compute() 283 int inum=hd_balancer.get_gpu_count(ago,inum_full); in compute() 289 build_nbor_list(inum, inum_full-inum, nall, host_x, host_type, in compute() 305 device->precompute(ago,inum_full,nall,host_x,host_type,success,host_q, in compute()
|
H A D | lal_morse_ext.cpp | 96 int** mor_gpu_compute_n(const int ago, const int inum_full, in mor_gpu_compute_n() argument 103 return MORMF.compute(ago, inum_full, nall, host_x, host_type, sublo, in mor_gpu_compute_n() 108 void mor_gpu_compute(const int ago, const int inum_full, const int nall, in mor_gpu_compute() argument 113 MORMF.compute(ago,inum_full,nall,host_x,host_type,ilist,numj, in mor_gpu_compute()
|
H A D | lal_lj96_ext.cpp | 95 int** lj96_gpu_compute_n(const int ago, const int inum_full, in lj96_gpu_compute_n() argument 102 return LJ96MF.compute(ago, inum_full, nall, host_x, host_type, sublo, in lj96_gpu_compute_n() 107 void lj96_gpu_compute(const int ago, const int inum_full, const int nall, in lj96_gpu_compute() argument 112 LJ96MF.compute(ago,inum_full,nall,host_x,host_type,ilist,numj,firstneigh, in lj96_gpu_compute()
|
H A D | lal_table_ext.cpp | 95 int ** table_gpu_compute_n(const int ago, const int inum_full, in table_gpu_compute_n() argument 102 return TBMF.compute(ago, inum_full, nall, host_x, host_type, sublo, in table_gpu_compute_n() 107 void table_gpu_compute(const int ago, const int inum_full, const int nall, in table_gpu_compute() argument 112 TBMF.compute(ago,inum_full,nall,host_x,host_type,ilist,numj, in table_gpu_compute()
|
H A D | lal_beck_ext.cpp | 95 int ** beck_gpu_compute_n(const int ago, const int inum_full, in beck_gpu_compute_n() argument 102 return BLMF.compute(ago, inum_full, nall, host_x, host_type, sublo, in beck_gpu_compute_n() 107 void beck_gpu_compute(const int ago, const int inum_full, const int nall, in beck_gpu_compute() argument 112 BLMF.compute(ago,inum_full,nall,host_x,host_type,ilist,numj, in beck_gpu_compute()
|
H A D | lal_yukawa_colloid_ext.cpp | 95 int ** ykcolloid_gpu_compute_n(const int ago, const int inum_full, in ykcolloid_gpu_compute_n() argument 102 return YKCOLLMF.compute(ago, inum_full, nall, host_x, host_type, sublo, in ykcolloid_gpu_compute_n() 108 void ykcolloid_gpu_compute(const int ago, const int inum_full, in ykcolloid_gpu_compute() argument 114 YKCOLLMF.compute(ago,inum_full,nall,host_x,host_type,ilist,numj, in ykcolloid_gpu_compute()
|
H A D | lal_yukawa_ext.cpp | 95 int ** yukawa_gpu_compute_n(const int ago, const int inum_full, in yukawa_gpu_compute_n() argument 102 return YKMF.compute(ago, inum_full, nall, host_x, host_type, sublo, in yukawa_gpu_compute_n() 107 void yukawa_gpu_compute(const int ago, const int inum_full, const int nall, in yukawa_gpu_compute() argument 112 YKMF.compute(ago,inum_full,nall,host_x,host_type,ilist,numj, in yukawa_gpu_compute()
|
/dports/science/liggghts/LIGGGHTS-PUBLIC-3.8.0-26-g6e873439/lib/gpu/ |
H A D | lal_balance.h | 65 inline int get_gpu_count(const int ago, const int inum_full); 104 inline int balance(const int ago,const int inum_full,const double cpu_time) { in balance() argument 106 return get_gpu_count(ago,inum_full); in balance() 149 int BalanceT::get_gpu_count(const int ago, const int inum_full) { in get_gpu_count() argument 154 _inum_full=inum_full; in get_gpu_count() 161 _inum=static_cast<int>(floor(_actual_split*inum_full)); in get_gpu_count()
|
H A D | lal_base_charge.cpp | 179 void BaseChargeT::compute(const int f_ago, const int inum_full, in compute() argument 188 if (inum_full==0) { in compute() 197 int inum=hd_balancer.balance(ago,inum_full,cpu_time); in compute() 226 int** BaseChargeT::compute(const int ago, const int inum_full, in compute() argument 236 if (inum_full==0) { in compute() 245 int inum=hd_balancer.get_gpu_count(ago,inum_full); in compute() 251 build_nbor_list(inum, inum_full-inum, nall, host_x, host_type, in compute() 267 device->precompute(ago,inum_full,nall,host_x,host_type,success,host_q, in compute()
|
H A D | lal_base_dipole.cpp | 181 void BaseDipoleT::compute(const int f_ago, const int inum_full, in compute() argument 190 if (inum_full==0) { in compute() 199 int inum=hd_balancer.balance(ago,inum_full,cpu_time); in compute() 230 int** BaseDipoleT::compute(const int ago, const int inum_full, in compute() argument 241 if (inum_full==0) { in compute() 250 int inum=hd_balancer.get_gpu_count(ago,inum_full); in compute() 256 build_nbor_list(inum, inum_full-inum, nall, host_x, host_type, in compute() 275 device->precompute(ago,inum_full,nall,host_x,host_type,success,host_q, in compute()
|
H A D | lal_yukawa_colloid.cpp | 134 void YukawaColloidT::compute(const int f_ago, const int inum_full, in compute() argument 153 if (inum_full==0) { in compute() 162 int inum=this->hd_balancer.balance(ago,inum_full,cpu_time); in compute() 190 int** YukawaColloidT::compute(const int ago, const int inum_full, const int nall, in compute() argument 211 if (inum_full==0) { in compute() 221 int inum=this->hd_balancer.get_gpu_count(ago,inum_full); in compute() 227 this->build_nbor_list(inum, inum_full-inum, nall, host_x, host_type, in compute()
|
H A D | lal_re_squared_ext.cpp | 102 int** compute(const int ago, const int inum_full, const int nall, 110 int** re_gpu_compute_n(const int ago, const int inum_full, const int nall, in re_gpu_compute_n() argument 117 return REMF.compute(ago, inum_full, nall, host_x, host_type, sublo, subhi, in re_gpu_compute_n() 122 int * re_gpu_compute(const int ago, const int inum_full, const int nall, in re_gpu_compute() argument 127 return REMF.compute(ago, inum_full, nall, host_x, host_type, ilist, in re_gpu_compute()
|
H A D | lal_gayberne_ext.cpp | 105 int** compute(const int ago, const int inum_full, const int nall, 113 int** gb_gpu_compute_n(const int ago, const int inum_full, const int nall, in gb_gpu_compute_n() argument 120 return GBMF.compute(ago, inum_full, nall, host_x, host_type, sublo, subhi, in gb_gpu_compute_n() 125 int * gb_gpu_compute(const int ago, const int inum_full, const int nall, in gb_gpu_compute() argument 130 return GBMF.compute(ago, inum_full, nall, host_x, host_type, ilist, in gb_gpu_compute()
|
H A D | lal_base_atomic.cpp | 176 void BaseAtomicT::compute(const int f_ago, const int inum_full, in compute() argument 184 if (inum_full==0) { in compute() 193 int inum=hd_balancer.balance(ago,inum_full,cpu_time); in compute() 217 int ** BaseAtomicT::compute(const int ago, const int inum_full, in compute() argument 226 if (inum_full==0) { in compute() 235 int inum=hd_balancer.get_gpu_count(ago,inum_full); in compute() 241 build_nbor_list(inum, inum_full-inum, nall, host_x, host_type, in compute()
|
H A D | lal_table_ext.cpp | 95 int ** table_gpu_compute_n(const int ago, const int inum_full, in table_gpu_compute_n() argument 102 return TBMF.compute(ago, inum_full, nall, host_x, host_type, sublo, in table_gpu_compute_n() 107 void table_gpu_compute(const int ago, const int inum_full, const int nall, in table_gpu_compute() argument 112 TBMF.compute(ago,inum_full,nall,host_x,host_type,ilist,numj, in table_gpu_compute()
|
H A D | lal_yukawa_colloid_ext.cpp | 95 int ** ykcolloid_gpu_compute_n(const int ago, const int inum_full, in ykcolloid_gpu_compute_n() argument 102 return YKCOLLMF.compute(ago, inum_full, nall, host_x, host_type, sublo, in ykcolloid_gpu_compute_n() 108 void ykcolloid_gpu_compute(const int ago, const int inum_full, in ykcolloid_gpu_compute() argument 114 YKCOLLMF.compute(ago,inum_full,nall,host_x,host_type,ilist,numj, in ykcolloid_gpu_compute()
|
H A D | lal_cg_cmm_ext.cpp | 96 int** cmm_gpu_compute_n(const int ago, const int inum_full, in cmm_gpu_compute_n() argument 103 return CMMMF.compute(ago, inum_full, nall, host_x, host_type, sublo, in cmm_gpu_compute_n() 108 void cmm_gpu_compute(const int ago, const int inum_full, const int nall, in cmm_gpu_compute() argument 113 CMMMF.compute(ago,inum_full,nall,host_x,host_type,ilist,numj, in cmm_gpu_compute()
|
H A D | lal_coul_long_ext.cpp | 94 int** cl_gpu_compute_n(const int ago, const int inum_full, in cl_gpu_compute_n() argument 102 return CLMF.compute(ago, inum_full, nall, host_x, host_type, sublo, in cl_gpu_compute_n() 108 void cl_gpu_compute(const int ago, const int inum_full, const int nall, in cl_gpu_compute() argument 114 CLMF.compute(ago,inum_full,nall,host_x,host_type,ilist,numj, in cl_gpu_compute()
|
H A D | lal_lj96_ext.cpp | 95 int** lj96_gpu_compute_n(const int ago, const int inum_full, in lj96_gpu_compute_n() argument 102 return LJ96MF.compute(ago, inum_full, nall, host_x, host_type, sublo, in lj96_gpu_compute_n() 107 void lj96_gpu_compute(const int ago, const int inum_full, const int nall, in lj96_gpu_compute() argument 112 LJ96MF.compute(ago,inum_full,nall,host_x,host_type,ilist,numj,firstneigh, in lj96_gpu_compute()
|
/dports/science/lammps/lammps-stable_29Sep2021/src/INTEL/ |
H A D | npair_halffull_newton_intel.cpp | 51 const int inum_full = list->listfull->inum; in build_t() local 68 IP_PRE_omp_range_id(ifrom, ito, tid, inum_full, comm->nthreads); in build_t() 129 list->inum = inum_full; in build_t() 140 const int inum_full = list->listfull->inum; in build_t3() local 158 IP_PRE_omp_range_id(ifrom, ito, tid, inum_full, packthreads); in build_t3() 204 list->inum = inum_full; in build_t3()
|