/dports/science/py-dlib/dlib-19.22/dlib/cuda/ |
H A D | gpu_data_abstract.h | 14 class gpu_data 39 gpu_data( 52 gpu_data(const gpu_data&) = delete; 53 gpu_data& operator=(const gpu_data&) = delete; 54 gpu_data(gpu_data&& item); 55 gpu_data& operator=(gpu_data&& item); 205 gpu_data& item 221 gpu_data& dest, 222 const gpu_data& src 237 gpu_data& dest, [all …]
|
H A D | gpu_data.h | 17 class gpu_data 42 gpu_data( in gpu_data() function 48 gpu_data(const gpu_data&) = delete; 49 gpu_data& operator=(const gpu_data&) = delete; 52 gpu_data(gpu_data&& item) : gpu_data() { swap(item); } in gpu_data() function 53 gpu_data& operator=(gpu_data&& item) { swap(item); return *this; } 210 void memcpy (gpu_data& dest, const gpu_data& src); 213 gpu_data& dest, 215 const gpu_data& src, 222 inline void memcpy (gpu_data& dest, const gpu_data& src) in memcpy() [all …]
|
H A D | gpu_data.cpp | 22 gpu_data& dest, in memcpy() 23 const gpu_data& src in memcpy() 34 gpu_data& dest, in memcpy() 36 const gpu_data& src, in memcpy() 108 void gpu_data:: 121 void gpu_data:: 130 void gpu_data:: 147 void gpu_data:: 165 void gpu_data::
|
H A D | tensor.h | 200 virtual gpu_data& data() = 0; 201 virtual const gpu_data& data() const = 0; 438 gpu_data data_instance; 440 virtual gpu_data& data() { return data_instance; } in data() 441 virtual const gpu_data& data() const { return data_instance; } in data() 557 gpu_data* data_instance; 560 virtual gpu_data& data() { return *data_instance; } in data() 561 virtual const gpu_data& data() const { return *data_instance; } in data()
|
/dports/science/dlib-cpp/dlib-19.22/dlib/cuda/ |
H A D | gpu_data_abstract.h | 14 class gpu_data 39 gpu_data( 52 gpu_data(const gpu_data&) = delete; 53 gpu_data& operator=(const gpu_data&) = delete; 54 gpu_data(gpu_data&& item); 55 gpu_data& operator=(gpu_data&& item); 205 gpu_data& item 221 gpu_data& dest, 222 const gpu_data& src 237 gpu_data& dest, [all …]
|
H A D | gpu_data.h | 17 class gpu_data 42 gpu_data( in gpu_data() function 48 gpu_data(const gpu_data&) = delete; 49 gpu_data& operator=(const gpu_data&) = delete; 52 gpu_data(gpu_data&& item) : gpu_data() { swap(item); } in gpu_data() function 53 gpu_data& operator=(gpu_data&& item) { swap(item); return *this; } 210 void memcpy (gpu_data& dest, const gpu_data& src); 213 gpu_data& dest, 215 const gpu_data& src, 222 inline void memcpy (gpu_data& dest, const gpu_data& src) in memcpy() [all …]
|
H A D | gpu_data.cpp | 22 gpu_data& dest, in memcpy() 23 const gpu_data& src in memcpy() 34 gpu_data& dest, in memcpy() 36 const gpu_data& src, in memcpy() 108 void gpu_data:: 121 void gpu_data:: 130 void gpu_data:: 147 void gpu_data:: 165 void gpu_data::
|
H A D | tensor.h | 200 virtual gpu_data& data() = 0; 201 virtual const gpu_data& data() const = 0; 438 gpu_data data_instance; 440 virtual gpu_data& data() { return data_instance; } in data() 441 virtual const gpu_data& data() const { return data_instance; } in data() 557 gpu_data* data_instance; 560 virtual gpu_data& data() { return *data_instance; } in data() 561 virtual const gpu_data& data() const { return *data_instance; } in data()
|
/dports/devel/py-numba/numba-0.51.2/numba/cuda/cudadrv/ |
H A D | devicearray.py | 113 self.gpu_data = gpu_data 300 gpu_data = self.gpu_data.view(begin * itemsize, end * itemsize) 302 gpu_data=gpu_data) 345 gpu_data=self.gpu_data, 382 gpu_data=self.gpu_data, 515 dtype=self.dtype, gpu_data=self.gpu_data) 521 dtype=self.dtype, gpu_data=self.gpu_data) 536 dtype=self.dtype, gpu_data=self.gpu_data, 682 self.gpu_data = gpu_data 688 writeback=ary, stream=stream, gpu_data=gpu_data) [all …]
|
H A D | ndarray.py | 19 gpu_data = devices.get_context().memalloc(datasize) 20 return gpu_data
|
/dports/devel/taskflow/taskflow-3.2.0/doxygen/sycl_algorithms/ |
H A D | sycl_for_each.dox | 32 The following example creates a kernel that assigns each element of @c gpu_data 38 // assigns each element in gpu_data to 1 over the range [0, 100) with step size 1 39 sf.for_each_index(0, 100, 1, [gpu_data] (int idx) { 40 gpu_data[idx] = 1; 61 @c gpu_data to 1 over the range <tt>[gpu_data, gpu_data + 1000)</tt>. 66 // assigns each element to 1 over the range [gpu_data, gpu_data + 1000) 67 cf.for_each(gpu_data, gpu_data + 1000, [] (int& item) {
|
H A D | sycl_transform.dox | 31 starting from @c gpu_data to <tt>gpu_data + 1000</tt>, 37 // gpu_data[i] = gpu_data_x[i] + gpu_data_y[i] + gpu_data_z[i] 39 gpu_data, gpu_data + 1000,
|
/dports/devel/taskflow/taskflow-3.2.0/doxygen/cudaflow_algorithms/ |
H A D | cudaflow_for_each.dox | 32 The following example creates a kernel that assigns each entry of @c gpu_data 38 // assigns each element in gpu_data to 1 over the range [0, 100) with step size 1 39 cf.for_each_index(0, 100, 1, [gpu_data] __device__ (int idx) { 40 gpu_data[idx] = 1; 61 @c gpu_data to 1 over the range <tt>[gpu_data, gpu_data + 1000)</tt>. 66 // assigns each element to 1 over the range [gpu_data, gpu_data + 1000) 67 cf.for_each(gpu_data, gpu_data + 1000, [] __device__ (int& item) {
|
/dports/devel/py-distributed/distributed-2021.11.2/distributed/protocol/ |
H A D | numba.py | 28 shape=(x.nbytes,), strides=(1,), dtype=np.dtype("u1"), gpu_data=x.gpu_data 45 gpu_data=numba.cuda.as_cuda_array(frame).gpu_data,
|
/dports/devel/py-numba/numba-0.51.2/numba/cuda/ |
H A D | dispatcher.py | 133 gpu_data=ary.gpu_data) 142 gpu_data=ary.gpu_data) 183 gpu_data=ary.gpu_data)
|
/dports/www/qt5-webengine/qtwebengine-everywhere-src-5.15.2/src/3rdparty/chromium/gpu/command_buffer/service/ |
H A D | gles2_cmd_decoder_unittest_buffers.cc | 175 std::vector<int8_t> gpu_data(kTotalSize); in TEST_P() local 177 gpu_data[ii] = static_cast<int8_t>(ii % 128); in TEST_P() 179 DoBufferSubData(kTarget, 0, kTotalSize, &gpu_data[0]); in TEST_P() 194 .WillOnce(Return(&gpu_data[kOffset])) in TEST_P() 226 EXPECT_EQ(kValue0, gpu_data[ii]); in TEST_P() 259 std::vector<int8_t> gpu_data(kTotalSize); in TEST_P() local 261 gpu_data[ii] = static_cast<int8_t>(ii % 128); in TEST_P() 263 DoBufferSubData(kTarget, 0, kTotalSize, &gpu_data[0]); in TEST_P() 278 .WillOnce(Return(&gpu_data[kMappedOffset])) in TEST_P() 311 EXPECT_EQ(kValue0, gpu_data[ii]); in TEST_P() [all …]
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/gpu/command_buffer/service/ |
H A D | gles2_cmd_decoder_unittest_buffers.cc | 175 std::vector<int8_t> gpu_data(kTotalSize); in TEST_P() local 177 gpu_data[ii] = static_cast<int8_t>(ii % 128); in TEST_P() 179 DoBufferSubData(kTarget, 0, kTotalSize, &gpu_data[0]); in TEST_P() 194 .WillOnce(Return(&gpu_data[kOffset])) in TEST_P() 226 EXPECT_EQ(kValue0, gpu_data[ii]); in TEST_P() 259 std::vector<int8_t> gpu_data(kTotalSize); in TEST_P() local 261 gpu_data[ii] = static_cast<int8_t>(ii % 128); in TEST_P() 263 DoBufferSubData(kTarget, 0, kTotalSize, &gpu_data[0]); in TEST_P() 278 .WillOnce(Return(&gpu_data[kMappedOffset])) in TEST_P() 311 EXPECT_EQ(kValue0, gpu_data[ii]); in TEST_P() [all …]
|
/dports/math/deal.ii/dealii-803d21ff957e349b3799cd3ef2c840bc78734305/examples/step-64/ |
H A D | step-64.cu | 76 const typename CUDAWrappers::MatrixFree<dim, double>::Data *gpu_data); 99 const typename CUDAWrappers::MatrixFree<dim, double>::Data *gpu_data) in operator ()() argument 102 cell, gpu_data, n_dofs_1d, n_q_points); in operator ()() 105 gpu_data, in operator ()() 175 const typename CUDAWrappers::MatrixFree<dim, double>::Data *gpu_data, 200 const typename CUDAWrappers::MatrixFree<dim, double>::Data *gpu_data, in operator ()() argument 206 cell, gpu_data, n_dofs_1d, n_q_points); in operator ()() 209 fe_eval(cell, gpu_data, shared_data); in operator ()()
|
/dports/devel/py-numba/numba-0.51.2/numba/cuda/tests/cudapy/ |
H A D | test_datetime.py | 80 self.assertEqual(viewed.gpu_data, darr.gpu_data) 90 self.assertEqual(viewed.gpu_data, darr.gpu_data)
|
/dports/graphics/blender/blender-2.91.0/release/scripts/addons/mesh_snap_utilities_line/snap_context_l/ |
H A D | __init__.py | 327 gpu_data = snap_obj.data[1] 329 if gpu_data.draw_tris: 332 tri_verts = gpu_data.get_tri_verts(index) 333 tri_co = [snap_obj.mat @ Vector(v) for v in gpu_data.get_tri_co(index)] 341 if gpu_data.draw_edges: 344 edge_verts = gpu_data.get_edge_verts(index) 345 edge_co = [snap_obj.mat @ Vector(v) for v in gpu_data.get_edge_co(index)] 366 if gpu_data.draw_verts: 368 co = snap_obj.mat @ Vector(gpu_data.get_loosevert_co(index)) 369 return co, (gpu_data.get_loosevert_index(index),), co
|
/dports/devel/py-numba/numba-0.51.2/numba/cuda/tests/cudadrv/ |
H A D | test_cuda_devicerecord.py | 31 self.assertIsNotNone(rec.gpu_data) 69 self.assertNotEqual(devrec.gpu_data, devrec2.gpu_data)
|
/dports/misc/py-mxnet/incubator-mxnet-1.9.0/tests/python/gpu/ |
H A D | test_gluon_model_zoo_gpu.py | 63 gpu_data = data.as_in_context(mx.gpu()) 89 gpu_out = gpu_model(gpu_data) 127 gpu_data = data.as_in_context(mx.gpu()) 159 gpu_out = gpu_model(gpu_data)
|
/dports/misc/mxnet/incubator-mxnet-1.9.0/tests/python/gpu/ |
H A D | test_gluon_model_zoo_gpu.py | 63 gpu_data = data.as_in_context(mx.gpu()) 89 gpu_out = gpu_model(gpu_data) 127 gpu_data = data.as_in_context(mx.gpu()) 159 gpu_out = gpu_model(gpu_data)
|
/dports/www/grafana8/grafana-8.3.6/vendor/github.com/apache/arrow/python/pyarrow/tests/ |
H A D | test_cuda_numba_interop.py | 168 darr = DeviceNDArray((size,), (dtype.itemsize,), dtype, gpu_data=mem) 190 darr = DeviceNDArray(arr.shape, arr.strides, arr.dtype, gpu_data=mem) 205 darr = DeviceNDArray(arr.shape, arr.strides, arr.dtype, gpu_data=mem) 231 darr = DeviceNDArray(arr.shape, arr.strides, arr.dtype, gpu_data=mem)
|
/dports/databases/arrow/apache-arrow-6.0.1/python/pyarrow/tests/ |
H A D | test_cuda_numba_interop.py | 168 darr = DeviceNDArray((size,), (dtype.itemsize,), dtype, gpu_data=mem) 190 darr = DeviceNDArray(arr.shape, arr.strides, arr.dtype, gpu_data=mem) 205 darr = DeviceNDArray(arr.shape, arr.strides, arr.dtype, gpu_data=mem) 231 darr = DeviceNDArray(arr.shape, arr.strides, arr.dtype, gpu_data=mem)
|