/dports/databases/py-partd/partd-1.2.0/partd/ |
H A D | encode.py | 8 def __init__(self, encode, decode, join, partd=None): argument 9 if not partd or isinstance(partd, str): 10 partd = File(partd) 11 self.partd = partd 25 self.partd.append(data, **kwargs) 28 raw = self.partd._get(keys, **kwargs) 33 return self.partd.delete(keys, **kwargs) 36 return self.partd.iset(key, frame(self.encode(value)), **kwargs) 39 return self.partd.drop() 43 return self.partd.lock [all …]
|
H A D | numpy.py | 43 def __init__(self, partd=None): argument 44 if not partd or isinstance(partd, str): 45 partd = File(partd) 46 self.partd = partd 50 return {'partd': self.partd} 66 self.partd.delete(keys2, **kwargs) 69 return self.partd._iset(key, value) 72 return self.partd.drop() 75 self.partd.__del__() 79 return self.partd.lock [all …]
|
H A D | pandas.py | 39 def __init__(self, partd=None): argument 40 self.partd = pnp.Numpy(partd) 56 self.partd.append(arrays, **kwargs) 60 columns = self.partd.partd.get([extend(k, '.columns') for k in keys], 65 index_names = self.partd.partd.get([extend(k, '.index-name') 79 return {'partd': self.partd} 82 return self.partd._iset(key, value) 85 return self.partd.drop() 89 return self.partd.partd.lock 93 self.partd.__exit__(self, *args) [all …]
|
H A D | zmq.py | 45 if partd is None: 46 partd = Buffer(Dict(), File()) 47 self.partd = partd 64 self.partd.lock.acquire() 111 self.partd.append(data, lock=False) 131 self.partd.delete(keys, lock=False) 158 self.partd.append(data, lock=False) 163 self.partd.drop() 169 result = self.partd.get(keys, lock=False) 180 self.partd.lock.release() [all …]
|
H A D | core.py | 88 with self.partd.lock: 89 result = self.partd.get(keys, lock=False) 90 self.partd.delete(keys, lock=False)
|
/dports/databases/py-partd/partd-1.2.0/partd/tests/ |
H A D | test_pandas.py | 11 from partd.pandas import PandasColumns, PandasBlocks, serialize, deserialize 27 assert os.path.exists(p.partd.partd.path) 31 assert os.path.exists(p.partd.partd.filename('x')) 32 assert os.path.exists(p.partd.partd.filename(('x', 'a'))) 33 assert os.path.exists(p.partd.partd.filename(('x', '.index'))) 34 assert os.path.exists(p.partd.partd.filename('y')) 43 assert not os.path.exists(p.partd.partd.path) 56 assert os.path.exists(p.partd.path) 60 assert os.path.exists(p.partd.filename('x')) 61 assert os.path.exists(p.partd.filename('y')) [all …]
|
H A D | test_pickle.py | 1 from partd.pickle import Pickle 11 assert os.path.exists(p.partd.filename('x')) 12 assert os.path.exists(p.partd.filename('y')) 21 assert not os.path.exists(p.partd.path)
|
H A D | test_compressed.py | 1 from partd.compressed import ZLib 13 assert os.path.exists(p.partd.filename('x')) 14 assert os.path.exists(p.partd.filename('y')) 24 assert not os.path.exists(p.partd.path)
|
H A D | test_numpy.py | 8 import partd 9 from partd.numpy import Numpy 71 s = partd.numpy.serialize(a) 72 assert (partd.numpy.deserialize(s, 'O') == a).all()
|
H A D | test_partd.py | 1 from partd import File 2 from partd.core import token, escape_filename, filename 3 from partd import core
|
H A D | test_buffer.py | 1 from partd.dict import Dict 2 from partd.file import File 3 from partd.buffer import Buffer, keys_to_flush
|
H A D | test_encode.py | 1 from partd.file import File 2 from partd.encode import Encode
|
H A D | test_zmq.py | 4 from partd.zmq import Server, keys_to_flush, File, Client 5 from partd import core, Dict
|
H A D | test_python.py | 1 from partd.python import dumps, loads
|
H A D | test_utils.py | 1 from partd.utils import frame, framesplit
|
/dports/databases/py-partd/partd-1.2.0/ |
H A D | PKG-INFO | 2 Name: partd 5 Home-page: http://github.com/dask/partd/ 33 >>> import partd 34 >>> p = partd.File('/path/to/new/dataset/') 49 4. Destroy partd dataset:: 59 We can back a partd by an in-memory dictionary:: 68 …ne partd with another, keeping a fixed maximum of data in the buffering partd. This writes the la… 96 another partd as an argument.:: 135 :target: https://github.com/dask/partd/actions?query=workflow%3ACI 136 .. |Version Status| image:: https://img.shields.io/pypi/v/partd.svg [all …]
|
H A D | README.rst | 25 >>> import partd 26 >>> p = partd.File('/path/to/new/dataset/') 41 4. Destroy partd dataset:: 51 We can back a partd by an in-memory dictionary:: 60 …ne partd with another, keeping a fixed maximum of data in the buffering partd. This writes the la… 80 Once we can robustly and efficiently append bytes to a partd we consider 82 partd, which accepts three functions, one to apply on bytes as they are 88 another partd as an argument.:: 127 :target: https://github.com/dask/partd/actions?query=workflow%3ACI 128 .. |Version Status| image:: https://img.shields.io/pypi/v/partd.svg [all …]
|
H A D | setup.cfg | 4 versionfile_source = partd/_version.py 5 versionfile_build = partd/_version.py 7 parentdir_prefix = partd-
|
H A D | MANIFEST.in | 1 recursive-include partd *.py 9 include partd/_version.py
|
/dports/databases/py-partd/partd-1.2.0/partd.egg-info/ |
H A D | PKG-INFO | 2 Name: partd 5 Home-page: http://github.com/dask/partd/ 33 >>> import partd 34 >>> p = partd.File('/path/to/new/dataset/') 49 4. Destroy partd dataset:: 59 We can back a partd by an in-memory dictionary:: 68 …ne partd with another, keeping a fixed maximum of data in the buffering partd. This writes the la… 96 another partd as an argument.:: 135 :target: https://github.com/dask/partd/actions?query=workflow%3ACI 136 .. |Version Status| image:: https://img.shields.io/pypi/v/partd.svg [all …]
|
/dports/devel/py-dask/dask-2021.11.2/dask/dataframe/ |
H A D | shuffle.py | 494 import partd 500 getattr(partd.compressed, self.compression) 511 file = partd.File(path) 512 partd.file.cleanup_files.append(path) 517 return partd.PandasBlocks(partd.Buffer(partd.Dict(), file)) 519 return partd.PandasBlocks(file) 772 import partd 774 if isinstance(p, partd.Encode): 775 maybe_file = p.partd 779 if isinstance(maybe_file, partd.File):
|
/dports/devel/py-pip/pip-20.3.4/tests/yaml/ |
H A D | huge.yml | 264 …re 0.15.2.*', 'distributed >=1.16.0', 'numpy >=1.10', 'pandas >=0.19.0', 'partd >=0.3.8', 'toolz >… 267 …re 0.15.3.*', 'distributed >=1.19.0', 'numpy >=1.10', 'pandas >=0.19.0', 'partd >=0.3.8', 'toolz >… 270 …re 0.15.4.*', 'distributed >=1.19.0', 'numpy >=1.10', 'pandas >=0.19.0', 'partd >=0.3.8', 'toolz >… 273 …re 0.16.0.*', 'distributed >=1.20.0', 'numpy >=1.10', 'pandas >=0.19.0', 'partd >=0.3.8', 'toolz >… 276 …re 0.16.1.*', 'distributed >=1.20.0', 'numpy >=1.10', 'pandas >=0.19.0', 'partd >=0.3.8', 'toolz >… 279 …re 0.17.0.*', 'distributed >=1.21.0', 'numpy >=1.10', 'pandas >=0.19.0', 'partd >=0.3.8', 'toolz >… 282 …re 0.17.1.*', 'distributed >=1.21.1', 'numpy >=1.10', 'pandas >=0.19.0', 'partd >=0.3.8', 'toolz >… 285 … 0.17.2.*', 'distributed >=1.21.0', 'numpy >=1.10.4', 'pandas >=0.19.0', 'partd >=0.3.8', 'toolz >… 288 … 0.17.3.*', 'distributed >=1.21.0', 'numpy >=1.11.0', 'pandas >=0.19.0', 'partd >=0.3.8', 'toolz >… 291 … 0.17.4.*', 'distributed >=1.21.0', 'numpy >=1.11.0', 'pandas >=0.19.0', 'partd >=0.3.8', 'toolz >… [all …]
|
/dports/databases/py-partd/ |
H A D | .license-catalog.mk | 5 _LICENSE_DISTFILES=partd-1.2.0.tar.gz
|
/dports/databases/percona56-client/percona-server-5.6.51-91.0/mysql-test/suite/parts/inc/ |
H A D | partition_methods1.inc | 160 PARTITION partd VALUES LESS THAN ($max_row_div2 + $max_row_div4), 175 PARTITION partd VALUES LESS THAN ($max_row_div2 + $max_row_div4) 205 PARTITION partd VALUES LESS THAN $MAX_VALUE); 219 PARTITION partd VALUES LESS THAN $MAX_VALUE
|
H A D | partition_methods2.inc | 154 PARTITION partd VALUES LESS THAN ($max_row_div2 + $max_row_div4), 169 PARTITION partd VALUES LESS THAN ($max_row_div2 + $max_row_div4) 198 PARTITION partd VALUES LESS THAN $MAX_VALUE); 212 PARTITION partd VALUES LESS THAN $MAX_VALUE
|