/dports/devel/py-dask/dask-2021.11.2/dask/dataframe/ |
H A D | shuffle.py | 51 npartitions = min(npartitions, df.npartitions) 98 npartitions = max(100, df.npartitions) 101 npartitions = df.npartitions 126 and npartitions == df.npartitions 182 npartitions = df.npartitions 369 npartitions=npartitions, 389 npartitions=npartitions or df.npartitions, 398 npartitions=npartitions, 461 if npartitions is not None and npartitions < df.npartitions: 532 npartitions = df.npartitions [all …]
|
H A D | multi.py | 331 npartitions = max(lhs.npartitions, rhs.npartitions) 652 n_small = min(left.npartitions, right.npartitions) 653 n_big = max(left.npartitions, right.npartitions) 1391 if lhs.npartitions < rhs.npartitions: 1392 rhs = rhs.repartition(npartitions=npartitions) 1394 lhs = lhs.repartition(npartitions=npartitions) 1402 if how == "left" and lhs.npartitions < rhs.npartitions: 1422 if lhs.npartitions < rhs.npartitions: 1478 if lhs.npartitions < rhs.npartitions: 1479 npartitions = rhs.npartitions [all …]
|
H A D | partitionquantiles.py | 296 def process_val_weights(vals_and_weights, npartitions, dtype_info): argument 335 if len(vals) == npartitions + 1: 337 elif len(vals) < npartitions + 1: 347 0, len(vals) - 1, npartitions - len(vals) + 1, dtype=int 353 target_weight = weights.sum() / npartitions 359 trimmed_npartitions = npartitions - len(jumbo_vals) 450 qs = np.linspace(0, 1, npartitions + 1) 454 state_data = random_state_data(df.npartitions, random_state) 466 df.npartitions, 467 npartitions, [all …]
|
/dports/devel/py-dask/dask-2021.11.2/dask/bag/ |
H A D | core.py | 463 self.npartitions = npartitions 1417 npartitions = self.npartitions 1418 if npartitions > self.npartitions: 1969 npartitions = bags[0].npartitions 2154 npartitions = {b.npartitions for b in bags} 2157 npartitions = npartitions.pop() 2257 npartitions = npartitions.pop() 2398 npartitions = b.npartitions 2542 if npartitions == bag.npartitions: 2546 if bag.npartitions > npartitions: [all …]
|
/dports/math/py-sympy/sympy-1.9/sympy/ntheory/tests/ |
H A D | test_partitions.py | 1 from sympy.ntheory import npartitions 5 assert [npartitions(k) for k in range(13)] == \ 7 assert npartitions(100) == 190569292 8 assert npartitions(200) == 3972999029388 9 assert npartitions(1000) == 24061467864032622473692149727991 10 assert npartitions(2000) == 4720819175619413888601432406799959512200344166 11 assert npartitions(10000) % 10**10 == 6916435144 12 assert npartitions(100000) % 10**10 == 9421098519
|
/dports/devel/py-dask/dask-2021.11.2/dask/dataframe/io/tests/ |
H A D | test_sql.py | 89 npartitions=2, 94 assert dask_df.npartitions == 2 104 ddf = dd.from_pandas(df, npartitions=2) 233 assert data.npartitions == 1 244 assert data_1.npartitions == 1 255 assert data.npartitions == 2 275 npartitions=3, 301 npartitions=2, 369 npartitions=2, 443 def test_to_sql(npartitions, parallel): argument [all …]
|
H A D | test_parquet.py | 86 npartitions = 15 variable 95 ddf = dd.from_pandas(df, npartitions=npartitions) 2517 npartitions = 2 2520 expect = dd.from_pandas(df, npartitions=npartitions) 2605 npartitions = 4 2615 dd.from_pandas(df.iloc[:half], npartitions=npartitions).to_parquet( 2618 dd.from_pandas(df.iloc[half:], npartitions=npartitions).to_parquet( 2773 assert ddf2.npartitions < ddf1.npartitions 3138 npartitions = 4 3139 b = np.arange(npartitions).repeat(size // npartitions) [all …]
|
H A D | test_orc.py | 55 assert d.npartitions == 8 / split_stripes 65 keys = [(d3._name, i) for i in range(d3.npartitions)] 124 df = dd.from_pandas(data, npartitions=8) 131 assert df2.npartitions == df.npartitions / int(split_stripes) 133 assert df2.npartitions == df.npartitions 143 assert df2.npartitions == 2 165 ddf = dd.from_pandas(df, npartitions=4)
|
/dports/devel/py-dask/dask-2021.11.2/dask/dataframe/tests/ |
H A D | test_shuffle.py | 54 assert s.npartitions == d.npartitions 66 assert shuffle(d, d.b).npartitions == d.npartitions 74 assert s.npartitions == 17 87 assert s.npartitions == 5 187 ddf = dd.from_pandas(df, npartitions=npartitions) 261 "x", shuffle=shuffle_method, max_branch=2, npartitions=ddf.npartitions 265 assert ddf2.npartitions == ddf.npartitions 287 assert result.npartitions == ddf.npartitions 500 assert ddf2.npartitions <= ddf.npartitions 1010 ddf.set_index("x", npartitions=npartitions) [all …]
|
H A D | test_multi.py | 242 assert c.npartitions == 3 1668 assert result.npartitions == case[0].npartitions + case[1].npartitions 1683 assert result.npartitions == case[0].npartitions + case[1].npartitions 1715 assert result.npartitions == ddf1.npartitions + ddf2.npartitions 1734 assert result.npartitions == ( 1735 ddf1.npartitions + ddf2.npartitions + ddf3.npartitions 2182 assert ddf1.npartitions == 2 2231 npartitions = 3 2428 npartitions=npartitions, 2438 if npartitions: [all …]
|
H A D | test_dataframe.py | 132 d.head(2, npartitions=5) 148 d.head(7, npartitions=2) 289 ddf = dd.from_pandas(pdf, npartitions=npartitions) 1593 npartitions=2, 2063 assert b.npartitions == k 2090 assert b.npartitions == 1 2115 assert a.npartitions == 15 2124 assert b.npartitions == 2 2147 ddf = dd.from_pandas(df, npartitions=npartitions, name="x") 3498 npartitions=4, [all …]
|
H A D | test_categorical.py | 23 ds = dd.from_pandas(s, npartitions=2) 26 ds = dd.from_pandas(s, npartitions=2) 268 a = dd.from_pandas(df, npartitions=2) 271 b = a.set_index("y", npartitions=a.npartitions) 276 b = a.set_index(a.y, npartitions=a.npartitions) 281 b = a.set_index("y", divisions=["a", "b", "c"], npartitions=a.npartitions) 298 ddf = dd.from_pandas(pdf, npartitions=npartitions) 305 def test_repartition_on_categoricals(npartitions): argument 307 ddf = dd.from_pandas(df, npartitions=2) 309 ddf2 = ddf.repartition(npartitions=npartitions) [all …]
|
H A D | test_hyperloglog.py | 66 def test_basic(df, npartitions): argument 67 ddf = dd.from_pandas(df, npartitions=npartitions) 76 def test_split_every(split_every, npartitions): argument 78 ddf = dd.from_pandas(df, npartitions=npartitions)
|
H A D | test_groupby.py | 130 ddf = dd.from_pandas(df, npartitions=3) 151 ddf = dd.from_pandas(df, npartitions=3) 241 ddf = dd.from_pandas(df, npartitions=3) 257 ddf = dd.from_pandas(df, npartitions=2) 412 s = dd.from_pandas(ps, npartitions=3) 421 s = dd.from_pandas(ps, npartitions=3) 440 ss = dd.from_pandas(s, npartitions=2) 460 ss = dd.from_pandas(s, npartitions=2) 1478 a = dd.from_pandas(d, npartitions=2) 2272 ddf = data_frame(df, npartitions=1) [all …]
|
H A D | test_rolling.py | 47 def test_map_overlap(npartitions): argument 48 ddf = dd.from_pandas(df, npartitions) 62 npartitions = 3 63 ddf = dd.from_pandas(df, npartitions) 73 assert len(diff) == npartitions 101 ddf = dd.from_pandas(df, npartitions=2) 207 a = dd.from_pandas(df, npartitions=2) 213 ddf = dd.from_pandas(df, npartitions=3) 233 ddf = dd.from_pandas(df, npartitions=5) 358 ddf = dd.from_pandas(df, npartitions=3) [all …]
|
/dports/devel/py-dask/dask-2021.11.2/dask/bag/tests/ |
H A D | test_bag.py | 373 b = db.range(15, npartitions=npartitions) 383 c = db.from_sequence(data, npartitions=npartitions) 739 assert b2.npartitions == b.npartitions ** 2 771 assert c.npartitions == b.npartitions 822 assert c.npartitions == d.npartitions 1166 b = db.range(100, npartitions=npartitions) 1168 assert b.npartitions == npartitions 1174 evens = db.from_sequence(range(0, hi, 2), npartitions=npartitions) 1175 odds = db.from_sequence(range(1, hi, 2), npartitions=npartitions) 1177 assert pairs.npartitions == npartitions [all …]
|
H A D | test_avro.py | 32 assert b.npartitions == 1 45 assert b.npartitions == 2 58 assert b.npartitions == 2 62 assert b.npartitions > 2 71 b = db.from_sequence([{"a": i} for i in [1, 2, 3, 4, 5]], npartitions=2) 91 b = db.from_sequence(expected, npartitions=3) 99 b = db.from_sequence(expected, npartitions=3)
|
/dports/devel/py-dask/dask-2021.11.2/dask/dataframe/tseries/tests/ |
H A D | test_resample.py | 38 ds = dd.from_pandas(ps, npartitions=npartitions) 57 ds = dd.from_pandas(ps, npartitions=2) 67 ds = dd.from_pandas(ps, npartitions=2) 79 ds = dd.from_pandas(ps, npartitions=2) 92 ds = dd.from_pandas(ps, npartitions=5) 140 ddf = dd.from_pandas(df, npartitions=5) 154 ds = dd.from_pandas(s, npartitions=5) 164 ddf = dd.from_pandas(df, npartitions=2, sort=False) 184 ddf = dd.from_pandas(df, npartitions=4) 197 ddf = dd.from_pandas(df, npartitions=1) [all …]
|
/dports/devel/py-dask/dask-2021.11.2/docs/source/ |
H A D | dataframe-indexing.rst | 33 >>> ddf = dd.from_pandas(df, npartitions=2) 37 npartitions=1 49 npartitions=1 61 npartitions=1 74 npartitions=1 82 npartitions=1 90 npartitions=1 103 npartitions=11 114 npartitions=1 133 npartitions=1
|
/dports/devel/py-dask/dask-2021.11.2/dask/ |
H A D | datasets.py | 81 def _make_mimesis(field, schema, npartitions, records_per_partition, seed=None): argument 110 seeds = [random_state.randint(0, 1 << 32) for _ in range(npartitions)] 113 field, schema, npartitions, records_per_partition, seed 120 return db.Bag(dsk, name, npartitions) 123 def make_people(npartitions=10, records_per_partition=1000, seed=None, locale="en"): argument 163 {"locale": locale}, schema, npartitions, records_per_partition, seed
|
H A D | layers.py | 397 npartitions, argument 408 self.npartitions = npartitions 448 self.name, self.npartitions 505 self.npartitions, 624 self.npartitions, 676 npartitions, argument 691 npartitions, 772 self.npartitions, 885 npartitions, argument 896 self.npartitions = npartitions [all …]
|
/dports/devel/py-dask/dask-2021.11.2/dask/dataframe/io/ |
H A D | sql.py | 17 npartitions=None, argument 120 if divisions and npartitions: 147 return from_pandas(head, npartitions=1) 155 if divisions is None and npartitions is None: 173 if npartitions is None: 176 npartitions = ( 189 freq="%iS" % ((maxi - mini).total_seconds() / npartitions), 195 divisions = np.linspace(mini, maxi, npartitions + 1).tolist()
|
/dports/devel/py-dask/dask-2021.11.2/dask/tests/ |
H A D | test_layers.py | 41 return dd.from_pandas(df, npartitions=2).shuffle("a", shuffle="tasks") 50 ddf1 = dd.from_pandas(df, npartitions=4) 51 ddf2 = dd.from_pandas(df, npartitions=1) 98 dd.from_pandas(pd.DataFrame({"a": range(10)}), npartitions=2,).to_parquet( 131 dd.from_pandas(pd.DataFrame({"a": range(10)}), npartitions=2,).to_parquet( 144 npartitions=2,
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/block/partitions/ |
H A D | osf.c | 20 unsigned int npartitions; in osf_partition() local 70 npartitions = le16_to_cpu(label->d_npartitions); in osf_partition() 71 if (npartitions > MAX_OSF_PARTITIONS) { in osf_partition() 75 for (i = 0 ; i < npartitions; i++, partition++) { in osf_partition()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/block/partitions/ |
H A D | osf.c | 20 unsigned int npartitions; in osf_partition() local 70 npartitions = le16_to_cpu(label->d_npartitions); in osf_partition() 71 if (npartitions > MAX_OSF_PARTITIONS) { in osf_partition() 75 for (i = 0 ; i < npartitions; i++, partition++) { in osf_partition()
|