Home
last modified time | relevance | path

Searched refs:ShuffledRDD (Results 1 – 9 of 9) sorted by relevance

/dports/devel/spark/spark-2.1.1/core/src/main/scala/org/apache/spark/rdd/
H A DShuffledRDD.scala45 class ShuffledRDD[K: ClassTag, V: ClassTag, C: ClassTag]( class
59 def setSerializer(serializer: Serializer): ShuffledRDD[K, V, C] = {
65 def setKeyOrdering(keyOrdering: Ordering[K]): ShuffledRDD[K, V, C] = {
71 def setAggregator(aggregator: Aggregator[K, V, C]): ShuffledRDD[K, V, C] = {
77 def setMapSideCombine(mapSideCombine: Boolean): ShuffledRDD[K, V, C] = {
H A DOrderedRDDFunctions.scala63 new ShuffledRDD[K, V, V](self, part)
75 new ShuffledRDD[K, V, V](self, partitioner).setKeyOrdering(ordering)
H A DPairRDDFunctions.scala102 new ShuffledRDD[K, V, C](self, partitioner)
541 new ShuffledRDD[K, V, V](self, partitioner)
H A DRDD.scala462 new ShuffledRDD[Int, T, T](mapPartitionsWithIndex(distributePartition),
/dports/devel/spark/spark-2.1.1/core/src/test/scala/org/apache/spark/
H A DSortShuffleSuite.scala28 import org.apache.spark.rdd.ShuffledRDD
62 val shuffledRdd = new ShuffledRDD[Int, Int, Int](rdd, new HashPartitioner(4))
73 val shuffledRdd = new ShuffledRDD[Int, Int, Int](rdd, new HashPartitioner(4))
80 private def ensureFilesAreCleanedUp(shuffledRdd: ShuffledRDD[_, _, _]): Unit = {
H A DShuffleSuite.scala27 import org.apache.spark.rdd.{CoGroupedRDD, OrderedRDDFunctions, RDD, ShuffledRDD, SubtractedRDD}
64 val c = new ShuffledRDD[Int,
88 val c = new ShuffledRDD[Int,
106 val c = new ShuffledRDD[Int, Int, Int](b, new HashPartitioner(NUM_BLOCKS))
132 val c = new ShuffledRDD[Int, Int, Int](b, new HashPartitioner(NUM_BLOCKS))
153 val results = new ShuffledRDD[Int, Int, Int](pairs,
H A DCheckpointSuite.scala370 new ShuffledRDD[Int, Int, Int](rdd.map(x => (x % 2, 1)), partitioner)
/dports/devel/spark/spark-2.1.1/core/src/test/scala/org/apache/spark/rdd/
H A DRDDSuite.scala371 asInstanceOf[ShuffledRDD[_, _, _]] != null
896 assert(ancestors4.count(_.isInstanceOf[ShuffledRDD[_, _, _]]) === 0)
898 assert(ancestors5.count(_.isInstanceOf[ShuffledRDD[_, _, _]]) === 1)
/dports/devel/spark/spark-2.1.1/docs/
H A Dquick-start.md126 wordCounts: org.apache.spark.rdd.RDD[(String, Int)] = ShuffledRDD[8] at reduceByKey at <console>:28