Searched refs:ShuffledRDD (Results 1 – 9 of 9) sorted by relevance
/dports/devel/spark/spark-2.1.1/core/src/main/scala/org/apache/spark/rdd/ |
H A D | ShuffledRDD.scala | 45 class ShuffledRDD[K: ClassTag, V: ClassTag, C: ClassTag]( class 59 def setSerializer(serializer: Serializer): ShuffledRDD[K, V, C] = { 65 def setKeyOrdering(keyOrdering: Ordering[K]): ShuffledRDD[K, V, C] = { 71 def setAggregator(aggregator: Aggregator[K, V, C]): ShuffledRDD[K, V, C] = { 77 def setMapSideCombine(mapSideCombine: Boolean): ShuffledRDD[K, V, C] = {
|
H A D | OrderedRDDFunctions.scala | 63 new ShuffledRDD[K, V, V](self, part) 75 new ShuffledRDD[K, V, V](self, partitioner).setKeyOrdering(ordering)
|
H A D | PairRDDFunctions.scala | 102 new ShuffledRDD[K, V, C](self, partitioner) 541 new ShuffledRDD[K, V, V](self, partitioner)
|
H A D | RDD.scala | 462 new ShuffledRDD[Int, T, T](mapPartitionsWithIndex(distributePartition),
|
/dports/devel/spark/spark-2.1.1/core/src/test/scala/org/apache/spark/ |
H A D | SortShuffleSuite.scala | 28 import org.apache.spark.rdd.ShuffledRDD 62 val shuffledRdd = new ShuffledRDD[Int, Int, Int](rdd, new HashPartitioner(4)) 73 val shuffledRdd = new ShuffledRDD[Int, Int, Int](rdd, new HashPartitioner(4)) 80 private def ensureFilesAreCleanedUp(shuffledRdd: ShuffledRDD[_, _, _]): Unit = {
|
H A D | ShuffleSuite.scala | 27 import org.apache.spark.rdd.{CoGroupedRDD, OrderedRDDFunctions, RDD, ShuffledRDD, SubtractedRDD} 64 val c = new ShuffledRDD[Int, 88 val c = new ShuffledRDD[Int, 106 val c = new ShuffledRDD[Int, Int, Int](b, new HashPartitioner(NUM_BLOCKS)) 132 val c = new ShuffledRDD[Int, Int, Int](b, new HashPartitioner(NUM_BLOCKS)) 153 val results = new ShuffledRDD[Int, Int, Int](pairs,
|
H A D | CheckpointSuite.scala | 370 new ShuffledRDD[Int, Int, Int](rdd.map(x => (x % 2, 1)), partitioner)
|
/dports/devel/spark/spark-2.1.1/core/src/test/scala/org/apache/spark/rdd/ |
H A D | RDDSuite.scala | 371 asInstanceOf[ShuffledRDD[_, _, _]] != null 896 assert(ancestors4.count(_.isInstanceOf[ShuffledRDD[_, _, _]]) === 0) 898 assert(ancestors5.count(_.isInstanceOf[ShuffledRDD[_, _, _]]) === 1)
|
/dports/devel/spark/spark-2.1.1/docs/ |
H A D | quick-start.md | 126 wordCounts: org.apache.spark.rdd.RDD[(String, Int)] = ShuffledRDD[8] at reduceByKey at <console>:28
|