/dports/devel/spark/spark-2.1.1/core/src/test/scala/org/apache/spark/storage/ |
H A D | ShuffleBlockFetcherIteratorSuite.scala | 77 ShuffleBlockId(0, 0, 0) -> createMockManagedBuffer(), 78 ShuffleBlockId(0, 1, 0) -> createMockManagedBuffer(), 79 ShuffleBlockId(0, 2, 0) -> createMockManagedBuffer()) 87 ShuffleBlockId(0, 3, 0) -> createMockManagedBuffer(), 88 ShuffleBlockId(0, 4, 0) -> createMockManagedBuffer()) 180 verify(blocks(ShuffleBlockId(0, 0, 0)), times(1)).release() 192 verify(blocks(ShuffleBlockId(0, 2, 0)), times(0)).retain() 204 ShuffleBlockId(0, 0, 0) -> mock(classOf[ManagedBuffer]), 205 ShuffleBlockId(0, 1, 0) -> mock(classOf[ManagedBuffer]), 206 ShuffleBlockId(0, 2, 0) -> mock(classOf[ManagedBuffer]) [all …]
|
H A D | BlockIdSuite.scala | 58 val id = ShuffleBlockId(1, 2, 3) 59 assertSame(id, ShuffleBlockId(1, 2, 3)) 60 assertDifferent(id, ShuffleBlockId(3, 2, 3))
|
/dports/devel/spark/spark-2.1.1/core/src/main/scala/org/apache/spark/shuffle/ |
H A D | ShuffleBlockResolver.scala | 21 import org.apache.spark.storage.ShuffleBlockId 37 def getBlockData(blockId: ShuffleBlockId): ManagedBuffer
|
H A D | IndexShuffleBlockResolver.scala | 194 override def getBlockData(blockId: ShuffleBlockId): ManagedBuffer = {
|
/dports/devel/spark/spark-2.1.1/core/src/test/scala/org/apache/spark/shuffle/ |
H A D | BlockStoreShuffleReaderSuite.scala | 28 import org.apache.spark.storage.{BlockManager, BlockManagerId, ShuffleBlockId} 96 val shuffleBlockId = ShuffleBlockId(shuffleId, mapId, reduceId) 108 val shuffleBlockId = ShuffleBlockId(shuffleId, mapId, reduceId)
|
/dports/devel/spark/spark-2.1.1/core/src/main/scala/org/apache/spark/storage/ |
H A D | BlockId.scala | 40 def isShuffle: Boolean = isInstanceOf[ShuffleBlockId] 59 case class ShuffleBlockId(shuffleId: Int, mapId: Int, reduceId: Int) extends BlockId { caseClass 119 ShuffleBlockId(shuffleId.toInt, mapId.toInt, reduceId.toInt)
|
H A D | ShuffleBlockFetcherIterator.scala | 360 case ShuffleBlockId(shufId, mapId, reduceId) =>
|
H A D | BlockManager.scala | 305 shuffleManager.shuffleBlockResolver.getBlockData(blockId.asInstanceOf[ShuffleBlockId]) 502 shuffleBlockResolver.getBlockData(blockId.asInstanceOf[ShuffleBlockId]).nioByteBuffer()))
|
/dports/devel/spark/spark-2.1.1/core/src/test/scala/org/apache/spark/ |
H A D | MapOutputTrackerSuite.scala | 29 import org.apache.spark.storage.{BlockManagerId, ShuffleBlockId} 69 Seq((BlockManagerId("a", "hostA", 1000), ArrayBuffer((ShuffleBlockId(10, 0, 0), size1000))), 70 (BlockManagerId("b", "hostB", 1000), ArrayBuffer((ShuffleBlockId(10, 1, 0), size10000)))) 151 Seq((BlockManagerId("a", "hostA", 1000), ArrayBuffer((ShuffleBlockId(10, 0, 0), size1000)))))
|
H A D | ShuffleSuite.scala | 31 import org.apache.spark.storage.{ShuffleBlockId, ShuffleDataBlockId} 278 val hashFile = sc.env.blockManager.diskBlockManager.getFile(new ShuffleBlockId(0, 0, 0))
|
H A D | ContextCleanerSuite.scala | 534 case ShuffleBlockId(`shuffleId`, _, _) => true
|
/dports/devel/spark/spark-2.1.1/core/src/main/scala/org/apache/spark/shuffle/sort/ |
H A D | SortShuffleWriter.scala | 24 import org.apache.spark.storage.ShuffleBlockId 71 val blockId = ShuffleBlockId(dep.shuffleId, mapId, IndexShuffleBlockResolver.NOOP_REDUCE_ID)
|
/dports/devel/spark/spark-2.1.1/core/src/test/scala/org/apache/spark/network/netty/ |
H A D | NettyBlockTransferSecuritySuite.scala | 38 import org.apache.spark.storage.{BlockId, ShuffleBlockId} 104 val blockId = ShuffleBlockId(0, 1, 2)
|
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/execution/ |
H A D | UnsafeRowSerializerSuite.scala | 31 import org.apache.spark.storage.ShuffleBlockId 130 sorter.writePartitionedFile(ShuffleBlockId(0, 0, 0), outputFile)
|
/dports/devel/spark/spark-2.1.1/core/src/main/scala/org/apache/spark/serializer/ |
H A D | SerializerManager.scala | 108 case _: ShuffleBlockId => compressShuffle
|
/dports/devel/spark/spark-2.1.1/core/src/main/scala/org/apache/spark/ |
H A D | MapOutputTracker.scala | 34 import org.apache.spark.storage.{BlockId, BlockManagerId, ShuffleBlockId} 701 ((ShuffleBlockId(shuffleId, mapId, part), status.getSizeForBlock(part)))
|
/dports/devel/spark/spark-2.1.1/core/src/test/scala/org/apache/spark/util/ |
H A D | JsonProtocolSuite.scala | 164 testBlockId(ShuffleBlockId(1, 2, 3))
|