/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/ |
H A D | TestJsonUtil.java | 41 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 131 DatanodeInfo di = JsonUtil.toDatanodeInfo(response); in testToDatanodeInfoWithName()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/ |
H A D | BlockReaderFactory.java | 37 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 127 private DatanodeInfo datanode; 222 public BlockReaderFactory setDatanodeInfo(DatanodeInfo datanode) { in setDatanodeInfo()
|
H A D | DFSClient.java | 157 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 1321 Map<DatanodeInfo, List<LocatedBlock>> datanodeBlocks = in getBlockStorageLocations() 1322 new LinkedHashMap<DatanodeInfo, List<LocatedBlock>>(); in getBlockStorageLocations() 1324 for (DatanodeInfo info : b.getLocations()) { in getBlockStorageLocations() 1336 Map<DatanodeInfo, HdfsBlocksMetadata> metadatas; in getBlockStorageLocations() 2237 final DatanodeInfo[] datanodes = lb.getLocations(); in getFileChecksum() 2372 private IOStreamPair connectToDN(DatanodeInfo dn, int timeout, in connectToDN() 2409 private Type inferChecksumTypeByReading(LocatedBlock lb, DatanodeInfo dn) in inferChecksumTypeByReading() 2549 public DatanodeInfo[] datanodeReport(DatanodeReportType type) in datanodeReport() 3130 void reportChecksumFailure(String file, ExtendedBlock blk, DatanodeInfo dn) { in reportChecksumFailure() [all …]
|
/dports/databases/hbase/hbase-1.2.1/hbase-server/src/test/java/org/apache/hadoop/hbase/fs/ |
H A D | TestBlockReorder.java | 59 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 178 DatanodeInfo[] infos = lb.getLocations(); in testBlockLocationReorder() 181 DatanodeInfo tmp = infos[0]; in testBlockLocationReorder()
|
/dports/devel/hadoop/hadoop-1.2.1/src/test/org/apache/hadoop/hdfs/server/namenode/ |
H A D | NNThroughputBenchmark.java | 36 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 815 DatanodeInfo xferTargets[][] in transferBlocks() 818 DatanodeInfo blockTargets[] = xferTargets[i]; in transferBlocks() 820 DatanodeInfo dnInfo = blockTargets[t]; in transferBlocks() 933 for(DatanodeInfo dnInfo : loc.getLocations()) { in addBlocks()
|
H A D | TestStartup.java | 40 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 344 client.namenode.addBlock("/tmp1.txt", "clientName", new DatanodeInfo[0]); in testGetBlocks()
|
/dports/devel/hadoop/hadoop-1.2.1/src/test/org/apache/hadoop/hdfs/ |
H A D | TestDatanodeDeath.java | 31 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 372 DatanodeInfo[] targets = dfstream.getPipeline(); in simpleTest()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/net/ |
H A D | TestNetworkTopology.java | 35 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 329 DatanodeInfo[] info; in testInvalidNetworkTopologiesNotCachedInHdfs()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ |
H A D | TestNamenodeCapacityReport.java | 41 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 257 for (DatanodeInfo dni : streams[i].getPipeline()) { in testXceiverCount()
|
/dports/devel/hadoop/hadoop-1.2.1/src/hdfs/org/apache/hadoop/hdfs/server/namenode/ |
H A D | DatanodeDescriptor.java | 27 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 47 public class DatanodeDescriptor extends DatanodeInfo {
|
H A D | FSNamesystem.java | 75 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 2155 void addToInvalidates(Block b, DatanodeInfo dn, boolean log) { in addToInvalidates() 2172 void addToInvalidates(Block b, DatanodeInfo dn) { in addToInvalidates() 2182 void addToInvalidatesNoLog(Block b, DatanodeInfo n) { in addToInvalidatesNoLog() 4353 DatanodeInfo cur = null; in chooseExcessReplicates() 4636 public DatanodeInfo[] datanodeReport(DatanodeReportType type in datanodeReport() 4642 DatanodeInfo[] arr = new DatanodeInfo[results.size()]; in datanodeReport() 4644 arr[i] = new DatanodeInfo(results.get(i)); in datanodeReport() 4727 public DatanodeInfo getDataNodeInfo(String name) { in getDataNodeInfo() 4937 ((node instanceof DatanodeInfo) && in inHostsList() [all …]
|
/dports/devel/hadoop/hadoop-1.2.1/src/test/org/apache/hadoop/hdfs/security/token/block/ |
H A D | TestBlockToken.java | 35 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 202 any(Block.class), anyBoolean(), any(DatanodeInfo[].class)); in testBlockTokenRpc()
|
/dports/devel/hadoop/hadoop-1.2.1/src/webapps/datanode/ |
H A D | browseDirectory.jsp | 75 DatanodeInfo [] locations = null; 83 DatanodeInfo chosenNode = jspHelper.bestNode(firstBlock);
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/ |
H A D | TestDatanodeDeath.java | 32 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 373 DatanodeInfo[] targets = dfstream.getPipeline(); in simpleTest()
|
H A D | TestFileAppend3.java | 42 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 314 DatanodeInfo[] datanodeinfos = lb.getLocations(); in testTC7() 389 for(DatanodeInfo datanodeinfo : lb.getLocations()) { in testTC11()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/ |
H A D | BlockReceiver.java | 43 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 101 private DatanodeInfo srcDataNode = null; 146 final String clientname, final DatanodeInfo srcDataNode, in BlockReceiver() 876 DatanodeInfo[] downstreams, in receiveBlock() 1134 final DataInputStream downstreamIn, final DatanodeInfo[] downstreams) { in PacketResponder()
|
H A D | DataNode.java | 121 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 1000 public void reportRemoteBadBlock(DatanodeInfo srcDataNode, ExtendedBlock block) in reportRemoteBadBlock() 1863 private void transferBlock(ExtendedBlock block, DatanodeInfo[] xferTargets, 1927 DatanodeInfo xferTargets[][], StorageType[][] xferTargetStorageTypes) { 2033 final DatanodeInfo[] targets; 2045 DataTransfer(DatanodeInfo targets[], StorageType[] targetStorageTypes, 2115 DatanodeInfo srcNode = new DatanodeInfo(bpReg); 2787 DatanodeInfo[] targets = rb.getLocations(); 2838 final DatanodeInfo[] targets, final StorageType[] targetStorageTypes,
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/ |
H A D | TestPendingReplication.java | 33 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 284 DatanodeInfo existingDn = (locatedBlock.getLocations())[0]; in testBlockReceived()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/ |
H A D | LocatedBlocks.java | 122 new ExtendedBlock(), new DatanodeInfo[0], 0L, false); in findBlock()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/web/resources/ |
H A D | NamenodeWebHdfsMethods.java | 62 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 164 static DatanodeInfo chooseDatanode(final NameNode namenode, in chooseDatanode() 230 private static DatanodeInfo bestNode(DatanodeInfo[] nodes, in bestNode() 232 for (DatanodeInfo dn: nodes) { in bestNode() 261 final DatanodeInfo dn; in redirectURI()
|
/dports/devel/hadoop/hadoop-1.2.1/src/hdfs/org/apache/hadoop/hdfs/server/datanode/ |
H A D | DataBlockScanner.java | 50 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 375 DatanodeInfo[] dnArr = { new DatanodeInfo(datanode.dnRegistration) }; in handleScanFailure()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/balancer/ |
H A D | TestBalancerWithMultipleNameNodes.java | 41 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 169 final DatanodeInfo[] datanodes = s.clients[n].getDatanodeReport( in runBalancer()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/ |
H A D | TestBlockHasMultipleReplicasOnSameDN.java | 134 DatanodeInfo[] locations = locatedBlock.getLocations(); in testBlockHasMultipleReplicasOnSameDN()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/mover/ |
H A D | Mover.java | 474 final DatanodeInfo datanode; 478 MLocation(DatanodeInfo datanode, StorageType storageType, long size) { in MLocation() 485 final DatanodeInfo[] datanodeInfos = lb.getLocations(); in toLocations()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/ |
H A D | BlockPlacementPolicyDefault.java | 31 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 865 DatanodeInfo[] locs = lBlk.getLocations(); in verifyBlockPlacement() 876 for (DatanodeInfo dn : locs) in verifyBlockPlacement()
|