/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/ |
H A D | BPServiceActor.java | 39 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 873 void reportRemoteBadBlock(DatanodeInfo dnInfo, ExtendedBlock block) in reportRemoteBadBlock() 876 new DatanodeInfo[] {dnInfo}); in reportRemoteBadBlock()
|
H A D | DataXceiver.java | 54 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 609 final DatanodeInfo[] targets, in writeBlock() 611 final DatanodeInfo srcDataNode, in writeBlock() 856 final DatanodeInfo[] targets, in transferBlock() 1051 final DatanodeInfo proxySource) throws IOException {
|
H A D | BPOfferService.java | 30 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 439 void reportRemoteBadBlock(DatanodeInfo dnInfo, ExtendedBlock block) { in reportRemoteBadBlock()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ |
H A D | NNThroughputBenchmark.java | 42 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 1026 DatanodeInfo xferTargets[][], in transferBlocks() 1030 DatanodeInfo blockTargets[] = xferTargets[i]; in transferBlocks() 1032 DatanodeInfo dnInfo = blockTargets[t]; in transferBlocks() 1159 for(DatanodeInfo dnInfo : loc.getLocations()) { in addBlocks()
|
H A D | TestStartup.java | 52 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 601 DatanodeInfo[] info = nn.getDatanodeReport(DatanodeReportType.LIVE); in testNNRestart()
|
H A D | TestCacheDirectives.java | 72 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 894 DatanodeInfo[] live = dfs.getDataNodeStats(DatanodeReportType.LIVE); 897 for (DatanodeInfo dn : live) {
|
/dports/devel/hadoop/hadoop-1.2.1/src/hdfs/org/apache/hadoop/hdfs/server/namenode/web/resources/ |
H A D | NamenodeWebHdfsMethods.java | 51 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 136 static DatanodeInfo chooseDatanode(final NameNode namenode, in chooseDatanode() 143 final DatanodeInfo dn = ns.chooseDatanode(path, getRemoteAddress(), blocksize); in chooseDatanode() 194 final DatanodeInfo dn = chooseDatanode(namenode, path, op, openOffset, in redirectURI()
|
/dports/devel/hadoop/hadoop-1.2.1/src/test/org/apache/hadoop/hdfs/server/namenode/ |
H A D | TestBlockTokenWithDFS.java | 36 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 128 DatanodeInfo[] nodes = lblock.getLocations(); in tryRead()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/ |
H A D | NameNodeRpcServer.java | 85 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 475 public BlocksWithLocations getBlocks(DatanodeInfo datanode, long size) in getBlocks() 707 ExtendedBlock previous, DatanodeInfo[] excludedNodes, long fileId, in addBlock() 734 final DatanodeInfo[] existings, final String[] existingStorageIDs, in getAdditionalDatanode() 735 final DatanodeInfo[] excludes, in getAdditionalDatanode() 1035 public DatanodeInfo[] getDatanodeReport(DatanodeReportType type) in getDatanodeReport() 1038 DatanodeInfo results[] = namesystem.datanodeReport(type); in getDatanodeReport()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/security/token/block/ |
H A D | TestBlockToken.java | 52 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 297 LocatedBlock fakeBlock = new LocatedBlock(b, new DatanodeInfo[0]); in testBlockTokenRpcLeak()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/ |
H A D | TestBlockTokenWithDFS.java | 49 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 145 DatanodeInfo[] nodes = lblock.getLocations(); in tryRead()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/ |
H A D | DatanodeDescriptor.java | 43 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 64 public class DatanodeDescriptor extends DatanodeInfo {
|
H A D | BlockManager.java | 52 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 1081 void addToInvalidates(final Block block, final DatanodeInfo datanode) { in addToInvalidates() 1112 void removeFromInvalidates(final DatanodeInfo datanode) { in removeFromInvalidates() 1128 final DatanodeInfo dn, String storageID, String reason) throws IOException { in findAndMarkBlockAsCorrupt() 1211 private boolean invalidateBlock(BlockToMarkCorrupt b, DatanodeInfo dn in invalidateBlock() 1273 final List<DatanodeInfo> nodes = invalidateBlocks.getDatanodes(); in computeInvalidateWork() 1279 for (DatanodeInfo dnInfo : nodes) { in computeInvalidateWork() 3008 private void addToExcessReplicate(DatanodeInfo dn, Block block) { in addToExcessReplicate() 3453 private int invalidateWorkForOneNode(DatanodeInfo dn) { in invalidateWorkForOneNode()
|
/dports/devel/hadoop/hadoop-1.2.1/src/test/org/apache/hadoop/hdfs/ |
H A D | TestFileCreation.java | 39 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 492 DatanodeInfo[] info = client.datanodeReport( in testFileCreationError1() 898 for(DatanodeInfo datanodeinfo: locatedblock.getLocations()) { in testLeaseExpireHardLimit()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/ |
H A D | TestRetryCacheWithHA.java | 66 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 700 private DatanodeInfo[] nodes; 735 DatanodeInfo[] newNodes = new DatanodeInfo[2]; in invoke()
|
H A D | TestHASafeMode.java | 53 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 833 new DatanodeInfo[0], in testOpenFileWhenNNAndClientCrashAfterAddBlock()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/ |
H A D | TestDataTransferProtocol.java | 45 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 568 new DatanodeInfo[1], new StorageType[1], null, stage, in writeBlock()
|
H A D | TestFileCreation.java | 68 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 465 DatanodeInfo[] info = client.datanodeReport( in testFileCreationError1() 992 for(DatanodeInfo datanodeinfo: locatedblock.getLocations()) { in testLeaseExpireHardLimit()
|
/dports/databases/hbase/hbase-1.2.1/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/ |
H A D | TestLogRolling.java | 68 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 357 final DatanodeInfo[] pipeline = log.getPipeLine(); in testLogRollOnDatanodeDeath()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocolPB/ |
H A D | ClientNamenodeProtocolTranslatorPB.java | 59 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 403 ExtendedBlock previous, DatanodeInfo[] excludeNodes, long fileId, in addBlock() 426 ExtendedBlock blk, DatanodeInfo[] existings, String[] existingStorageIDs, in getAdditionalDatanode() 427 DatanodeInfo[] excludes, in getAdditionalDatanode() 620 public DatanodeInfo[] getDatanodeReport(DatanodeReportType type) in getDatanodeReport()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/shortcircuit/ |
H A D | TestShortCircuitLocalRead.java | 51 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 366 final DatanodeInfo dnInfo = lb.get(0).getLocations()[0];
|
/dports/devel/hadoop/hadoop-1.2.1/src/webapps/datanode/ |
H A D | tail.jsp | 110 DatanodeInfo chosenNode;
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/shortcircuit/ |
H A D | ShortCircuitCache.java | 44 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 1011 public Slot allocShmSlot(DatanodeInfo datanode, in allocShmSlot()
|
/dports/databases/hbase/hbase-1.2.1/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/ |
H A D | FSHLog.java | 84 import org.apache.hadoop.hdfs.protocol.DatanodeInfo; 1979 DatanodeInfo[] getPipeLine() { 1985 return new DatanodeInfo[0];
|
/dports/devel/hadoop/hadoop-1.2.1/src/hdfs/org/apache/hadoop/hdfs/tools/offlineImageViewer/ |
H A D | ImageLoaderCurrent.java | 28 import org.apache.hadoop.hdfs.protocol.DatanodeInfo.AdminStates;
|