Home
last modified time | relevance | path

Searched refs:DatanodeInfo (Results 151 – 175 of 186) sorted by relevance

12345678

/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/
H A DBPServiceActor.java39 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
873 void reportRemoteBadBlock(DatanodeInfo dnInfo, ExtendedBlock block) in reportRemoteBadBlock()
876 new DatanodeInfo[] {dnInfo}); in reportRemoteBadBlock()
H A DDataXceiver.java54 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
609 final DatanodeInfo[] targets, in writeBlock()
611 final DatanodeInfo srcDataNode, in writeBlock()
856 final DatanodeInfo[] targets, in transferBlock()
1051 final DatanodeInfo proxySource) throws IOException {
H A DBPOfferService.java30 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
439 void reportRemoteBadBlock(DatanodeInfo dnInfo, ExtendedBlock block) { in reportRemoteBadBlock()
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/
H A DNNThroughputBenchmark.java42 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
1026 DatanodeInfo xferTargets[][], in transferBlocks()
1030 DatanodeInfo blockTargets[] = xferTargets[i]; in transferBlocks()
1032 DatanodeInfo dnInfo = blockTargets[t]; in transferBlocks()
1159 for(DatanodeInfo dnInfo : loc.getLocations()) { in addBlocks()
H A DTestStartup.java52 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
601 DatanodeInfo[] info = nn.getDatanodeReport(DatanodeReportType.LIVE); in testNNRestart()
H A DTestCacheDirectives.java72 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
894 DatanodeInfo[] live = dfs.getDataNodeStats(DatanodeReportType.LIVE);
897 for (DatanodeInfo dn : live) {
/dports/devel/hadoop/hadoop-1.2.1/src/hdfs/org/apache/hadoop/hdfs/server/namenode/web/resources/
H A DNamenodeWebHdfsMethods.java51 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
136 static DatanodeInfo chooseDatanode(final NameNode namenode, in chooseDatanode()
143 final DatanodeInfo dn = ns.chooseDatanode(path, getRemoteAddress(), blocksize); in chooseDatanode()
194 final DatanodeInfo dn = chooseDatanode(namenode, path, op, openOffset, in redirectURI()
/dports/devel/hadoop/hadoop-1.2.1/src/test/org/apache/hadoop/hdfs/server/namenode/
H A DTestBlockTokenWithDFS.java36 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
128 DatanodeInfo[] nodes = lblock.getLocations(); in tryRead()
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/
H A DNameNodeRpcServer.java85 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
475 public BlocksWithLocations getBlocks(DatanodeInfo datanode, long size) in getBlocks()
707 ExtendedBlock previous, DatanodeInfo[] excludedNodes, long fileId, in addBlock()
734 final DatanodeInfo[] existings, final String[] existingStorageIDs, in getAdditionalDatanode()
735 final DatanodeInfo[] excludes, in getAdditionalDatanode()
1035 public DatanodeInfo[] getDatanodeReport(DatanodeReportType type) in getDatanodeReport()
1038 DatanodeInfo results[] = namesystem.datanodeReport(type); in getDatanodeReport()
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/security/token/block/
H A DTestBlockToken.java52 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
297 LocatedBlock fakeBlock = new LocatedBlock(b, new DatanodeInfo[0]); in testBlockTokenRpcLeak()
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/
H A DTestBlockTokenWithDFS.java49 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
145 DatanodeInfo[] nodes = lblock.getLocations(); in tryRead()
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/
H A DDatanodeDescriptor.java43 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
64 public class DatanodeDescriptor extends DatanodeInfo {
H A DBlockManager.java52 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
1081 void addToInvalidates(final Block block, final DatanodeInfo datanode) { in addToInvalidates()
1112 void removeFromInvalidates(final DatanodeInfo datanode) { in removeFromInvalidates()
1128 final DatanodeInfo dn, String storageID, String reason) throws IOException { in findAndMarkBlockAsCorrupt()
1211 private boolean invalidateBlock(BlockToMarkCorrupt b, DatanodeInfo dn in invalidateBlock()
1273 final List<DatanodeInfo> nodes = invalidateBlocks.getDatanodes(); in computeInvalidateWork()
1279 for (DatanodeInfo dnInfo : nodes) { in computeInvalidateWork()
3008 private void addToExcessReplicate(DatanodeInfo dn, Block block) { in addToExcessReplicate()
3453 private int invalidateWorkForOneNode(DatanodeInfo dn) { in invalidateWorkForOneNode()
/dports/devel/hadoop/hadoop-1.2.1/src/test/org/apache/hadoop/hdfs/
H A DTestFileCreation.java39 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
492 DatanodeInfo[] info = client.datanodeReport( in testFileCreationError1()
898 for(DatanodeInfo datanodeinfo: locatedblock.getLocations()) { in testLeaseExpireHardLimit()
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/
H A DTestRetryCacheWithHA.java66 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
700 private DatanodeInfo[] nodes;
735 DatanodeInfo[] newNodes = new DatanodeInfo[2]; in invoke()
H A DTestHASafeMode.java53 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
833 new DatanodeInfo[0], in testOpenFileWhenNNAndClientCrashAfterAddBlock()
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/
H A DTestDataTransferProtocol.java45 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
568 new DatanodeInfo[1], new StorageType[1], null, stage, in writeBlock()
H A DTestFileCreation.java68 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
465 DatanodeInfo[] info = client.datanodeReport( in testFileCreationError1()
992 for(DatanodeInfo datanodeinfo: locatedblock.getLocations()) { in testLeaseExpireHardLimit()
/dports/databases/hbase/hbase-1.2.1/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/
H A DTestLogRolling.java68 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
357 final DatanodeInfo[] pipeline = log.getPipeLine(); in testLogRollOnDatanodeDeath()
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocolPB/
H A DClientNamenodeProtocolTranslatorPB.java59 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
403 ExtendedBlock previous, DatanodeInfo[] excludeNodes, long fileId, in addBlock()
426 ExtendedBlock blk, DatanodeInfo[] existings, String[] existingStorageIDs, in getAdditionalDatanode()
427 DatanodeInfo[] excludes, in getAdditionalDatanode()
620 public DatanodeInfo[] getDatanodeReport(DatanodeReportType type) in getDatanodeReport()
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/shortcircuit/
H A DTestShortCircuitLocalRead.java51 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
366 final DatanodeInfo dnInfo = lb.get(0).getLocations()[0];
/dports/devel/hadoop/hadoop-1.2.1/src/webapps/datanode/
H A Dtail.jsp110 DatanodeInfo chosenNode;
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/shortcircuit/
H A DShortCircuitCache.java44 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
1011 public Slot allocShmSlot(DatanodeInfo datanode, in allocShmSlot()
/dports/databases/hbase/hbase-1.2.1/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/
H A DFSHLog.java84 import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
1979 DatanodeInfo[] getPipeLine() {
1985 return new DatanodeInfo[0];
/dports/devel/hadoop/hadoop-1.2.1/src/hdfs/org/apache/hadoop/hdfs/tools/offlineImageViewer/
H A DImageLoaderCurrent.java28 import org.apache.hadoop.hdfs.protocol.DatanodeInfo.AdminStates;

12345678