/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/ |
H A D | ZombieJob.java | 422 int taskAttemptNumber) { in getTaskAttemptInfo() argument 435 taskNumber, taskAttemptNumber); in getTaskAttemptInfo() 471 int taskAttemptNumber, int locality) { in getMapTaskAttemptInfoAdjusted() argument 481 taskNumber, taskAttemptNumber); in getMapTaskAttemptInfoAdjusted() 699 int taskAttemptNumber) { in makeTaskAttemptID() argument 701 taskAttemptNumber); in makeTaskAttemptID() 705 int taskAttemptNumber, int taskNumber, int locality) { in makeUpTaskAttemptInfo() argument 906 if (taskAttemptNumber >= numAttempts.length - 1) { in makeUpState() 910 double pSucceed = numAttempts[taskAttemptNumber]; in makeUpState() 930 int taskNumber, int taskAttemptNumber) { in getLoggedTaskAttempt() argument [all …]
|
H A D | JobStory.java | 98 int taskAttemptNumber); in getTaskAttemptInfo() argument 110 int taskAttemptNumber, in getMapTaskAttemptInfoAdjusted() argument
|
/dports/devel/hadoop/hadoop-1.2.1/src/tools/org/apache/hadoop/tools/rumen/ |
H A D | ZombieJob.java | 416 int taskAttemptNumber) { in getTaskAttemptInfo() argument 429 taskNumber, taskAttemptNumber); in getTaskAttemptInfo() 465 int taskAttemptNumber, int locality) { in getMapTaskAttemptInfoAdjusted() argument 475 taskNumber, taskAttemptNumber); in getMapTaskAttemptInfoAdjusted() 685 int taskAttemptNumber) { in makeTaskAttemptID() argument 691 int taskAttemptNumber, int taskNumber, int locality) { in makeUpTaskAttemptInfo() argument 700 taskNumber, taskAttemptNumber).toString()); in makeUpTaskAttemptInfo() 879 if (taskAttemptNumber >= numAttempts.length - 1) { in makeUpState() 883 double pSucceed = numAttempts[taskAttemptNumber]; in makeUpState() 903 int taskNumber, int taskAttemptNumber) { in getLoggedTaskAttempt() argument [all …]
|
H A D | JobStory.java | 98 int taskAttemptNumber); in getTaskAttemptInfo() argument 110 int taskAttemptNumber, in getMapTaskAttemptInfoAdjusted() argument
|
/dports/devel/spark/spark-2.1.1/core/src/main/scala/org/apache/spark/mapred/ |
H A D | SparkHadoopMapRedUtil.scala | 72 val taskAttemptNumber = TaskContext.get().attemptNumber() constant 73 val canCommit = outputCommitCoordinator.canCommit(jobId, splitId, taskAttemptNumber) 83 throw new CommitDeniedException(message, jobId, splitId, taskAttemptNumber)
|
/dports/devel/hadoop/hadoop-1.2.1/src/contrib/gridmix/src/java/org/apache/hadoop/mapred/gridmix/ |
H A D | JobFactory.java | 154 int taskAttemptNumber) { in getTaskAttemptInfo() argument 155 return job.getTaskAttemptInfo(taskType, taskNumber, taskAttemptNumber); in getTaskAttemptInfo() 158 int taskNumber, int taskAttemptNumber, int locality) { in getMapTaskAttemptInfoAdjusted() argument 160 taskNumber, taskAttemptNumber, locality); in getMapTaskAttemptInfoAdjusted()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ |
H A D | JobFactory.java | 156 int taskAttemptNumber) { in getTaskAttemptInfo() argument 157 return job.getTaskAttemptInfo(taskType, taskNumber, taskAttemptNumber); in getTaskAttemptInfo() 160 int taskNumber, int taskAttemptNumber, int locality) { in getMapTaskAttemptInfoAdjusted() argument 162 taskNumber, taskAttemptNumber, locality); in getMapTaskAttemptInfoAdjusted()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-tools/hadoop-gridmix/src/test/java/org/apache/hadoop/mapred/gridmix/ |
H A D | DebugJobProducer.java | 268 TaskType taskType, int taskNumber, int taskAttemptNumber) { in getTaskAttemptInfo() argument 291 int taskNumber, int taskAttemptNumber, int locality) { in getMapTaskAttemptInfoAdjusted() argument
|
/dports/devel/hadoop/hadoop-1.2.1/src/contrib/gridmix/src/test/org/apache/hadoop/mapred/gridmix/ |
H A D | DebugJobProducer.java | 268 TaskType taskType, int taskNumber, int taskAttemptNumber) { in getTaskAttemptInfo() argument 287 int taskNumber, int taskAttemptNumber, int locality) { in getMapTaskAttemptInfoAdjusted() argument
|
/dports/devel/spark/spark-2.1.1/core/src/main/scala/org/apache/spark/util/ |
H A D | Utils.scala | 2611 taskAttemptNumber: Option[Int] = None) extends Logging { 2620 if (taskAttemptNumber.isDefined) s"_${taskAttemptNumber.get}" else ""
|