/dports/devel/hadoop/hadoop-1.2.1/src/test/org/apache/hadoop/mapred/ |
H A D | TestTaskLogsTruncater.java | 178 writeRandomBytes(attemptID, attemptID, log, 500); in testNoTruncationNeeded() 180 File logIndex = TaskLog.getIndexFile(attemptID, false); in testNoTruncationNeeded() 210 logLengths = getAllLogsFileLengths(attemptID, false); in testNoTruncationNeeded() 239 writeRandomBytes(attemptID, attemptID, log, 1500); in testDisabledLogTruncation() 279 writeRandomBytes(attemptID, attemptID, log, 1500); in testLogTruncationOnFinishing() 301 logLengths = getAllLogsFileLengths(attemptID, false); in testLogTruncationOnFinishing() 330 writeRandomBytes(attemptID, attemptID, LogName.SYSLOG, 1500); in testLogTruncation() 331 writeRandomBytes(attemptID, attemptID, LogName.STDERR, 500); in testLogTruncation() 350 attemptID, false); in testLogTruncation() 357 attemptID, false); in testLogTruncation() [all …]
|
H A D | TestTaskLauncher.java | 73 TaskAttemptID attemptID = new TaskAttemptID(jtId, 1, true, 0, 0); in testExternalKillForLaunchTask() local 74 Task task = new MapTask(null, attemptID, 0, null, 5); in testExternalKillForLaunchTask() 77 TaskInProgress killTip = tt.runningTasks.get(attemptID); in testExternalKillForLaunchTask() 88 assertEquals("Launcher didnt pick up the task " + attemptID + "to launch", in testExternalKillForLaunchTask() 94 tt.processKillTaskAction(new KillTaskAction(attemptID)); in testExternalKillForLaunchTask()
|
H A D | TestJvmManager.java | 119 TaskAttemptID attemptID = new TaskAttemptID("test", 0, true, 0, 0); in testJvmKill() local 120 Task task = new MapTask(null, attemptID, 0, null, MAP_SLOTS); in testJvmKill() 125 RunningJob rjob = new RunningJob(attemptID.getJobID()); in testJvmKill() 130 newTaskDistributedCacheManager(attemptID.getJobID(), taskConf); in testJvmKill() 199 attemptID = new TaskAttemptID("test", 0, true, 0, 1); in testJvmKill() 200 task = new MapTask(null, attemptID, 0, null, MAP_SLOTS); in testJvmKill()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/speculate/ |
H A D | LegacyTaskRuntimeEstimator.java | 49 TaskAttemptId attemptID = status.id; in updateAttempt() local 50 TaskId taskID = attemptID.getTaskId(); in updateAttempt() 64 TaskAttempt taskAttempt = task.getAttempt(attemptID); in updateAttempt() 70 Long boxedStart = startTimes.get(attemptID); in updateAttempt() 118 (Map<TaskAttempt, AtomicLong> data, TaskAttemptId attemptID) { in storedPerAttemptValue() argument 119 TaskId taskID = attemptID.getTaskId(); in storedPerAttemptValue() 129 TaskAttempt taskAttempt = task.getAttempt(attemptID); in storedPerAttemptValue() 142 public long estimatedRuntime(TaskAttemptId attemptID) { in estimatedRuntime() argument 143 return storedPerAttemptValue(attemptRuntimeEstimates, attemptID); in estimatedRuntime() 147 public long runtimeEstimateVariance(TaskAttemptId attemptID) { in runtimeEstimateVariance() argument [all …]
|
H A D | ExponentiallySmoothedTaskRuntimeEstimator.java | 94 (TaskAttemptId attemptID, float newProgress, long newTime) { 96 AtomicReference<EstimateVector> vectorRef = estimates.get(attemptID); 99 estimates.putIfAbsent(attemptID, new AtomicReference<EstimateVector>(null)); 100 incorporateReading(attemptID, newProgress, newTime); 112 incorporateReading(attemptID, newProgress, newTime); 122 private EstimateVector getEstimateVector(TaskAttemptId attemptID) { 123 AtomicReference<EstimateVector> vectorRef = estimates.get(attemptID); 186 TaskAttemptId attemptID = status.id; 190 incorporateReading(attemptID, progress, timestamp);
|
H A D | StartEndTimesBase.java | 75 public long attemptEnrolledTime(TaskAttemptId attemptID) { in attemptEnrolledTime() argument 76 Long result = startTimes.get(attemptID); in attemptEnrolledTime() 162 TaskAttemptId attemptID = status.id; in updateAttempt() local 163 TaskId taskID = attemptID.getTaskId(); in updateAttempt() 177 Long boxedStart = startTimes.get(attemptID); in updateAttempt() 180 TaskAttempt taskAttempt = task.getAttempt(attemptID); in updateAttempt()
|
H A D | SpeculatorEvent.java | 49 public SpeculatorEvent(TaskAttemptId attemptID, boolean flag, long timestamp) { in SpeculatorEvent() argument 52 this.reportedStatus.id = attemptID; in SpeculatorEvent() 53 this.taskID = attemptID.getTaskId(); in SpeculatorEvent()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapred/ |
H A D | TaskAttemptListenerImpl.java | 191 taskHeartbeatHandler.progressing(attemptID); in canCommit() 200 Job job = context.getJob(attemptID.getTaskId().getJobId()); in canCommit() 201 Task task = job.getTask(attemptID.getTaskId()); in canCommit() 202 return task.canCommit(attemptID); in canCommit() 226 taskHeartbeatHandler.progressing(attemptID); in commitPending() 229 new TaskAttemptEvent(attemptID, in commitPending() 240 taskHeartbeatHandler.progressing(attemptID); in done() 293 taskHeartbeatHandler.progressing(attemptID); in getMapCompletionEvents() 315 taskHeartbeatHandler.progressing(attemptID); in reportDiagnosticInfo() 471 taskHeartbeatHandler.register(attemptID); in registerLaunchedTask() [all …]
|
H A D | LocalContainerLauncher.java | 264 TaskAttemptId attemptID = launchEv.getTaskAttemptID(); in runTask() local 266 Job job = context.getAllJobs().get(attemptID.getTaskId().getJobId()); in runTask() 272 job.getTask(attemptID.getTaskId()); in runTask() 284 new TaskAttemptContainerLaunchedEvent(attemptID, -1)); in runTask() 301 runSubtask(remoteTask, ytask.getType(), attemptID, numMapTasks, in runTask() 312 context.getEventHandler().handle(new TaskAttemptEvent(attemptID, in runTask() 323 if (futures.remove(attemptID) != null) { in runTask() 324 LOG.info("removed attempt " + attemptID + in runTask() 332 TaskAttemptId attemptID, in runSubtask() argument 338 TypeConverter.fromYarn(attemptID); in runSubtask() [all …]
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/rm/ |
H A D | ContainerRequestEvent.java | 32 public ContainerRequestEvent(TaskAttemptId attemptID, in ContainerRequestEvent() argument 35 super(attemptID, ContainerAllocator.EventType.CONTAINER_REQ); in ContainerRequestEvent() 41 ContainerRequestEvent(TaskAttemptId attemptID, Resource capability) { in ContainerRequestEvent() argument 42 this(attemptID, capability, new String[0], new String[0]); in ContainerRequestEvent() 47 TaskAttemptId attemptID, in createContainerRequestEventForFailedContainer() argument 50 return new ContainerRequestEvent(attemptID, capability); in createContainerRequestEventForFailedContainer()
|
H A D | ContainerAllocatorEvent.java | 27 private TaskAttemptId attemptID; field in ContainerAllocatorEvent 29 public ContainerAllocatorEvent(TaskAttemptId attemptID, in ContainerAllocatorEvent() argument 32 this.attemptID = attemptID; in ContainerAllocatorEvent() 36 return attemptID; in getAttemptID()
|
H A D | ContainerFailedEvent.java | 27 public ContainerFailedEvent(TaskAttemptId attemptID, String contMgrAddr) { in ContainerFailedEvent() argument 28 super(attemptID, ContainerAllocator.EventType.CONTAINER_FAILED); in ContainerFailedEvent()
|
H A D | RMContainerAllocator.java | 750 if (attemptID == null) { in getResources() 755 assignedRequests.remove(attemptID); in getResources() 818 TaskAttemptId attemptID) { in createContainerFinishedEvent() argument 822 return new TaskAttemptEvent(attemptID, in createContainerFinishedEvent() 825 return new TaskAttemptEvent(attemptID, in createContainerFinishedEvent() 967 reduces.put(req.attemptID, req); in addReduce() 1043 + toBeReplacedReq.attemptID); in assign() 1049 maps.put(newReq.attemptID, newReq); in assign() 1052 reduces.put(newReq.attemptID, newReq); in assign() 1087 assigned.attemptID, allocated, applicationACLs)); in containerAssigned() [all …]
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/commit/ |
H A D | CommitterTaskAbortEvent.java | 26 private final TaskAttemptId attemptID; field in CommitterTaskAbortEvent 29 public CommitterTaskAbortEvent(TaskAttemptId attemptID, in CommitterTaskAbortEvent() argument 32 this.attemptID = attemptID; in CommitterTaskAbortEvent() 37 return attemptID; in getAttemptID()
|
/dports/devel/hadoop/hadoop-1.2.1/src/test/tools/data/rumen/small-trace-test/ |
H A D | truncated-trace-output | 21 "attemptID" : "attempt_200904211745_0002_m_000010_0", 66 "attemptID" : "attempt_200904211745_0002_m_000017_0", 111 "attemptID" : "attempt_200904211745_0002_m_000019_0", 156 "attemptID" : "attempt_200904211745_0002_m_000005_0", 201 "attemptID" : "attempt_200904211745_0002_m_000000_0", 246 "attemptID" : "attempt_200904211745_0002_m_000009_0", 289 "attemptID" : "attempt_200904211745_0002_m_000002_0", 314 "attemptID" : "attempt_200904211745_0002_m_000002_1", 359 "attemptID" : "attempt_200904211745_0002_m_000004_0", 404 "attemptID" : "attempt_200904211745_0002_m_000007_0", [all …]
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/ |
H A D | TaskHeartbeatHandler.java | 112 public void progressing(TaskAttemptId attemptID) { in progressing() argument 115 ReportTime time = runningAttempts.get(attemptID); in progressing() 122 public void register(TaskAttemptId attemptID) { in register() argument 123 runningAttempts.put(attemptID, new ReportTime(clock.getTime())); in register() 126 public void unregister(TaskAttemptId attemptID) { in unregister() argument 127 runningAttempts.remove(attemptID); in unregister()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/job/event/ |
H A D | TaskTAttemptEvent.java | 26 private TaskAttemptId attemptID; field in TaskTAttemptEvent 30 this.attemptID = id; in TaskTAttemptEvent() 34 return attemptID; in getTaskAttemptID()
|
H A D | TaskAttemptEvent.java | 30 private TaskAttemptId attemptID; field in TaskAttemptEvent 39 this.attemptID = id; in TaskAttemptEvent() 43 return attemptID; in getTaskAttemptID()
|
H A D | TaskAttemptDiagnosticsUpdateEvent.java | 28 public TaskAttemptDiagnosticsUpdateEvent(TaskAttemptId attemptID, in TaskAttemptDiagnosticsUpdateEvent() argument 30 super(attemptID, TaskAttemptEventType.TA_DIAGNOSTICS_UPDATE); in TaskAttemptDiagnosticsUpdateEvent()
|
H A D | TaskAttemptKillEvent.java | 28 public TaskAttemptKillEvent(TaskAttemptId attemptID, in TaskAttemptKillEvent() argument 30 super(attemptID, TaskAttemptEventType.TA_KILL); in TaskAttemptKillEvent()
|
/dports/www/grafana8/grafana-8.3.6/pkg/services/alerting/ |
H A D | engine.go | 153 case attemptID, more := <-attemptChan: 157 go e.processJob(attemptID, attemptChan, cancelChan, job) 171 func (e *AlertEngine) processJob(attemptID int, attemptChan chan int, cancelChan chan context.Cance… 206 span.SetTag("attemptID", attemptID) 214 if attemptID < setting.AlertingMaxAttempts { 216 …ntext.Rule.ID, "name", evalContext.Rule.Name, "firing", evalContext.Firing, "attemptID", attemptID) 217 attemptChan <- (attemptID + 1) 244 …ntext.Rule.ID, "name", evalContext.Rule.Name, "firing", evalContext.Firing, "attemptID", attemptID)
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapreduce/v2/app/ |
H A D | TestFail.java | 283 protected void attemptLaunched(TaskAttemptId attemptID) { 284 if (attemptID.getTaskId().getId() == 0) {//check if it is first task 287 new TaskAttemptEvent(attemptID, 291 new TaskAttemptEvent(attemptID, 304 protected void attemptLaunched(TaskAttemptId attemptID) { 305 if (attemptID.getTaskId().getId() == 0 && attemptID.getId() == 0) { 309 new TaskAttemptEvent(attemptID, 313 new TaskAttemptEvent(attemptID,
|
/dports/devel/hadoop/hadoop-1.2.1/src/tools/org/apache/hadoop/tools/rumen/ |
H A D | LoggedTaskAttempt.java | 41 String attemptID; field in LoggedTaskAttempt 110 return attemptID; in getAttemptID() 113 void setAttemptID(String attemptID) { in setAttemptID() argument 114 this.attemptID = attemptID; in setAttemptID() 487 compare1(attemptID, other.attemptID, loc, "attemptID"); in deepCompare()
|
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/ |
H A D | TestTaskCommit.java | 195 TaskAttemptID attemptID = TaskAttemptID.forName(dummyAttemptID); in testTaskCleanupDoesNotCommit() local 197 JobContext jContext = new JobContextImpl(job, attemptID.getJobID()); in testTaskCleanupDoesNotCommit() 203 attemptID = TaskAttemptID.forName(dummyAttemptID); in testTaskCleanupDoesNotCommit() 204 Task task = new MapTask(null, attemptID, 0, null, 1); in testTaskCleanupDoesNotCommit() 207 task.initialize(job, attemptID.getJobID(), Reporter.NULL, false); in testTaskCleanupDoesNotCommit()
|
/dports/devel/spark/spark-2.1.1/core/src/main/scala/org/apache/spark/ |
H A D | SparkHadoopWriter.scala | 49 private var attemptID = 0 71 jobid, splitID, attemptID, conf.value) 154 attemptID = attemptid 158 new TaskAttemptID(new TaskID(jID.value, TaskType.MAP, splitID), attemptID))
|