Home
last modified time | relevance | path

Searched refs:attemptID (Results 1 – 25 of 55) sorted by relevance

123

/dports/devel/hadoop/hadoop-1.2.1/src/test/org/apache/hadoop/mapred/
H A DTestTaskLogsTruncater.java178 writeRandomBytes(attemptID, attemptID, log, 500); in testNoTruncationNeeded()
180 File logIndex = TaskLog.getIndexFile(attemptID, false); in testNoTruncationNeeded()
210 logLengths = getAllLogsFileLengths(attemptID, false); in testNoTruncationNeeded()
239 writeRandomBytes(attemptID, attemptID, log, 1500); in testDisabledLogTruncation()
279 writeRandomBytes(attemptID, attemptID, log, 1500); in testLogTruncationOnFinishing()
301 logLengths = getAllLogsFileLengths(attemptID, false); in testLogTruncationOnFinishing()
330 writeRandomBytes(attemptID, attemptID, LogName.SYSLOG, 1500); in testLogTruncation()
331 writeRandomBytes(attemptID, attemptID, LogName.STDERR, 500); in testLogTruncation()
350 attemptID, false); in testLogTruncation()
357 attemptID, false); in testLogTruncation()
[all …]
H A DTestTaskLauncher.java73 TaskAttemptID attemptID = new TaskAttemptID(jtId, 1, true, 0, 0); in testExternalKillForLaunchTask() local
74 Task task = new MapTask(null, attemptID, 0, null, 5); in testExternalKillForLaunchTask()
77 TaskInProgress killTip = tt.runningTasks.get(attemptID); in testExternalKillForLaunchTask()
88 assertEquals("Launcher didnt pick up the task " + attemptID + "to launch", in testExternalKillForLaunchTask()
94 tt.processKillTaskAction(new KillTaskAction(attemptID)); in testExternalKillForLaunchTask()
H A DTestJvmManager.java119 TaskAttemptID attemptID = new TaskAttemptID("test", 0, true, 0, 0); in testJvmKill() local
120 Task task = new MapTask(null, attemptID, 0, null, MAP_SLOTS); in testJvmKill()
125 RunningJob rjob = new RunningJob(attemptID.getJobID()); in testJvmKill()
130 newTaskDistributedCacheManager(attemptID.getJobID(), taskConf); in testJvmKill()
199 attemptID = new TaskAttemptID("test", 0, true, 0, 1); in testJvmKill()
200 task = new MapTask(null, attemptID, 0, null, MAP_SLOTS); in testJvmKill()
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/speculate/
H A DLegacyTaskRuntimeEstimator.java49 TaskAttemptId attemptID = status.id; in updateAttempt() local
50 TaskId taskID = attemptID.getTaskId(); in updateAttempt()
64 TaskAttempt taskAttempt = task.getAttempt(attemptID); in updateAttempt()
70 Long boxedStart = startTimes.get(attemptID); in updateAttempt()
118 (Map<TaskAttempt, AtomicLong> data, TaskAttemptId attemptID) { in storedPerAttemptValue() argument
119 TaskId taskID = attemptID.getTaskId(); in storedPerAttemptValue()
129 TaskAttempt taskAttempt = task.getAttempt(attemptID); in storedPerAttemptValue()
142 public long estimatedRuntime(TaskAttemptId attemptID) { in estimatedRuntime() argument
143 return storedPerAttemptValue(attemptRuntimeEstimates, attemptID); in estimatedRuntime()
147 public long runtimeEstimateVariance(TaskAttemptId attemptID) { in runtimeEstimateVariance() argument
[all …]
H A DExponentiallySmoothedTaskRuntimeEstimator.java94 (TaskAttemptId attemptID, float newProgress, long newTime) {
96 AtomicReference<EstimateVector> vectorRef = estimates.get(attemptID);
99 estimates.putIfAbsent(attemptID, new AtomicReference<EstimateVector>(null));
100 incorporateReading(attemptID, newProgress, newTime);
112 incorporateReading(attemptID, newProgress, newTime);
122 private EstimateVector getEstimateVector(TaskAttemptId attemptID) {
123 AtomicReference<EstimateVector> vectorRef = estimates.get(attemptID);
186 TaskAttemptId attemptID = status.id;
190 incorporateReading(attemptID, progress, timestamp);
H A DStartEndTimesBase.java75 public long attemptEnrolledTime(TaskAttemptId attemptID) { in attemptEnrolledTime() argument
76 Long result = startTimes.get(attemptID); in attemptEnrolledTime()
162 TaskAttemptId attemptID = status.id; in updateAttempt() local
163 TaskId taskID = attemptID.getTaskId(); in updateAttempt()
177 Long boxedStart = startTimes.get(attemptID); in updateAttempt()
180 TaskAttempt taskAttempt = task.getAttempt(attemptID); in updateAttempt()
H A DSpeculatorEvent.java49 public SpeculatorEvent(TaskAttemptId attemptID, boolean flag, long timestamp) { in SpeculatorEvent() argument
52 this.reportedStatus.id = attemptID; in SpeculatorEvent()
53 this.taskID = attemptID.getTaskId(); in SpeculatorEvent()
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapred/
H A DTaskAttemptListenerImpl.java191 taskHeartbeatHandler.progressing(attemptID); in canCommit()
200 Job job = context.getJob(attemptID.getTaskId().getJobId()); in canCommit()
201 Task task = job.getTask(attemptID.getTaskId()); in canCommit()
202 return task.canCommit(attemptID); in canCommit()
226 taskHeartbeatHandler.progressing(attemptID); in commitPending()
229 new TaskAttemptEvent(attemptID, in commitPending()
240 taskHeartbeatHandler.progressing(attemptID); in done()
293 taskHeartbeatHandler.progressing(attemptID); in getMapCompletionEvents()
315 taskHeartbeatHandler.progressing(attemptID); in reportDiagnosticInfo()
471 taskHeartbeatHandler.register(attemptID); in registerLaunchedTask()
[all …]
H A DLocalContainerLauncher.java264 TaskAttemptId attemptID = launchEv.getTaskAttemptID(); in runTask() local
266 Job job = context.getAllJobs().get(attemptID.getTaskId().getJobId()); in runTask()
272 job.getTask(attemptID.getTaskId()); in runTask()
284 new TaskAttemptContainerLaunchedEvent(attemptID, -1)); in runTask()
301 runSubtask(remoteTask, ytask.getType(), attemptID, numMapTasks, in runTask()
312 context.getEventHandler().handle(new TaskAttemptEvent(attemptID, in runTask()
323 if (futures.remove(attemptID) != null) { in runTask()
324 LOG.info("removed attempt " + attemptID + in runTask()
332 TaskAttemptId attemptID, in runSubtask() argument
338 TypeConverter.fromYarn(attemptID); in runSubtask()
[all …]
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/rm/
H A DContainerRequestEvent.java32 public ContainerRequestEvent(TaskAttemptId attemptID, in ContainerRequestEvent() argument
35 super(attemptID, ContainerAllocator.EventType.CONTAINER_REQ); in ContainerRequestEvent()
41 ContainerRequestEvent(TaskAttemptId attemptID, Resource capability) { in ContainerRequestEvent() argument
42 this(attemptID, capability, new String[0], new String[0]); in ContainerRequestEvent()
47 TaskAttemptId attemptID, in createContainerRequestEventForFailedContainer() argument
50 return new ContainerRequestEvent(attemptID, capability); in createContainerRequestEventForFailedContainer()
H A DContainerAllocatorEvent.java27 private TaskAttemptId attemptID; field in ContainerAllocatorEvent
29 public ContainerAllocatorEvent(TaskAttemptId attemptID, in ContainerAllocatorEvent() argument
32 this.attemptID = attemptID; in ContainerAllocatorEvent()
36 return attemptID; in getAttemptID()
H A DContainerFailedEvent.java27 public ContainerFailedEvent(TaskAttemptId attemptID, String contMgrAddr) { in ContainerFailedEvent() argument
28 super(attemptID, ContainerAllocator.EventType.CONTAINER_FAILED); in ContainerFailedEvent()
H A DRMContainerAllocator.java750 if (attemptID == null) { in getResources()
755 assignedRequests.remove(attemptID); in getResources()
818 TaskAttemptId attemptID) { in createContainerFinishedEvent() argument
822 return new TaskAttemptEvent(attemptID, in createContainerFinishedEvent()
825 return new TaskAttemptEvent(attemptID, in createContainerFinishedEvent()
967 reduces.put(req.attemptID, req); in addReduce()
1043 + toBeReplacedReq.attemptID); in assign()
1049 maps.put(newReq.attemptID, newReq); in assign()
1052 reduces.put(newReq.attemptID, newReq); in assign()
1087 assigned.attemptID, allocated, applicationACLs)); in containerAssigned()
[all …]
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/commit/
H A DCommitterTaskAbortEvent.java26 private final TaskAttemptId attemptID; field in CommitterTaskAbortEvent
29 public CommitterTaskAbortEvent(TaskAttemptId attemptID, in CommitterTaskAbortEvent() argument
32 this.attemptID = attemptID; in CommitterTaskAbortEvent()
37 return attemptID; in getAttemptID()
/dports/devel/hadoop/hadoop-1.2.1/src/test/tools/data/rumen/small-trace-test/
H A Dtruncated-trace-output21 "attemptID" : "attempt_200904211745_0002_m_000010_0",
66 "attemptID" : "attempt_200904211745_0002_m_000017_0",
111 "attemptID" : "attempt_200904211745_0002_m_000019_0",
156 "attemptID" : "attempt_200904211745_0002_m_000005_0",
201 "attemptID" : "attempt_200904211745_0002_m_000000_0",
246 "attemptID" : "attempt_200904211745_0002_m_000009_0",
289 "attemptID" : "attempt_200904211745_0002_m_000002_0",
314 "attemptID" : "attempt_200904211745_0002_m_000002_1",
359 "attemptID" : "attempt_200904211745_0002_m_000004_0",
404 "attemptID" : "attempt_200904211745_0002_m_000007_0",
[all …]
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/
H A DTaskHeartbeatHandler.java112 public void progressing(TaskAttemptId attemptID) { in progressing() argument
115 ReportTime time = runningAttempts.get(attemptID); in progressing()
122 public void register(TaskAttemptId attemptID) { in register() argument
123 runningAttempts.put(attemptID, new ReportTime(clock.getTime())); in register()
126 public void unregister(TaskAttemptId attemptID) { in unregister() argument
127 runningAttempts.remove(attemptID); in unregister()
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/job/event/
H A DTaskTAttemptEvent.java26 private TaskAttemptId attemptID; field in TaskTAttemptEvent
30 this.attemptID = id; in TaskTAttemptEvent()
34 return attemptID; in getTaskAttemptID()
H A DTaskAttemptEvent.java30 private TaskAttemptId attemptID; field in TaskAttemptEvent
39 this.attemptID = id; in TaskAttemptEvent()
43 return attemptID; in getTaskAttemptID()
H A DTaskAttemptDiagnosticsUpdateEvent.java28 public TaskAttemptDiagnosticsUpdateEvent(TaskAttemptId attemptID, in TaskAttemptDiagnosticsUpdateEvent() argument
30 super(attemptID, TaskAttemptEventType.TA_DIAGNOSTICS_UPDATE); in TaskAttemptDiagnosticsUpdateEvent()
H A DTaskAttemptKillEvent.java28 public TaskAttemptKillEvent(TaskAttemptId attemptID, in TaskAttemptKillEvent() argument
30 super(attemptID, TaskAttemptEventType.TA_KILL); in TaskAttemptKillEvent()
/dports/www/grafana8/grafana-8.3.6/pkg/services/alerting/
H A Dengine.go153 case attemptID, more := <-attemptChan:
157 go e.processJob(attemptID, attemptChan, cancelChan, job)
171 func (e *AlertEngine) processJob(attemptID int, attemptChan chan int, cancelChan chan context.Cance…
206 span.SetTag("attemptID", attemptID)
214 if attemptID < setting.AlertingMaxAttempts {
216 …ntext.Rule.ID, "name", evalContext.Rule.Name, "firing", evalContext.Firing, "attemptID", attemptID)
217 attemptChan <- (attemptID + 1)
244 …ntext.Rule.ID, "name", evalContext.Rule.Name, "firing", evalContext.Firing, "attemptID", attemptID)
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapreduce/v2/app/
H A DTestFail.java283 protected void attemptLaunched(TaskAttemptId attemptID) {
284 if (attemptID.getTaskId().getId() == 0) {//check if it is first task
287 new TaskAttemptEvent(attemptID,
291 new TaskAttemptEvent(attemptID,
304 protected void attemptLaunched(TaskAttemptId attemptID) {
305 if (attemptID.getTaskId().getId() == 0 && attemptID.getId() == 0) {
309 new TaskAttemptEvent(attemptID,
313 new TaskAttemptEvent(attemptID,
/dports/devel/hadoop/hadoop-1.2.1/src/tools/org/apache/hadoop/tools/rumen/
H A DLoggedTaskAttempt.java41 String attemptID; field in LoggedTaskAttempt
110 return attemptID; in getAttemptID()
113 void setAttemptID(String attemptID) { in setAttemptID() argument
114 this.attemptID = attemptID; in setAttemptID()
487 compare1(attemptID, other.attemptID, loc, "attemptID"); in deepCompare()
/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/
H A DTestTaskCommit.java195 TaskAttemptID attemptID = TaskAttemptID.forName(dummyAttemptID); in testTaskCleanupDoesNotCommit() local
197 JobContext jContext = new JobContextImpl(job, attemptID.getJobID()); in testTaskCleanupDoesNotCommit()
203 attemptID = TaskAttemptID.forName(dummyAttemptID); in testTaskCleanupDoesNotCommit()
204 Task task = new MapTask(null, attemptID, 0, null, 1); in testTaskCleanupDoesNotCommit()
207 task.initialize(job, attemptID.getJobID(), Reporter.NULL, false); in testTaskCleanupDoesNotCommit()
/dports/devel/spark/spark-2.1.1/core/src/main/scala/org/apache/spark/
H A DSparkHadoopWriter.scala49 private var attemptID = 0
71 jobid, splitID, attemptID, conf.value)
154 attemptID = attemptid
158 new TaskAttemptID(new TaskID(jID.value, TaskType.MAP, splitID), attemptID))

123