@Before public void before() { JobId mockJobId = mock(JobId.class); when(mockJobId.toString()).thenReturn("testJobId"); summary.setJobId(mockJobId); summary.setJobSubmitTime(2L); summary.setJobLaunchTime(3L); summary.setFirstMapTaskLaunchTime(4L); summary.setFirstReduceTaskLaunchTime(5L); summary.setJobFinishTime(6L); summary.setNumSucceededMaps(1); summary.setNumFailedMaps(0); summary.setNumSucceededReduces(1); summary.setNumFailedReduces(0); summary.setNumKilledMaps(0); summary.setNumKilledReduces(0); summary.setUser("testUser"); summary.setQueue("testQueue"); summary.setJobStatus("testJobStatus"); summary.setMapSlotSeconds(7); summary.setReduceSlotSeconds(8); summary.setJobName("testName"); }
if (summary.getJobStatus() == null) summary .setJobStatus(org.apache.hadoop.mapreduce.JobStatus.State.SUCCEEDED .toString()); case JOB_KILLED: JobUnsuccessfulCompletionEvent juce = (JobUnsuccessfulCompletionEvent) event; summary.setJobStatus(juce.getStatus()); summary.setNumFinishedMaps(context.getJob(jobId).getTotalMaps()); summary.setNumFinishedReduces(context.getJob(jobId).getTotalReduces());
if (summary.getJobStatus() == null) summary .setJobStatus(org.apache.hadoop.mapreduce.JobStatus.State.SUCCEEDED .toString()); case JOB_KILLED: JobUnsuccessfulCompletionEvent juce = (JobUnsuccessfulCompletionEvent) event; summary.setJobStatus(juce.getStatus()); summary.setNumFinishedMaps(context.getJob(jobId).getTotalMaps()); summary.setNumFinishedReduces(context.getJob(jobId).getTotalReduces());
if (summary.getJobStatus() == null) summary .setJobStatus(org.apache.hadoop.mapreduce.JobStatus.State.SUCCEEDED .toString()); - job.getFailedReduces() - job.getKilledReduces(); summary.setJobStatus(juce.getStatus()); summary.setNumSucceededMaps(successfulMaps); summary.setNumSucceededReduces(successfulReduces);