blob: 410207ee15e32a984e3629aa04f640cd2ddb4724 [file] [log] [blame]
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.mapred;
import java.io.BufferedReader;
import java.io.FileInputStream;
import java.io.File;
import java.io.InputStreamReader;
import java.io.IOException;
import junit.framework.TestCase;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.TaskType;
import org.apache.hadoop.mapreduce.SleepJob;
/**
* A JUnit test to test Kill Job & Fail Job functionality with local file
* system.
*/
public class TestJobKillAndFail extends TestCase {
static final Log LOG = LogFactory.getLog(TestJobKillAndFail.class);
private static String TEST_ROOT_DIR = new File(System.getProperty(
"test.build.data", "/tmp")).toURI().toString().replace(' ', '+');
/**
* TaskController instance that just sets a flag when a stack dump
* is performed in a child thread.
*/
static class MockStackDumpTaskController extends DefaultTaskController {
static volatile int numStackDumps = 0;
static final Log LOG = LogFactory.getLog(TestJobKillAndFail.class);
public MockStackDumpTaskController() {
LOG.info("Instantiated MockStackDumpTC");
}
@Override
void dumpTaskStack(TaskControllerContext context) {
LOG.info("Got stack-dump request in TaskController");
MockStackDumpTaskController.numStackDumps++;
super.dumpTaskStack(context);
}
}
/** If a task was killed, then dumpTaskStack() should have been
* called. Test whether or not the counter was incremented
* and succeed/fail based on this. */
private void checkForStackDump(boolean expectDump, int lastNumDumps) {
int curNumDumps = MockStackDumpTaskController.numStackDumps;
LOG.info("curNumDumps=" + curNumDumps + "; lastNumDumps=" + lastNumDumps
+ "; expect=" + expectDump);
if (expectDump) {
assertTrue("No stack dump recorded!", lastNumDumps < curNumDumps);
} else {
assertTrue("Stack dump happened anyway!", lastNumDumps == curNumDumps);
}
}
public void testJobFailAndKill() throws Exception {
MiniMRCluster mr = null;
try {
JobConf jtConf = new JobConf();
jtConf.set("mapred.jobtracker.instrumentation",
JTInstrumentation.class.getName());
jtConf.set("mapreduce.tasktracker.taskcontroller",
MockStackDumpTaskController.class.getName());
mr = new MiniMRCluster(2, "file:///", 3, null, null, jtConf);
JTInstrumentation instr = (JTInstrumentation)
mr.getJobTrackerRunner().getJobTracker().getInstrumentation();
// run the TCs
JobConf conf = mr.createJobConf();
conf.setInt(Job.COMPLETION_POLL_INTERVAL_KEY, 50);
Path inDir = new Path(TEST_ROOT_DIR + "/failkilljob/input");
Path outDir = new Path(TEST_ROOT_DIR + "/failkilljob/output");
RunningJob runningJob = UtilsForTests.runJobFail(conf, inDir, outDir);
// Checking that the Job got failed
assertEquals(runningJob.getJobState(), JobStatus.FAILED);
assertTrue(instr.verifyJob());
assertEquals(1, instr.failed);
instr.reset();
int prevNumDumps = MockStackDumpTaskController.numStackDumps;
runningJob = UtilsForTests.runJobKill(conf, inDir, outDir);
// Checking that the Job got killed
assertTrue(runningJob.isComplete());
assertEquals(runningJob.getJobState(), JobStatus.KILLED);
assertTrue(instr.verifyJob());
assertEquals(1, instr.killed);
// check that job kill does not put a stacktrace in task logs.
checkForStackDump(false, prevNumDumps);
// Test that a task that times out does have a stack trace
conf = mr.createJobConf();
conf.setInt(JobContext.TASK_TIMEOUT, 10000);
conf.setInt(Job.COMPLETION_POLL_INTERVAL_KEY, 50);
SleepJob sleepJob = new SleepJob();
sleepJob.setConf(conf);
Job job = sleepJob.createJob(1, 0, 30000, 1,0, 0);
job.setMaxMapAttempts(1);
prevNumDumps = MockStackDumpTaskController.numStackDumps;
job.waitForCompletion(true);
checkForStackDump(true, prevNumDumps);
} finally {
if (mr != null) {
mr.shutdown();
}
}
}
static class JTInstrumentation extends JobTrackerInstrumentation {
volatile int failed;
volatile int killed;
volatile int addPrep;
volatile int decPrep;
volatile int addRunning;
volatile int decRunning;
void reset() {
failed = 0;
killed = 0;
addPrep = 0;
decPrep = 0;
addRunning = 0;
decRunning = 0;
}
boolean verifyJob() {
return addPrep==1 && decPrep==1 && addRunning==1 && decRunning==1;
}
public JTInstrumentation(JobTracker jt, JobConf conf) {
super(jt, conf);
}
public synchronized void addPrepJob(JobConf conf, JobID id)
{
addPrep++;
}
public synchronized void decPrepJob(JobConf conf, JobID id)
{
decPrep++;
}
public synchronized void addRunningJob(JobConf conf, JobID id)
{
addRunning++;
}
public synchronized void decRunningJob(JobConf conf, JobID id)
{
decRunning++;
}
public synchronized void failedJob(JobConf conf, JobID id)
{
failed++;
}
public synchronized void killedJob(JobConf conf, JobID id)
{
killed++;
}
}
}