blob: 5d36c92e4209d4e0bb28fdb0821e40fbcd6eb4e0 [file] [log] [blame]
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.mapreduce;
import java.io.File;
import java.io.IOException;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.PathFilter;
import org.apache.hadoop.fs.FileAlreadyExistsException;
import org.apache.hadoop.mapred.HadoopTestCase;
import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
import org.apache.hadoop.mapreduce.task.TaskAttemptContextImpl;
import org.junit.Ignore;
@Ignore
public class TestNoJobSetupCleanup extends HadoopTestCase {
private static String TEST_ROOT_DIR =
new File(System.getProperty("test.build.data","/tmp"))
.toURI().toString().replace(' ', '+');
private final Path inDir = new Path(TEST_ROOT_DIR, "./wc/input");
private final Path outDir = new Path(TEST_ROOT_DIR, "./wc/output");
public TestNoJobSetupCleanup() throws IOException {
super(HadoopTestCase.CLUSTER_MR , HadoopTestCase.LOCAL_FS, 2, 2);
}
private Job submitAndValidateJob(Configuration conf, int numMaps, int numReds)
throws IOException, InterruptedException, ClassNotFoundException {
Job job = MapReduceTestUtil.createJob(conf, inDir, outDir,
numMaps, numReds);
job.setJobSetupCleanupNeeded(false);
job.setOutputFormatClass(MyOutputFormat.class);
job.waitForCompletion(true);
assertTrue(job.isSuccessful());
assertTrue(job.getTaskReports(TaskType.JOB_SETUP).length == 0);
assertTrue(job.getTaskReports(TaskType.JOB_CLEANUP).length == 0);
assertTrue(job.getTaskReports(TaskType.MAP).length == numMaps);
assertTrue(job.getTaskReports(TaskType.REDUCE).length == numReds);
FileSystem fs = FileSystem.get(conf);
assertTrue("Job output directory doesn't exit!", fs.exists(outDir));
// job commit done only in cleanup
// therefore output should still be in temp location
String tempWorkingPathStr = outDir + Path.SEPARATOR + "_temporary"
+ Path.SEPARATOR + "0";
Path tempWorkingPath = new Path(tempWorkingPathStr);
FileStatus[] list = fs.listStatus(tempWorkingPath, new OutputFilter());
int numPartFiles = numReds == 0 ? numMaps : numReds;
assertTrue("Number of part-files is " + list.length + " and not "
+ numPartFiles, list.length == numPartFiles);
return job;
}
public void testNoJobSetupCleanup() throws Exception {
try {
Configuration conf = createJobConf();
// run a job without job-setup and cleanup
submitAndValidateJob(conf, 1, 1);
// run a map only job.
submitAndValidateJob(conf, 1, 0);
// run empty job without job setup and cleanup
submitAndValidateJob(conf, 0, 0);
// run empty job without job setup and cleanup, with non-zero reduces
submitAndValidateJob(conf, 0, 1);
} finally {
tearDown();
}
}
public static class MyOutputFormat extends TextOutputFormat {
public void checkOutputSpecs(JobContext job)
throws FileAlreadyExistsException, IOException{
super.checkOutputSpecs(job);
// creating dummy TaskAttemptID
TaskAttemptID tid = new TaskAttemptID("jt", 1, TaskType.JOB_SETUP, 0, 0);
getOutputCommitter(new TaskAttemptContextImpl(job.getConfiguration(), tid)).
setupJob(job);
}
}
private static class OutputFilter implements PathFilter {
public boolean accept(Path path) {
return !(path.getName().startsWith("_"));
}
}
}