blob: 37cb91f023b8437b2cd60e94fa2e1954c05f1c21 [file] [log] [blame]
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.mapred.lib;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapred.*;
import java.io.DataOutputStream;
import java.io.IOException;
import java.util.Iterator;
public class TestChainMapReduce extends HadoopTestCase {
private static Path getFlagDir(boolean local) {
Path flagDir = new Path("testing/chain/flags");
// Hack for local FS that does not have the concept of a 'mounting point'
if (local) {
String localPathRoot = System.getProperty("test.build.data", "/tmp")
.replace(' ', '+');
flagDir = new Path(localPathRoot, flagDir);
}
return flagDir;
}
private static void cleanFlags(JobConf conf) throws IOException {
FileSystem fs = FileSystem.get(conf);
fs.delete(getFlagDir(conf.getBoolean("localFS", true)), true);
fs.mkdirs(getFlagDir(conf.getBoolean("localFS", true)));
}
private static void writeFlag(JobConf conf, String flag) throws IOException {
FileSystem fs = FileSystem.get(conf);
if (getFlag(conf, flag)) {
fail("Flag " + flag + " already exists");
}
DataOutputStream file =
fs.create(new Path(getFlagDir(conf.getBoolean("localFS", true)), flag));
file.close();
}
private static boolean getFlag(JobConf conf, String flag) throws IOException {
FileSystem fs = FileSystem.get(conf);
return fs
.exists(new Path(getFlagDir(conf.getBoolean("localFS", true)), flag));
}
public TestChainMapReduce() throws IOException {
super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1);
}
public void testChain() throws Exception {
Path inDir = new Path("testing/chain/input");
Path outDir = new Path("testing/chain/output");
// Hack for local FS that does not have the concept of a 'mounting point'
if (isLocalFS()) {
String localPathRoot = System.getProperty("test.build.data", "/tmp")
.replace(' ', '+');
inDir = new Path(localPathRoot, inDir);
outDir = new Path(localPathRoot, outDir);
}
JobConf conf = createJobConf();
conf.setBoolean("localFS", isLocalFS());
conf.setInt("mapreduce.job.maps", 1);
cleanFlags(conf);
FileSystem fs = FileSystem.get(conf);
fs.delete(outDir, true);
if (!fs.mkdirs(inDir)) {
throw new IOException("Mkdirs failed to create " + inDir.toString());
}
DataOutputStream file = fs.create(new Path(inDir, "part-0"));
file.writeBytes("1\n2\n");
file.close();
conf.setJobName("chain");
conf.setInputFormat(TextInputFormat.class);
conf.setOutputFormat(TextOutputFormat.class);
conf.set("a", "X");
JobConf mapAConf = new JobConf(false);
mapAConf.set("a", "A");
ChainMapper.addMapper(conf, AMap.class, LongWritable.class, Text.class,
LongWritable.class, Text.class, true, mapAConf);
ChainMapper.addMapper(conf, BMap.class, LongWritable.class, Text.class,
LongWritable.class, Text.class, false, null);
JobConf reduceConf = new JobConf(false);
reduceConf.set("a", "C");
ChainReducer.setReducer(conf, CReduce.class, LongWritable.class, Text.class,
LongWritable.class, Text.class, true, reduceConf);
ChainReducer.addMapper(conf, DMap.class, LongWritable.class, Text.class,
LongWritable.class, Text.class, false, null);
JobConf mapEConf = new JobConf(false);
mapEConf.set("a", "E");
ChainReducer.addMapper(conf, EMap.class, LongWritable.class, Text.class,
LongWritable.class, Text.class, true, mapEConf);
FileInputFormat.setInputPaths(conf, inDir);
FileOutputFormat.setOutputPath(conf, outDir);
JobClient jc = new JobClient(conf);
RunningJob job = jc.submitJob(conf);
while (!job.isComplete()) {
Thread.sleep(100);
}
assertTrue(getFlag(conf, "configure.A"));
assertTrue(getFlag(conf, "configure.B"));
assertTrue(getFlag(conf, "configure.C"));
assertTrue(getFlag(conf, "configure.D"));
assertTrue(getFlag(conf, "configure.E"));
assertTrue(getFlag(conf, "map.A.value.1"));
assertTrue(getFlag(conf, "map.A.value.2"));
assertTrue(getFlag(conf, "map.B.value.1"));
assertTrue(getFlag(conf, "map.B.value.2"));
assertTrue(getFlag(conf, "reduce.C.value.2"));
assertTrue(getFlag(conf, "reduce.C.value.1"));
assertTrue(getFlag(conf, "map.D.value.1"));
assertTrue(getFlag(conf, "map.D.value.2"));
assertTrue(getFlag(conf, "map.E.value.1"));
assertTrue(getFlag(conf, "map.E.value.2"));
assertTrue(getFlag(conf, "close.A"));
assertTrue(getFlag(conf, "close.B"));
assertTrue(getFlag(conf, "close.C"));
assertTrue(getFlag(conf, "close.D"));
assertTrue(getFlag(conf, "close.E"));
}
public static class AMap extends IDMap {
public AMap() {
super("A", "A", true);
}
}
public static class BMap extends IDMap {
public BMap() {
super("B", "X", false);
}
}
public static class CReduce extends IDReduce {
public CReduce() {
super("C", "C");
}
}
public static class DMap extends IDMap {
public DMap() {
super("D", "X", false);
}
}
public static class EMap extends IDMap {
public EMap() {
super("E", "E", true);
}
}
public static class IDMap
implements Mapper<LongWritable, Text, LongWritable, Text> {
private JobConf conf;
private String name;
private String prop;
private boolean byValue;
public IDMap(String name, String prop, boolean byValue) {
this.name = name;
this.prop = prop;
this.byValue = byValue;
}
public void configure(JobConf conf) {
this.conf = conf;
assertEquals(prop, conf.get("a"));
try {
writeFlag(conf, "configure." + name);
} catch (IOException ex) {
throw new RuntimeException(ex);
}
}
public void map(LongWritable key, Text value,
OutputCollector<LongWritable, Text> output,
Reporter reporter) throws IOException {
writeFlag(conf, "map." + name + ".value." + value);
key.set(10);
output.collect(key, value);
if (byValue) {
assertEquals(10, key.get());
} else {
assertNotSame(10, key.get());
}
key.set(11);
}
public void close() throws IOException {
try {
writeFlag(conf, "close." + name);
} catch (IOException ex) {
throw new RuntimeException(ex);
}
}
}
public static class IDReduce
implements Reducer<LongWritable, Text, LongWritable, Text> {
private JobConf conf;
private String name;
private String prop;
private boolean byValue = false;
public IDReduce(String name, String prop) {
this.name = name;
this.prop = prop;
}
public void configure(JobConf conf) {
this.conf = conf;
assertEquals(prop, conf.get("a"));
try {
writeFlag(conf, "configure." + name);
} catch (IOException ex) {
throw new RuntimeException(ex);
}
}
public void reduce(LongWritable key, Iterator<Text> values,
OutputCollector<LongWritable, Text> output,
Reporter reporter) throws IOException {
while (values.hasNext()) {
Text value = values.next();
writeFlag(conf, "reduce." + name + ".value." + value);
key.set(10);
output.collect(key, value);
if (byValue) {
assertEquals(10, key.get());
} else {
assertNotSame(10, key.get());
}
key.set(11);
}
}
public void close() throws IOException {
try {
writeFlag(conf, "close." + name);
} catch (IOException ex) {
throw new RuntimeException(ex);
}
}
}
}