blob: 3c9411e305922854d2ab64018369ee094b10cea6 [file] [log] [blame]
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.mapred.pipes;
import java.io.*;
import java.util.*;
import org.apache.hadoop.fs.*;
import org.apache.hadoop.io.*;
import org.apache.hadoop.mapred.*;
/**
* This is a support class to test Hadoop Pipes when using C++ RecordReaders.
* It defines an InputFormat with InputSplits that are just strings. The
* RecordReaders are not implemented in Java, naturally...
*/
public class WordCountInputFormat
extends FileInputFormat<IntWritable, Text> {
static class WordCountInputSplit implements InputSplit {
private String filename;
WordCountInputSplit() { }
WordCountInputSplit(Path filename) {
this.filename = filename.toUri().getPath();
}
public void write(DataOutput out) throws IOException {
Text.writeString(out, filename);
}
public void readFields(DataInput in) throws IOException {
filename = Text.readString(in);
}
public long getLength() { return 0L; }
public String[] getLocations() { return new String[0]; }
}
public InputSplit[] getSplits(JobConf conf,
int numSplits) throws IOException {
ArrayList<InputSplit> result = new ArrayList<InputSplit>();
FileSystem local = FileSystem.getLocal(conf);
for(Path dir: getInputPaths(conf)) {
for(FileStatus file: local.listStatus(dir)) {
result.add(new WordCountInputSplit(file.getPath()));
}
}
return result.toArray(new InputSplit[result.size()]);
}
public RecordReader<IntWritable, Text> getRecordReader(InputSplit split,
JobConf conf,
Reporter reporter) {
return new RecordReader<IntWritable, Text>(){
public boolean next(IntWritable key, Text value) throws IOException {
return false;
}
public IntWritable createKey() {
return new IntWritable();
}
public Text createValue() {
return new Text();
}
public long getPos() {
return 0;
}
public void close() { }
public float getProgress() {
return 0.0f;
}
};
}
}