| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| */ |
| package org.apache.apex.examples.mroperator; |
| |
| import java.io.IOException; |
| import java.util.Iterator; |
| import java.util.StringTokenizer; |
| |
| import org.apache.hadoop.fs.Path; |
| import org.apache.hadoop.io.IntWritable; |
| import org.apache.hadoop.io.LongWritable; |
| import org.apache.hadoop.io.Text; |
| import org.apache.hadoop.mapred.FileInputFormat; |
| import org.apache.hadoop.mapred.FileOutputFormat; |
| import org.apache.hadoop.mapred.JobClient; |
| import org.apache.hadoop.mapred.JobConf; |
| import org.apache.hadoop.mapred.MapReduceBase; |
| import org.apache.hadoop.mapred.Mapper; |
| import org.apache.hadoop.mapred.OutputCollector; |
| import org.apache.hadoop.mapred.Reducer; |
| import org.apache.hadoop.mapred.Reporter; |
| import org.apache.hadoop.mapred.TextInputFormat; |
| import org.apache.hadoop.mapred.TextOutputFormat; |
| |
| /** |
| * <p>WordCount class.</p> |
| * |
| * @since 0.9.0 |
| */ |
| @SuppressWarnings("deprecation") |
| public class WordCount |
| { |
| |
| public static class Map extends MapReduceBase implements Mapper<LongWritable, Text, Text, IntWritable> |
| { |
| private static final IntWritable one = new IntWritable(1); |
| private Text word = new Text(); |
| |
| public void map(LongWritable key, Text value, OutputCollector<Text, IntWritable> output, Reporter reporter) throws IOException |
| { |
| String line = value.toString(); |
| StringTokenizer tokenizer = new StringTokenizer(line); |
| while (tokenizer.hasMoreTokens()) { |
| word.set(tokenizer.nextToken()); |
| output.collect(word, one); |
| } |
| } |
| } |
| |
| public static class Reduce extends MapReduceBase implements Reducer<Text, IntWritable, Text, IntWritable> |
| { |
| public void reduce(Text key, Iterator<IntWritable> values, OutputCollector<Text, IntWritable> output, Reporter reporter) throws IOException |
| { |
| int sum = 0; |
| while (values.hasNext()) { |
| sum += values.next().get(); |
| } |
| output.collect(key, new IntWritable(sum)); |
| } |
| } |
| |
| public void run(String[] args) throws Exception |
| { |
| |
| JobConf conf = new JobConf(this.getClass()); |
| conf.setJobName("wordcount"); |
| |
| conf.setOutputKeyClass(Text.class); |
| conf.setOutputValueClass(IntWritable.class); |
| |
| conf.setMapperClass(Map.class); |
| conf.setCombinerClass(Reduce.class); |
| conf.setReducerClass(Reduce.class); |
| |
| conf.setInputFormat(TextInputFormat.class); |
| conf.setOutputFormat(TextOutputFormat.class); |
| |
| FileInputFormat.setInputPaths(conf, new Path(args[0])); |
| FileOutputFormat.setOutputPath(conf, new Path(args[1])); |
| |
| JobClient.runJob(conf); |
| } |
| |
| public static void main(String[] args) throws Exception |
| { |
| new WordCount().run(args); |
| } |
| } |