| /* |
| * Licensed to the Apache Software Foundation (ASF) under one or more |
| * contributor license agreements. See the NOTICE file distributed with |
| * this work for additional information regarding copyright ownership. |
| * The ASF licenses this file to You under the Apache License, Version 2.0 |
| * (the "License"); you may not use this file except in compliance with |
| * the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.accumulo.testing.core.randomwalk.sequential; |
| |
| import java.io.IOException; |
| import java.util.Iterator; |
| |
| import org.apache.accumulo.core.client.Accumulo; |
| import org.apache.accumulo.core.client.ClientInfo; |
| import org.apache.accumulo.core.client.mapreduce.AccumuloInputFormat; |
| import org.apache.accumulo.core.client.mapreduce.AccumuloOutputFormat; |
| import org.apache.accumulo.core.data.Key; |
| import org.apache.accumulo.core.data.Mutation; |
| import org.apache.accumulo.core.data.Value; |
| import org.apache.hadoop.conf.Configured; |
| import org.apache.hadoop.io.IntWritable; |
| import org.apache.hadoop.io.NullWritable; |
| import org.apache.hadoop.io.Text; |
| import org.apache.hadoop.mapreduce.Job; |
| import org.apache.hadoop.mapreduce.Mapper; |
| import org.apache.hadoop.mapreduce.Reducer; |
| import org.apache.hadoop.util.Tool; |
| import org.slf4j.Logger; |
| import org.slf4j.LoggerFactory; |
| |
| public class MapRedVerifyTool extends Configured implements Tool { |
| protected final Logger log = LoggerFactory.getLogger(this.getClass()); |
| |
| public static class SeqMapClass extends Mapper<Key,Value,NullWritable,IntWritable> { |
| @Override |
| public void map(Key row, Value data, Context output) throws IOException, InterruptedException { |
| Integer num = Integer.valueOf(row.getRow().toString()); |
| output.write(NullWritable.get(), new IntWritable(num.intValue())); |
| } |
| } |
| |
| public static class SeqReduceClass extends Reducer<NullWritable,IntWritable,Text,Mutation> { |
| @Override |
| public void reduce(NullWritable ignore, Iterable<IntWritable> values, Context output) throws IOException, InterruptedException { |
| Iterator<IntWritable> iterator = values.iterator(); |
| |
| if (iterator.hasNext() == false) { |
| return; |
| } |
| |
| int start = iterator.next().get(); |
| int index = start; |
| while (iterator.hasNext()) { |
| int next = iterator.next().get(); |
| if (next != index + 1) { |
| writeMutation(output, start, index); |
| start = next; |
| } |
| index = next; |
| } |
| writeMutation(output, start, index); |
| } |
| |
| public void writeMutation(Context output, int start, int end) throws IOException, InterruptedException { |
| Mutation m = new Mutation(new Text(String.format("%010d", start))); |
| m.put(new Text(String.format("%010d", end)), new Text(""), new Value(new byte[0])); |
| output.write(null, m); |
| } |
| } |
| |
| @Override |
| public int run(String[] args) throws Exception { |
| Job job = Job.getInstance(getConf(), this.getClass().getSimpleName()); |
| job.setJarByClass(this.getClass()); |
| |
| if (job.getJar() == null) { |
| log.error("M/R requires a jar file! Run mvn package."); |
| return 1; |
| } |
| |
| ClientInfo info = Accumulo.newClient().from(args[0]).info(); |
| AccumuloInputFormat.setClientInfo(job, info); |
| AccumuloInputFormat.setInputTableName(job, args[1]); |
| |
| AccumuloOutputFormat.setClientInfo(job, info); |
| AccumuloOutputFormat.setDefaultTableName(job, args[2]); |
| |
| job.setInputFormatClass(AccumuloInputFormat.class); |
| job.setMapperClass(SeqMapClass.class); |
| job.setMapOutputKeyClass(NullWritable.class); |
| job.setMapOutputValueClass(IntWritable.class); |
| |
| job.setReducerClass(SeqReduceClass.class); |
| job.setNumReduceTasks(1); |
| |
| job.setOutputFormatClass(AccumuloOutputFormat.class); |
| AccumuloOutputFormat.setCreateTables(job, true); |
| |
| job.waitForCompletion(true); |
| return job.isSuccessful() ? 0 : 1; |
| } |
| } |