| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package org.apache.hadoop.tools; |
| |
| import java.io.IOException; |
| import java.net.URI; |
| import java.util.Iterator; |
| |
| import junit.framework.TestCase; |
| |
| import org.apache.hadoop.conf.Configuration; |
| import org.apache.hadoop.fs.FileStatus; |
| import org.apache.hadoop.fs.FileSystem; |
| import org.apache.hadoop.fs.FsShell; |
| import org.apache.hadoop.fs.FSDataInputStream; |
| import org.apache.hadoop.fs.FSDataOutputStream; |
| import org.apache.hadoop.fs.Path; |
| import org.apache.hadoop.hdfs.MiniDFSCluster; |
| import org.apache.hadoop.io.LongWritable; |
| import org.apache.hadoop.io.Text; |
| import org.apache.hadoop.mapred.*; |
| import org.apache.hadoop.tools.HadoopArchives; |
| import org.apache.hadoop.util.ToolRunner; |
| import org.mortbay.log.Log; |
| |
| /** |
| * test the har file system |
| * create a har filesystem |
| * run fs commands |
| * and then run a map reduce job |
| */ |
| public class TestHarFileSystem extends TestCase { |
| private Path inputPath, inputrelPath; |
| private MiniDFSCluster dfscluster; |
| private MiniMRCluster mapred; |
| private FileSystem fs; |
| private Path filea, fileb, filec; |
| private Path archivePath; |
| |
| protected void setUp() throws Exception { |
| super.setUp(); |
| dfscluster = new MiniDFSCluster(new Configuration(), 2, true, null); |
| fs = dfscluster.getFileSystem(); |
| mapred = new MiniMRCluster(2, fs.getUri().toString(), 1); |
| inputPath = new Path(fs.getHomeDirectory(), "test"); |
| inputrelPath = new Path(fs.getHomeDirectory().toUri(). |
| getPath().substring(1), "test"); |
| filea = new Path(inputPath,"a"); |
| fileb = new Path(inputPath,"b"); |
| filec = new Path(inputPath,"c c"); |
| archivePath = new Path(fs.getHomeDirectory(), "tmp"); |
| fs.mkdirs(inputPath); |
| FSDataOutputStream out = fs.create(filea); |
| out.write("a".getBytes()); |
| out.close(); |
| out = fs.create(fileb); |
| out.write("b".getBytes()); |
| out.close(); |
| out = fs.create(filec); |
| out.write("c".getBytes()); |
| out.close(); |
| } |
| |
| protected void tearDown() throws Exception { |
| try { |
| if (mapred != null) { |
| mapred.shutdown(); |
| } |
| if (dfscluster != null) { |
| dfscluster.shutdown(); |
| } |
| } catch(Exception e) { |
| System.err.println(e); |
| } |
| super.tearDown(); |
| } |
| |
| static class TextMapperReducer implements Mapper<LongWritable, Text, Text, Text>, |
| Reducer<Text, Text, Text, Text> { |
| |
| public void configure(JobConf conf) { |
| //do nothing |
| } |
| |
| public void map(LongWritable key, Text value, OutputCollector<Text, Text> output, Reporter reporter) throws IOException { |
| output.collect(value, new Text("")); |
| } |
| |
| public void close() throws IOException { |
| // do nothing |
| } |
| |
| public void reduce(Text key, Iterator<Text> values, OutputCollector<Text, Text> output, Reporter reporter) throws IOException { |
| while(values.hasNext()) { |
| values.next(); |
| output.collect(key, null); |
| } |
| } |
| } |
| |
| /* check bytes in the har output files */ |
| private void checkBytes(Path harPath, Configuration conf) throws IOException { |
| Path harFilea = new Path(harPath, "a"); |
| Path harFileb = new Path(harPath, "b"); |
| Path harFilec = new Path(harPath, "c c"); |
| FileSystem harFs = harFilea.getFileSystem(conf); |
| FSDataInputStream fin = harFs.open(harFilea); |
| byte[] b = new byte[4]; |
| int readBytes = fin.read(b); |
| fin.close(); |
| assertTrue("harFilea strings are unequal.", (b[0] == "a".getBytes()[0])); |
| fin = harFs.open(harFileb); |
| fin.read(b); |
| fin.close(); |
| assertTrue("harFileb strings are unequal.", (b[0] == "b".getBytes()[0])); |
| fin = harFs.open(harFilec); |
| fin.read(b); |
| fin.close(); |
| assertTrue("harFilec strings are unequal.", (b[0] == "c".getBytes()[0])); |
| } |
| |
| private void checkProperties(Path harPath, Configuration conf) throws IOException { |
| Path harFilea = new Path(harPath, "a"); |
| Path harFileb = new Path(harPath, "b"); |
| Path harFilec = new Path(harPath, "c c"); |
| FileSystem harFs = harFilea.getFileSystem(conf); |
| |
| Path nonharFilea = new Path(inputPath, "a"); |
| Path nonharFileb = new Path(inputPath, "b"); |
| Path nonharFilec = new Path(inputPath, "c c"); |
| FileSystem nonharFs = nonharFilea.getFileSystem(conf); |
| |
| assertEquals("Modification times do not match for a", |
| harFs.getFileStatus(harFilea).getModificationTime(), |
| nonharFs.getFileStatus(nonharFilea).getModificationTime()); |
| |
| assertEquals("Modification times do not match for b", |
| harFs.getFileStatus(harFileb).getModificationTime(), |
| nonharFs.getFileStatus(nonharFileb).getModificationTime()); |
| |
| assertEquals("Modification times do not match for c", |
| harFs.getFileStatus(harFilec).getModificationTime(), |
| nonharFs.getFileStatus(nonharFilec).getModificationTime()); |
| } |
| |
| /** |
| * check if the block size of the part files is what we had specified |
| */ |
| private void checkBlockSize(FileSystem fs, Path finalPath, long blockSize) throws IOException { |
| FileStatus[] statuses = fs.globStatus(new Path(finalPath, "part-*")); |
| for (FileStatus status: statuses) { |
| assertEquals("unexpected blocksize for part-* file", blockSize, |
| status.getBlockSize()); |
| } |
| } |
| |
| // test archives with a -p option |
| public void testRelativeArchives() throws Exception { |
| fs.delete(archivePath, true); |
| Configuration conf = mapred.createJobConf(); |
| HadoopArchives har = new HadoopArchives(conf); |
| |
| { |
| String[] args = new String[6]; |
| args[0] = "-archiveName"; |
| args[1] = "foo1.har"; |
| args[2] = "-p"; |
| args[3] = fs.getHomeDirectory().toString(); |
| args[4] = "test"; |
| args[5] = archivePath.toString(); |
| int ret = ToolRunner.run(har, args); |
| assertEquals("ToolRunner.run('-archiveName') failed.", 0, ret); |
| Path finalPath = new Path(archivePath, "foo1.har"); |
| Path fsPath = new Path(inputPath.toUri().getPath()); |
| Path filePath = new Path(finalPath, "test"); |
| // make it a har path |
| Path harPath = new Path("har://" + filePath.toUri().getPath()); |
| assertTrue("_index missing!", fs.exists(new Path(finalPath, "_index"))); |
| assertTrue("_masterindex missing!", |
| fs.exists(new Path(finalPath, "_masterindex"))); |
| // check for existence of only 1 part file, since part file size == 2GB |
| assertTrue("part-0 missing!", fs.exists(new Path(finalPath, "part-0"))); |
| assertTrue("part-1 exists!", !fs.exists(new Path(finalPath, "part-1"))); |
| assertTrue("part-2 exists!", !fs.exists(new Path(finalPath, "part-2"))); |
| assertTrue("_logs exists!", !fs.exists(new Path(finalPath, "_logs"))); |
| FileStatus[] statuses = fs.listStatus(finalPath); |
| args = new String[2]; |
| args[0] = "-ls"; |
| args[1] = harPath.toString(); |
| FsShell shell = new FsShell(conf); |
| ret = ToolRunner.run(shell, args); |
| // fileb and filec |
| assertEquals("ToolRunner.run('-ls') failed.", 0, ret); |
| checkBytes(harPath, conf); |
| checkProperties(harPath, conf); |
| /* check block size for path files */ |
| checkBlockSize(fs, finalPath, 512 * 1024 * 1024l); |
| } |
| |
| /** now try with different block size and part file size **/ |
| { |
| String[] args = new String[8]; |
| args[0] = "-Dhar.block.size=512"; |
| args[1] = "-Dhar.partfile.size=1"; |
| args[2] = "-archiveName"; |
| args[3] = "foo.har"; |
| args[4] = "-p"; |
| args[5] = fs.getHomeDirectory().toString(); |
| args[6] = "test"; |
| args[7] = archivePath.toString(); |
| int ret = ToolRunner.run(har, args); |
| assertEquals("ToolRunner.run('-archiveName') failed.", 0, ret); |
| Path finalPath = new Path(archivePath, "foo.har"); |
| Path fsPath = new Path(inputPath.toUri().getPath()); |
| Path filePath = new Path(finalPath, "test"); |
| // make it a har path |
| Path harPath = new Path("har://" + filePath.toUri().getPath()); |
| assertTrue(fs.exists(new Path(finalPath, "_index"))); |
| assertTrue(fs.exists(new Path(finalPath, "_masterindex"))); |
| /*check for existence of 3 part files, since part file size == 1 */ |
| assertTrue(fs.exists(new Path(finalPath, "part-0"))); |
| assertTrue(fs.exists(new Path(finalPath, "part-1"))); |
| assertTrue(fs.exists(new Path(finalPath, "part-2"))); |
| assertTrue(!fs.exists(new Path(finalPath, "_logs"))); |
| FileStatus[] statuses = fs.listStatus(finalPath); |
| args = new String[2]; |
| args[0] = "-ls"; |
| args[1] = harPath.toString(); |
| FsShell shell = new FsShell(conf); |
| ret = ToolRunner.run(shell, args); |
| // fileb and filec |
| assertEquals("ToolRunner.run('-ls') failed.", 0, ret); |
| checkBytes(harPath, conf); |
| checkProperties(harPath, conf); |
| checkBlockSize(fs, finalPath, 512); |
| } |
| } |
| |
| public void testArchivesWithMapred() throws Exception { |
| fs.delete(archivePath, true); |
| Configuration conf = mapred.createJobConf(); |
| HadoopArchives har = new HadoopArchives(conf); |
| String[] args = new String[4]; |
| |
| //check for destination not specified |
| args[0] = "-archiveName"; |
| args[1] = "foo.har"; |
| args[2] = "-p"; |
| args[3] = "/"; |
| int ret = ToolRunner.run(har, args); |
| assertFalse("bogus ToolRunner.run('-archiveName') didn't fail??", 0 == ret); |
| args = new String[6]; |
| //check for wrong archiveName |
| args[0] = "-archiveName"; |
| args[1] = "/d/foo.har"; |
| args[2] = "-p"; |
| args[3] = "/"; |
| args[4] = inputrelPath.toString(); |
| args[5] = archivePath.toString(); |
| ret = ToolRunner.run(har, args); |
| assertFalse("bogus ToolRunner.run('-archiveName') didn't fail??", 0 == ret); |
| // see if dest is a file |
| args[1] = "foo.har"; |
| args[5] = filec.toString(); |
| ret = ToolRunner.run(har, args); |
| assertFalse("bogus ToolRunner.run('-archiveName') didn't fail??", 0 == ret); |
| //this is a valid run |
| args[0] = "-archiveName"; |
| args[1] = "foo.har"; |
| args[2] = "-p"; |
| args[3] = "/"; |
| args[4] = inputrelPath.toString(); |
| args[5] = archivePath.toString(); |
| ret = ToolRunner.run(har, args); |
| // check for the existence of the archive |
| assertEquals("ToolRunner.run('-archiveName') failed.", 0, ret); |
| ///try running it again. it should not |
| // override the directory |
| ret = ToolRunner.run(har, args); |
| assertTrue("ToolRunner.run('-archiveName') rerun did NOT fail.", ret != 0); |
| Path finalPath = new Path(archivePath, "foo.har"); |
| Path fsPath = new Path(inputPath.toUri().getPath()); |
| String relative = fsPath.toString().substring(1); |
| Path filePath = new Path(finalPath, relative); |
| //make it a har path |
| URI uri = fs.getUri(); |
| Path harPath = new Path("har://" + "hdfs-" + uri.getHost() +":" + |
| uri.getPort() + filePath.toUri().getPath()); |
| assertTrue(fs.exists(new Path(finalPath, "_index"))); |
| assertTrue(fs.exists(new Path(finalPath, "_masterindex"))); |
| assertTrue(!fs.exists(new Path(finalPath, "_logs"))); |
| //creation tested |
| //check if the archive is same |
| // do ls and cat on all the files |
| |
| FsShell shell = new FsShell(conf); |
| args = new String[2]; |
| args[0] = "-ls"; |
| args[1] = harPath.toString(); |
| ret = ToolRunner.run(shell, args); |
| // ls should work. |
| assertEquals("ToolRunner.run('-ls') failed.", 0, ret); |
| // now check for contents of filea, fileb, and filec |
| Path harFilea = new Path(harPath, "a"); |
| Path harFileb = new Path(harPath, "b"); |
| Path harFilec = new Path(harPath, "c c"); |
| FileSystem harFs = harFilea.getFileSystem(conf); |
| FSDataInputStream fin = harFs.open(harFilea); |
| byte[] b = new byte[4]; |
| int readBytes = fin.read(b); |
| assertTrue("Empty read.", readBytes > 0); |
| fin.close(); |
| assertTrue("harFilea strings are unequal.", (b[0] == "a".getBytes()[0])); |
| fin = harFs.open(harFileb); |
| readBytes = fin.read(b); |
| assertTrue("Empty read.", readBytes > 0); |
| fin.close(); |
| assertTrue("harFileb strings are unequal.", (b[0] == "b".getBytes()[0])); |
| fin = harFs.open(harFilec); |
| readBytes = fin.read(b); |
| assertTrue("Empty read.", readBytes > 0); |
| fin.close(); |
| assertTrue("harFilec strings are unequal.", (b[0] == "c".getBytes()[0])); |
| // ok all files match |
| // run a map reduce job |
| FileSystem fsHar = harPath.getFileSystem(conf); |
| FileStatus[] bla = fsHar.listStatus(harPath); |
| Path outdir = new Path(fs.getHomeDirectory(), "mapout"); |
| JobConf jobconf = mapred.createJobConf(); |
| FileInputFormat.addInputPath(jobconf, harPath); |
| jobconf.setInputFormat(TextInputFormat.class); |
| jobconf.setOutputFormat(TextOutputFormat.class); |
| FileOutputFormat.setOutputPath(jobconf, outdir); |
| jobconf.setMapperClass(TextMapperReducer.class); |
| jobconf.setMapOutputKeyClass(Text.class); |
| jobconf.setMapOutputValueClass(Text.class); |
| jobconf.setReducerClass(TextMapperReducer.class); |
| jobconf.setNumReduceTasks(1); |
| JobClient.runJob(jobconf); |
| args[1] = outdir.toString(); |
| ret = ToolRunner.run(shell, args); |
| |
| FileStatus[] status = fs.globStatus(new Path(outdir, "part*")); |
| Path reduceFile = status[0].getPath(); |
| FSDataInputStream reduceIn = fs.open(reduceFile); |
| b = new byte[6]; |
| readBytes = reduceIn.read(b); |
| assertEquals("Should read 6 bytes instead of "+readBytes+".", 6, readBytes); |
| //assuming all the 6 bytes were read. |
| Text readTxt = new Text(b); |
| assertTrue("string read from reduceIn doesn't match expected value.", |
| "a\nb\nc\n".equals(readTxt.toString())); |
| assertEquals("number of bytes left should be -1.", -1, reduceIn.read(b)); |
| reduceIn.close(); |
| } |
| } |