| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.hadoop.hdfs; |
| |
| import java.io.File; |
| import java.io.IOException; |
| import java.security.MessageDigest; |
| import java.security.NoSuchAlgorithmException; |
| import java.util.Random; |
| |
| import org.apache.commons.logging.Log; |
| import org.apache.commons.logging.impl.Log4JLogger; |
| import org.apache.hadoop.conf.Configuration; |
| import org.apache.hadoop.fs.ContentSummary; |
| import org.apache.hadoop.fs.FSDataInputStream; |
| import org.apache.hadoop.fs.FSDataOutputStream; |
| import org.apache.hadoop.fs.FileSystem; |
| import org.apache.hadoop.fs.Path; |
| import org.apache.hadoop.hdfs.protocol.ExtendedBlock; |
| import org.apache.hadoop.hdfs.protocol.DatanodeInfo; |
| import org.apache.hadoop.hdfs.protocol.LocatedBlock; |
| import org.apache.hadoop.hdfs.protocol.LocatedBlocks; |
| import org.apache.hadoop.hdfs.server.datanode.DataNode; |
| import org.apache.hadoop.hdfs.server.datanode.FSDataset; |
| import org.apache.hadoop.io.IOUtils; |
| import org.apache.log4j.Level; |
| import org.junit.Assert; |
| import org.junit.Test; |
| |
| public class TestFiHftp { |
| final Log LOG = FileSystem.LOG; |
| { |
| ((Log4JLogger)LOG).getLogger().setLevel(Level.ALL); |
| } |
| |
| static final short DATANODE_NUM = 1; |
| static final Random ran = new Random(); |
| static final byte[] buffer = new byte[1 << 16]; |
| static final MessageDigest md5; |
| static { |
| try { |
| md5 = MessageDigest.getInstance("MD5"); |
| } catch (NoSuchAlgorithmException e) { |
| throw new RuntimeException(e); |
| } |
| } |
| |
| private static byte[] createFile(FileSystem fs, Path name, long length, |
| short replication, long blocksize) throws IOException { |
| final FSDataOutputStream out = fs.create(name, false, 4096, |
| replication, blocksize); |
| try { |
| for(long n = length; n > 0; ) { |
| ran.nextBytes(buffer); |
| final int w = n < buffer.length? (int)n: buffer.length; |
| out.write(buffer, 0, w); |
| md5.update(buffer, 0, w); |
| n -= w; |
| } |
| } finally { |
| IOUtils.closeStream(out); |
| } |
| return md5.digest(); |
| } |
| |
| @Test |
| public void testHftpOpen() throws IOException { |
| final Configuration conf = new Configuration(); |
| MiniDFSCluster cluster = null; |
| try { |
| cluster = new MiniDFSCluster.Builder(conf).numDataNodes(DATANODE_NUM).build(); |
| cluster.waitActive(); |
| |
| //test with a file |
| //which is larger than the servlet response buffer size |
| { |
| final long blocksize = 1L << 20; // |
| final long filesize = 2*blocksize + 100; |
| runTestHftpOpen(cluster, "/foo", blocksize, filesize); |
| } |
| |
| //test with a small file |
| //which is smaller than the servlet response buffer size |
| { |
| final long blocksize = 1L << 10; // |
| final long filesize = 2*blocksize + 100; |
| runTestHftpOpen(cluster, "/small", blocksize, filesize); |
| } |
| } finally { |
| if (cluster != null) {cluster.shutdown();} |
| } |
| } |
| |
| /** |
| * A test with a 3GB file. |
| * It may take ~6 minutes. |
| */ |
| void largeFileTest(final MiniDFSCluster cluster) throws IOException { |
| final long blocksize = 128L << 20; |
| final long filesize = 3L << 30; |
| runTestHftpOpen(cluster, "/large", blocksize, filesize); |
| } |
| |
| /** |
| * @param blocksize |
| * @param filesize must be > block size |
| */ |
| private void runTestHftpOpen(final MiniDFSCluster cluster, final String file, |
| final long blocksize, final long filesize) throws IOException { |
| //create a file |
| final DistributedFileSystem dfs = (DistributedFileSystem)cluster.getFileSystem(); |
| final Path filepath = new Path(file); |
| final byte[] filemd5 = createFile(dfs, filepath, filesize, DATANODE_NUM, |
| blocksize); |
| DFSTestUtil.waitReplication(dfs, filepath, DATANODE_NUM); |
| |
| //test hftp open and read |
| final HftpFileSystem hftpfs = cluster.getHftpFileSystem(0); |
| { |
| final FSDataInputStream in = hftpfs.open(filepath); |
| long bytesRead = 0; |
| try { |
| for(int r; (r = in.read(buffer)) != -1; ) { |
| bytesRead += r; |
| md5.update(buffer, 0, r); |
| } |
| } finally { |
| LOG.info("bytesRead=" + bytesRead); |
| in.close(); |
| } |
| Assert.assertEquals(filesize, bytesRead); |
| Assert.assertArrayEquals(filemd5, md5.digest()); |
| } |
| |
| //delete the second block |
| final DFSClient client = dfs.getClient(); |
| final LocatedBlocks locatedblocks = client.getNamenode().getBlockLocations( |
| file, 0, filesize); |
| Assert.assertEquals((filesize - 1)/blocksize + 1, |
| locatedblocks.locatedBlockCount()); |
| final LocatedBlock lb = locatedblocks.get(1); |
| final ExtendedBlock blk = lb.getBlock(); |
| Assert.assertEquals(blocksize, lb.getBlockSize()); |
| final DatanodeInfo[] datanodeinfos = lb.getLocations(); |
| Assert.assertEquals(DATANODE_NUM, datanodeinfos.length); |
| final DataNode dn = cluster.getDataNode(datanodeinfos[0].getIpcPort()); |
| LOG.info("dn=" + dn + ", blk=" + blk + " (length=" + blk.getNumBytes() + ")"); |
| final FSDataset data = (FSDataset)dn.getFSDataset(); |
| final File blkfile = data.getBlockFile(blk); |
| Assert.assertTrue(blkfile.delete()); |
| |
| //read again by hftp, should get an exception |
| LOG.info("hftpfs.getUri() = " + hftpfs.getUri()); |
| final ContentSummary cs = hftpfs.getContentSummary(filepath); |
| LOG.info("hftpfs.getContentSummary = " + cs); |
| Assert.assertEquals(filesize, cs.getLength()); |
| |
| final FSDataInputStream in = hftpfs.open(hftpfs.makeQualified(filepath)); |
| long bytesRead = 0; |
| try { |
| for(int r; (r = in.read(buffer)) != -1; ) { |
| bytesRead += r; |
| } |
| Assert.fail(); |
| } catch(IOException ioe) { |
| LOG.info("GOOD: get an exception", ioe); |
| } finally { |
| LOG.info("bytesRead=" + bytesRead); |
| in.close(); |
| } |
| } |
| } |