| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package org.apache.hadoop.hdfs.server.datanode; |
| |
| import java.io.IOException; |
| |
| import junit.framework.Assert; |
| |
| import org.apache.commons.logging.Log; |
| import org.apache.commons.logging.LogFactory; |
| import org.apache.commons.logging.impl.Log4JLogger; |
| import org.apache.hadoop.conf.Configuration; |
| import org.apache.hadoop.fs.FileSystem; |
| import org.apache.hadoop.fs.Path; |
| import org.apache.hadoop.hdfs.DFSConfigKeys; |
| import org.apache.hadoop.hdfs.DFSTestUtil; |
| import org.apache.hadoop.hdfs.DFSUtil; |
| import org.apache.hadoop.hdfs.HdfsConfiguration; |
| import org.apache.hadoop.hdfs.MiniDFSCluster; |
| import org.apache.hadoop.hdfs.MiniDFSNNTopology; |
| import org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner; |
| import static org.apache.hadoop.hdfs.server.datanode.DataBlockScanner.SLEEP_PERIOD_MS; |
| import org.apache.log4j.Level; |
| import org.apache.log4j.Logger; |
| import org.junit.Test; |
| import org.junit.Ignore; |
| import static org.junit.Assert.fail; |
| |
| |
| public class TestMultipleNNDataBlockScanner { |
| private static final Log LOG = |
| LogFactory.getLog(TestMultipleNNDataBlockScanner.class); |
| Configuration conf; |
| MiniDFSCluster cluster = null; |
| String bpids[] = new String[3]; |
| FileSystem fs[] = new FileSystem[3]; |
| |
| public void setUp() throws IOException { |
| conf = new HdfsConfiguration(); |
| conf.setInt(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, 100); |
| conf.setInt(DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_KEY, 100); |
| cluster = new MiniDFSCluster.Builder(conf) |
| .nnTopology(MiniDFSNNTopology.simpleFederatedTopology(3)) |
| .build(); |
| for (int i = 0; i < 3; i++) { |
| cluster.waitActive(i); |
| } |
| for (int i = 0; i < 3; i++) { |
| bpids[i] = cluster.getNamesystem(i).getBlockPoolId(); |
| } |
| for (int i = 0; i < 3; i++) { |
| fs[i] = cluster.getFileSystem(i); |
| } |
| // Create 2 files on each namenode with 10 blocks each |
| for (int i = 0; i < 3; i++) { |
| DFSTestUtil.createFile(fs[i], new Path("file1"), 1000, (short) 1, 0); |
| DFSTestUtil.createFile(fs[i], new Path("file2"), 1000, (short) 1, 1); |
| } |
| } |
| |
| @Test(timeout=120000) |
| public void testDataBlockScanner() throws IOException, InterruptedException { |
| setUp(); |
| try { |
| DataNode dn = cluster.getDataNodes().get(0); |
| for (int i = 0; i < 3; i++) { |
| long blocksScanned = 0; |
| while (blocksScanned != 20) { |
| blocksScanned = dn.blockScanner.getBlocksScannedInLastRun(bpids[i]); |
| LOG.info("Waiting for all blocks to be scanned for bpid=" + bpids[i] |
| + "; Scanned so far=" + blocksScanned); |
| Thread.sleep(5000); |
| } |
| } |
| |
| StringBuilder buffer = new StringBuilder(); |
| dn.blockScanner.printBlockReport(buffer, false); |
| LOG.info("Block Report\n" + buffer.toString()); |
| } finally { |
| cluster.shutdown(); |
| } |
| } |
| |
| @Test(timeout=120000) |
| public void testBlockScannerAfterRefresh() throws IOException, |
| InterruptedException { |
| setUp(); |
| try { |
| Configuration dnConf = cluster.getDataNodes().get(0).getConf(); |
| Configuration conf = new HdfsConfiguration(dnConf); |
| StringBuilder namenodesBuilder = new StringBuilder(); |
| |
| String bpidToShutdown = cluster.getNamesystem(2).getBlockPoolId(); |
| for (int i = 0; i < 2; i++) { |
| String nsId = DFSUtil.getNamenodeNameServiceId(cluster |
| .getConfiguration(i)); |
| namenodesBuilder.append(nsId); |
| namenodesBuilder.append(","); |
| } |
| |
| conf.set(DFSConfigKeys.DFS_NAMESERVICES, namenodesBuilder |
| .toString()); |
| DataNode dn = cluster.getDataNodes().get(0); |
| dn.refreshNamenodes(conf); |
| |
| try { |
| while (true) { |
| dn.blockScanner.getBlocksScannedInLastRun(bpidToShutdown); |
| Thread.sleep(1000); |
| } |
| } catch (IOException ex) { |
| // Expected |
| LOG.info(ex.getMessage()); |
| } |
| |
| namenodesBuilder.append(DFSUtil.getNamenodeNameServiceId(cluster |
| .getConfiguration(2))); |
| conf.set(DFSConfigKeys.DFS_NAMESERVICES, namenodesBuilder |
| .toString()); |
| dn.refreshNamenodes(conf); |
| |
| for (int i = 0; i < 3; i++) { |
| long blocksScanned = 0; |
| while (blocksScanned != 20) { |
| blocksScanned = dn.blockScanner.getBlocksScannedInLastRun(bpids[i]); |
| LOG.info("Waiting for all blocks to be scanned for bpid=" + bpids[i] |
| + "; Scanned so far=" + blocksScanned); |
| Thread.sleep(5000); |
| } |
| } |
| } finally { |
| cluster.shutdown(); |
| } |
| } |
| |
| @Test(timeout=120000) |
| public void testBlockScannerAfterRestart() throws IOException, |
| InterruptedException { |
| setUp(); |
| try { |
| cluster.restartDataNode(0); |
| cluster.waitActive(); |
| DataNode dn = cluster.getDataNodes().get(0); |
| for (int i = 0; i < 3; i++) { |
| while (!dn.blockScanner.isInitialized(bpids[i])) { |
| Thread.sleep(1000); |
| } |
| long blocksScanned = 0; |
| while (blocksScanned != 20) { |
| if (dn.blockScanner != null) { |
| blocksScanned = dn.blockScanner.getBlocksScannedInLastRun(bpids[i]); |
| LOG.info("Waiting for all blocks to be scanned for bpid=" |
| + bpids[i] + "; Scanned so far=" + blocksScanned); |
| } |
| Thread.sleep(5000); |
| } |
| } |
| } finally { |
| cluster.shutdown(); |
| } |
| } |
| |
| @Test(timeout=120000) |
| public void test2NNBlockRescanInterval() throws IOException { |
| ((Log4JLogger)BlockPoolSliceScanner.LOG).getLogger().setLevel(Level.ALL); |
| Configuration conf = new HdfsConfiguration(); |
| cluster = new MiniDFSCluster.Builder(conf) |
| .nnTopology(MiniDFSNNTopology.simpleFederatedTopology(3)) |
| .build(); |
| |
| try { |
| FileSystem fs = cluster.getFileSystem(1); |
| Path file2 = new Path("/test/testBlockScanInterval"); |
| DFSTestUtil.createFile(fs, file2, 30, (short) 1, 0); |
| |
| fs = cluster.getFileSystem(0); |
| Path file1 = new Path("/test/testBlockScanInterval"); |
| DFSTestUtil.createFile(fs, file1, 30, (short) 1, 0); |
| for (int i = 0; i < 8; i++) { |
| LOG.info("Verifying that the blockscanner scans exactly once"); |
| waitAndScanBlocks(1, 1); |
| } |
| } finally { |
| cluster.shutdown(); |
| } |
| } |
| |
| /** |
| * HDFS-3828: DN rescans blocks too frequently |
| * |
| * @throws Exception |
| */ |
| @Test(timeout=120000) |
| public void testBlockRescanInterval() throws IOException { |
| ((Log4JLogger)BlockPoolSliceScanner.LOG).getLogger().setLevel(Level.ALL); |
| Configuration conf = new HdfsConfiguration(); |
| cluster = new MiniDFSCluster.Builder(conf).build(); |
| |
| try { |
| FileSystem fs = cluster.getFileSystem(); |
| Path file1 = new Path("/test/testBlockScanInterval"); |
| DFSTestUtil.createFile(fs, file1, 30, (short) 1, 0); |
| for (int i = 0; i < 4; i++) { |
| LOG.info("Verifying that the blockscanner scans exactly once"); |
| waitAndScanBlocks(1, 1); |
| } |
| } finally { |
| cluster.shutdown(); |
| } |
| } |
| |
| void waitAndScanBlocks(long scansLastRun, long scansTotal) |
| throws IOException { |
| // DataBlockScanner will run for every 5 seconds so we are checking for |
| // every 5 seconds |
| int n = 5; |
| String bpid = cluster.getNamesystem(0).getBlockPoolId(); |
| DataNode dn = cluster.getDataNodes().get(0); |
| long blocksScanned, total; |
| do { |
| try { |
| Thread.sleep(SLEEP_PERIOD_MS); |
| } catch (InterruptedException e) { |
| fail("Interrupted: " + e); |
| } |
| blocksScanned = dn.blockScanner.getBlocksScannedInLastRun(bpid); |
| total = dn.blockScanner.getTotalScans(bpid); |
| LOG.info("bpid = " + bpid + " blocksScanned = " + blocksScanned + " total=" + total); |
| } while (n-- > 0 && (blocksScanned != scansLastRun || scansTotal != total)); |
| Assert.assertEquals(scansTotal, total); |
| Assert.assertEquals(scansLastRun, blocksScanned); |
| } |
| } |