blob: 32a7d0cddf55c951176ad53aeba5c7d8fb16850e [file] [log] [blame]
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
* <p/>
* http://www.apache.org/licenses/LICENSE-2.0
* <p/>
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.bigtop.itest.hbase.smoke;
import java.net.URL;
import java.util.Random;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.HColumnDescriptor;
import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.PerformanceEvaluation;
import org.apache.hadoop.hbase.client.HBaseAdmin;
import org.apache.hadoop.hbase.client.HTable;
import org.apache.hadoop.hbase.client.Result;
import org.apache.hadoop.hbase.client.ResultScanner;
import org.apache.hadoop.hbase.client.Scan;
import org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles;
import org.apache.hadoop.hbase.mapreduce.NMapInputFormat;
import org.apache.hadoop.hbase.util.Bytes;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertNotNull;
import static org.junit.Assert.assertTrue;
import org.junit.AfterClass;
import org.junit.Test;
import org.junit.Ignore;
import org.apache.bigtop.itest.JarContent;
import org.apache.bigtop.itest.shell.Shell;
import org.apache.bigtop.itest.hbase.util.HBaseTestUtil;
public class TestHFileOutputFormat {
private static final int ROWSPERSPLIT = 1024;
private static final byte[][] FAMILIES =
{Bytes.add(PerformanceEvaluation.FAMILY_NAME, Bytes.toBytes("-A")),
Bytes.add(PerformanceEvaluation.FAMILY_NAME, Bytes.toBytes("-B"))};
private static final String HBASE_HOME = System.getenv("HBASE_HOME");
private static final String HBASE_CONF_DIR = System.getenv("HBASE_CONF_DIR");
static {
assertNotNull("HBASE_HOME has to be set to run this test",
HBASE_HOME);
assertNotNull("HBASE_CONF_DIR has to be set to run this test",
HBASE_CONF_DIR);
}
private static String hbase_jar =
JarContent.getJarName(HBASE_HOME, "hbase-.*(?<!tests).jar");
private static String hbase_tests_jar =
JarContent.getJarName(HBASE_HOME, "hbase-.*tests.jar");
private static URL incrload_jar_url =
JarContent.getJarURL(org.apache.bigtop.itest.hbase.smoke.IncrementalPELoad.class);
static {
assertNotNull("Can't find hbase.jar", hbase_jar);
assertNotNull("Can't find hbase-tests.jar", hbase_tests_jar);
assertNotNull("Can't find jar containing IncrementalPELoad class", incrload_jar_url);
}
private static final String HBASE_JAR = HBASE_HOME + "/" + hbase_jar;
private static final String HBASE_TESTS_JAR = HBASE_HOME + "/" + hbase_tests_jar;
private static final String ZOOKEEPER_JAR = HBASE_HOME + "/lib/zookeeper.jar";
private static final String INCRLOAD_JAR = incrload_jar_url.getFile();
private static final String INCRLOAD = "org.apache.bigtop.itest.hbase.smoke.IncrementalPELoad";
private static final String USER = System.getProperty("user.name");
private static Shell sh = new Shell("/bin/bash -s");
@AfterClass
public static void cleanUp() {
// delete junk from HDFS
sh.exec("hadoop fs -rmr -skipTrash /user/" + USER + "/partitions_*");
assertTrue("HDFS cleanup failed", sh.getRet() == 0);
}
@Ignore("HBASE-1861")
@Test
public void testMRIncrementalLoad() throws Exception {
doIncrementalLoadTest("testMRIncrementalLoad", false);
}
@Ignore("HBASE-1861")
@Test
public void testMRIncrementalLoadWithSplit() throws Exception {
doIncrementalLoadTest("testMRIncrementalLoadWithSplit", true);
}
private byte[][] generateRandomSplitKeys(int numKeys) {
Random random = new Random();
byte[][] ret = new byte[numKeys][];
for (int i = 0; i < numKeys; i++) {
ret[i] = PerformanceEvaluation.generateData(random, 10);
}
return ret;
}
private void doIncrementalLoadTest(String testName, boolean shouldChangeRegions)
throws Exception {
FileSystem fs = HBaseTestUtil.getClusterFileSystem();
Path testDir = HBaseTestUtil.getMROutputDir(testName);
byte[][] splitKeys = generateRandomSplitKeys(4);
Configuration conf = HBaseConfiguration.create();
HBaseAdmin admin = new HBaseAdmin(conf);
final byte[] TABLE_NAME = HBaseTestUtil.getTestTableName(testName);
HTableDescriptor tbldesc = new HTableDescriptor(TABLE_NAME);
HColumnDescriptor coldesc1 = new HColumnDescriptor(FAMILIES[0]);
HColumnDescriptor coldesc2 = new HColumnDescriptor(FAMILIES[1]);
tbldesc.addFamily(coldesc1);
tbldesc.addFamily(coldesc2);
admin.createTable(tbldesc, splitKeys);
HTable table = new HTable(conf, TABLE_NAME);
assertEquals("Should start with empty table",
0, HBaseTestUtil.countRows(table));
// Generate the bulk load files
runIncrementalPELoad(Bytes.toString(TABLE_NAME), testDir.toString());
// This doesn't write into the table, just makes files
assertEquals("HFOF should not touch actual table",
0, HBaseTestUtil.countRows(table));
// Make sure that a directory was created for every CF
int dir = 0;
for (FileStatus f : fs.listStatus(testDir)) {
for (byte[] family : FAMILIES) {
if (Bytes.toString(family).equals(f.getPath().getName())) {
++dir;
}
}
}
assertEquals("Column family not found in FS.", FAMILIES.length, dir);
// handle the split case
if (shouldChangeRegions) {
admin.disableTable(TABLE_NAME);
admin.deleteTable(TABLE_NAME);
byte[][] newSplitKeys = generateRandomSplitKeys(14);
admin.createTable(tbldesc, newSplitKeys);
}
// Before we can load the HFiles, we need to set the permissions so that
// HBase has write access to testDir's contents
chmod(testDir.toString());
// Perform the actual load
new LoadIncrementalHFiles(conf).doBulkLoad(testDir, table);
// Ensure data shows up
int expectedRows = NMapInputFormat.getNumMapTasks(conf) * ROWSPERSPLIT;
assertEquals("LoadIncrementalHFiles should put expected data in table",
expectedRows, HBaseTestUtil.countRows(table));
Scan scan = new Scan();
ResultScanner results = table.getScanner(scan);
int count = 0;
for (Result res : results) {
count++;
assertEquals(FAMILIES.length, res.raw().length);
KeyValue first = res.raw()[0];
for (KeyValue kv : res.raw()) {
assertTrue(KeyValue.COMPARATOR.matchingRows(first, kv));
assertTrue(Bytes.equals(first.getValue(), kv.getValue()));
}
}
results.close();
String tableDigestBefore = HBaseTestUtil.checksumRows(table);
// Cause regions to reopen
admin.disableTable(TABLE_NAME);
admin.enableTable(TABLE_NAME);
assertEquals("Data should remain after reopening of regions",
tableDigestBefore, HBaseTestUtil.checksumRows(table));
// cleanup
// - disable and drop table
admin.disableTable(TABLE_NAME);
admin.deleteTable(TABLE_NAME);
// - remove incremental load output
fs.delete(testDir, true);
}
private void chmod(String uri) {
sh.exec("hadoop fs -chmod -R 777 " + uri);
assertEquals("chmod failed", 0, sh.getRet());
}
private void runIncrementalPELoad(String table, String outDir) {
sh.exec("export HADOOP_CLASSPATH=" + HBASE_CONF_DIR + ":" + HBASE_JAR + ":" + HBASE_TESTS_JAR + ":" + ZOOKEEPER_JAR,
"hadoop jar " + INCRLOAD_JAR + " " + INCRLOAD +
" -libjars " + HBASE_JAR + "," + HBASE_TESTS_JAR +
" " + table + " " + outDir);
assertEquals("MR job failed", 0, sh.getRet());
}
}