| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| */ |
| package org.apache.accumulo.test; |
| |
| import static org.junit.Assert.assertEquals; |
| import static org.junit.Assert.assertNotNull; |
| import static org.junit.Assert.assertTrue; |
| import static org.junit.Assert.fail; |
| |
| import java.io.File; |
| import java.io.IOException; |
| import java.security.SecureRandom; |
| import java.util.Arrays; |
| import java.util.HashMap; |
| import java.util.HashSet; |
| import java.util.Map; |
| import java.util.Map.Entry; |
| import java.util.Random; |
| import java.util.Set; |
| import java.util.SortedSet; |
| import java.util.TreeSet; |
| |
| import org.apache.accumulo.core.client.Accumulo; |
| import org.apache.accumulo.core.client.AccumuloClient; |
| import org.apache.accumulo.core.client.AccumuloException; |
| import org.apache.accumulo.core.client.BatchWriter; |
| import org.apache.accumulo.core.client.IteratorSetting; |
| import org.apache.accumulo.core.client.Scanner; |
| import org.apache.accumulo.core.client.TableNotFoundException; |
| import org.apache.accumulo.core.client.admin.CompactionConfig; |
| import org.apache.accumulo.core.client.admin.CompactionStrategyConfig; |
| import org.apache.accumulo.core.client.admin.NewTableConfiguration; |
| import org.apache.accumulo.core.clientImpl.ClientContext; |
| import org.apache.accumulo.core.conf.Property; |
| import org.apache.accumulo.core.data.Key; |
| import org.apache.accumulo.core.data.Mutation; |
| import org.apache.accumulo.core.data.Value; |
| import org.apache.accumulo.core.iterators.user.RegExFilter; |
| import org.apache.accumulo.core.security.Authorizations; |
| import org.apache.accumulo.harness.AccumuloClusterHarness; |
| import org.apache.accumulo.test.functional.ConfigurableCompactionIT; |
| import org.apache.accumulo.test.functional.FunctionalTestUtils; |
| import org.apache.accumulo.test.functional.SlowIterator; |
| import org.apache.commons.io.FileUtils; |
| import org.apache.hadoop.io.Text; |
| import org.junit.After; |
| import org.junit.Assume; |
| import org.junit.Test; |
| |
| import edu.umd.cs.findbugs.annotations.SuppressFBWarnings; |
| |
| @SuppressWarnings("removal") |
| public class UserCompactionStrategyIT extends AccumuloClusterHarness { |
| |
| @Override |
| public int defaultTimeoutSeconds() { |
| return 3 * 60; |
| } |
| |
| @After |
| public void checkForDanglingFateLocks() { |
| if (getClusterType() == ClusterType.MINI) { |
| try (AccumuloClient c = Accumulo.newClient().from(getClientProps()).build()) { |
| assertNotNull(c); |
| FunctionalTestUtils.assertNoDanglingFateLocks((ClientContext) c, getCluster()); |
| } |
| } |
| } |
| |
| @Test |
| public void testDropA() throws Exception { |
| try (AccumuloClient c = Accumulo.newClient().from(getClientProps()).build()) { |
| |
| String tableName = getUniqueNames(1)[0]; |
| c.tableOperations().create(tableName); |
| |
| writeFlush(c, tableName, "a"); |
| writeFlush(c, tableName, "b"); |
| // create a file that starts with A containing rows 'a' and 'b' |
| c.tableOperations().compact(tableName, new CompactionConfig().setWait(true)); |
| |
| writeFlush(c, tableName, "c"); |
| writeFlush(c, tableName, "d"); |
| |
| // drop files that start with A |
| CompactionStrategyConfig csConfig = |
| new CompactionStrategyConfig(TestCompactionStrategy.class.getName()); |
| csConfig.setOptions(Map.of("dropPrefix", "A", "inputPrefix", "F")); |
| c.tableOperations().compact(tableName, |
| new CompactionConfig().setWait(true).setCompactionStrategy(csConfig)); |
| |
| assertEquals(Set.of("c", "d"), getRows(c, tableName)); |
| |
| // this compaction should not drop files starting with A |
| c.tableOperations().compact(tableName, new CompactionConfig().setWait(true)); |
| c.tableOperations().compact(tableName, new CompactionConfig().setWait(true)); |
| |
| assertEquals(Set.of("c", "d"), getRows(c, tableName)); |
| } |
| } |
| |
| private void testDropNone(Map<String,String> options) throws Exception { |
| |
| try (AccumuloClient c = Accumulo.newClient().from(getClientProps()).build()) { |
| |
| String tableName = getUniqueNames(1)[0]; |
| c.tableOperations().create(tableName); |
| |
| writeFlush(c, tableName, "a"); |
| writeFlush(c, tableName, "b"); |
| |
| CompactionStrategyConfig csConfig = |
| new CompactionStrategyConfig(TestCompactionStrategy.class.getName()); |
| csConfig.setOptions(options); |
| c.tableOperations().compact(tableName, |
| new CompactionConfig().setWait(true).setCompactionStrategy(csConfig)); |
| |
| assertEquals(Set.of("a", "b"), getRows(c, tableName)); |
| } |
| } |
| |
| @Test |
| public void testDropNone() throws Exception { |
| // test a compaction strategy that selects no files. In this case there is no work to do, want |
| // to ensure it does not hang. |
| |
| testDropNone(Map.of("inputPrefix", "Z")); |
| } |
| |
| @Test |
| public void testDropNone2() throws Exception { |
| // test a compaction strategy that selects no files. This differs testDropNone() in that |
| // shouldCompact() will return true and getCompactionPlan() will |
| // return no work to do. |
| |
| testDropNone(Map.of("inputPrefix", "Z", "shouldCompact", "true")); |
| } |
| |
| @Test |
| public void testPerTableClasspath() throws Exception { |
| // Can't assume that a test-resource will be on the server's classpath |
| Assume.assumeTrue(getClusterType() == ClusterType.MINI); |
| |
| // test per-table classpath + user specified compaction strategy |
| |
| try (AccumuloClient c = Accumulo.newClient().from(getClientProps()).build()) { |
| final String tableName = getUniqueNames(1)[0]; |
| File target = new File(System.getProperty("user.dir"), "target"); |
| assertTrue(target.mkdirs() || target.isDirectory()); |
| File destFile = installJar(target, "/TestCompactionStrat.jar"); |
| c.instanceOperations().setProperty( |
| Property.VFS_CONTEXT_CLASSPATH_PROPERTY.getKey() + "context1", destFile.toString()); |
| HashMap<String,String> props = new HashMap<>(); |
| props.put(Property.TABLE_CLASSLOADER_CONTEXT.getKey(), "context1"); |
| SortedSet<Text> splits = new TreeSet<>(Arrays.asList(new Text("efg"))); |
| var ntc = new NewTableConfiguration().setProperties(props).withSplits(splits); |
| c.tableOperations().create(tableName, ntc); |
| |
| writeFlush(c, tableName, "a"); |
| writeFlush(c, tableName, "b"); |
| |
| writeFlush(c, tableName, "h"); |
| writeFlush(c, tableName, "i"); |
| |
| assertEquals(4, FunctionalTestUtils.countRFiles(c, tableName)); |
| |
| // EfgCompactionStrat will only compact a tablet w/ end row of 'efg'. No other tablets are |
| // compacted. |
| CompactionStrategyConfig csConfig = |
| new CompactionStrategyConfig("org.apache.accumulo.test.EfgCompactionStrat"); |
| c.tableOperations().compact(tableName, |
| new CompactionConfig().setWait(true).setCompactionStrategy(csConfig)); |
| |
| assertEquals(3, FunctionalTestUtils.countRFiles(c, tableName)); |
| |
| c.tableOperations().compact(tableName, new CompactionConfig().setWait(true)); |
| |
| assertEquals(2, FunctionalTestUtils.countRFiles(c, tableName)); |
| } |
| } |
| |
| @SuppressFBWarnings(value = "PATH_TRAVERSAL_IN", justification = "path provided by test") |
| private static File installJar(File destDir, String jarFile) throws IOException { |
| File destName = new File(destDir, new File(jarFile).getName()); |
| FileUtils.copyInputStreamToFile(ConfigurableCompactionIT.class.getResourceAsStream(jarFile), |
| destName); |
| return destName; |
| } |
| |
| @Test |
| public void testIterators() throws Exception { |
| // test compaction strategy + iterators |
| |
| try (AccumuloClient c = Accumulo.newClient().from(getClientProps()).build()) { |
| |
| String tableName = getUniqueNames(1)[0]; |
| c.tableOperations().create(tableName); |
| |
| writeFlush(c, tableName, "a"); |
| writeFlush(c, tableName, "b"); |
| // create a file that starts with A containing rows 'a' and 'b' |
| c.tableOperations().compact(tableName, new CompactionConfig().setWait(true)); |
| |
| writeFlush(c, tableName, "c"); |
| writeFlush(c, tableName, "d"); |
| |
| assertEquals(3, FunctionalTestUtils.countRFiles(c, tableName)); |
| |
| // drop files that start with A |
| CompactionStrategyConfig csConfig = |
| new CompactionStrategyConfig(TestCompactionStrategy.class.getName()); |
| csConfig.setOptions(Map.of("inputPrefix", "F")); |
| |
| IteratorSetting iterConf = new IteratorSetting(21, "myregex", RegExFilter.class); |
| RegExFilter.setRegexs(iterConf, "a|c", null, null, null, false); |
| |
| c.tableOperations().compact(tableName, new CompactionConfig().setWait(true) |
| .setCompactionStrategy(csConfig).setIterators(Arrays.asList(iterConf))); |
| |
| // compaction strategy should only be applied to one file. If its applied to both, then row |
| // 'b' |
| // would be dropped by filter. |
| assertEquals(Set.of("a", "b", "c"), getRows(c, tableName)); |
| |
| assertEquals(2, FunctionalTestUtils.countRFiles(c, tableName)); |
| |
| c.tableOperations().compact(tableName, new CompactionConfig().setWait(true)); |
| |
| // ensure that iterator is not applied |
| assertEquals(Set.of("a", "b", "c"), getRows(c, tableName)); |
| |
| assertEquals(1, FunctionalTestUtils.countRFiles(c, tableName)); |
| } |
| } |
| |
| @Test |
| public void testFileSize() throws Exception { |
| try (AccumuloClient c = Accumulo.newClient().from(getClientProps()).build()) { |
| |
| String tableName = getUniqueNames(1)[0]; |
| c.tableOperations().create(tableName); |
| |
| // write random data because its very unlikely it will compress |
| writeRandomValue(c, tableName, 1 << 16); |
| writeRandomValue(c, tableName, 1 << 16); |
| |
| writeRandomValue(c, tableName, 1 << 9); |
| writeRandomValue(c, tableName, 1 << 7); |
| writeRandomValue(c, tableName, 1 << 6); |
| |
| assertEquals(5, FunctionalTestUtils.countRFiles(c, tableName)); |
| |
| CompactionStrategyConfig csConfig = |
| new CompactionStrategyConfig(SizeCompactionStrategy.class.getName()); |
| csConfig.setOptions(Map.of("size", "" + (1 << 15))); |
| c.tableOperations().compact(tableName, |
| new CompactionConfig().setWait(true).setCompactionStrategy(csConfig)); |
| |
| assertEquals(3, FunctionalTestUtils.countRFiles(c, tableName)); |
| |
| csConfig = new CompactionStrategyConfig(SizeCompactionStrategy.class.getName()); |
| csConfig.setOptions(Map.of("size", "" + (1 << 17))); |
| c.tableOperations().compact(tableName, |
| new CompactionConfig().setWait(true).setCompactionStrategy(csConfig)); |
| |
| assertEquals(1, FunctionalTestUtils.countRFiles(c, tableName)); |
| } |
| |
| } |
| |
| @Test |
| public void testConcurrent() throws Exception { |
| // two compactions without iterators or strategy should be able to run concurrently |
| |
| try (AccumuloClient c = Accumulo.newClient().from(getClientProps()).build()) { |
| |
| String tableName = getUniqueNames(1)[0]; |
| c.tableOperations().create(tableName); |
| |
| // write random data because its very unlikely it will compress |
| writeRandomValue(c, tableName, 1 << 16); |
| writeRandomValue(c, tableName, 1 << 16); |
| |
| c.tableOperations().compact(tableName, new CompactionConfig().setWait(false)); |
| c.tableOperations().compact(tableName, new CompactionConfig().setWait(true)); |
| |
| assertEquals(1, FunctionalTestUtils.countRFiles(c, tableName)); |
| |
| writeRandomValue(c, tableName, 1 << 16); |
| |
| IteratorSetting iterConfig = new IteratorSetting(30, SlowIterator.class); |
| SlowIterator.setSleepTime(iterConfig, 1000); |
| |
| long t1 = System.currentTimeMillis(); |
| c.tableOperations().compact(tableName, |
| new CompactionConfig().setWait(false).setIterators(Arrays.asList(iterConfig))); |
| try { |
| // this compaction should fail because previous one set iterators |
| c.tableOperations().compact(tableName, new CompactionConfig().setWait(true)); |
| if (System.currentTimeMillis() - t1 < 2000) { |
| fail("Expected compaction to fail because another concurrent compaction set iterators"); |
| } |
| } catch (AccumuloException e) {} |
| } |
| } |
| |
| void writeRandomValue(AccumuloClient c, String tableName, int size) throws Exception { |
| Random rand = new SecureRandom(); |
| |
| byte[] data1 = new byte[size]; |
| rand.nextBytes(data1); |
| |
| try (BatchWriter bw = c.createBatchWriter(tableName)) { |
| Mutation m1 = new Mutation("r" + rand.nextInt(909090)); |
| m1.put("data", "bl0b", new Value(data1)); |
| bw.addMutation(m1); |
| } |
| c.tableOperations().flush(tableName, null, null, true); |
| } |
| |
| private Set<String> getRows(AccumuloClient c, String tableName) throws TableNotFoundException { |
| Set<String> rows = new HashSet<>(); |
| try (Scanner scanner = c.createScanner(tableName, Authorizations.EMPTY)) { |
| for (Entry<Key,Value> entry : scanner) { |
| rows.add(entry.getKey().getRowData().toString()); |
| } |
| } |
| return rows; |
| } |
| |
| private void writeFlush(AccumuloClient client, String tablename, String row) throws Exception { |
| try (BatchWriter bw = client.createBatchWriter(tablename)) { |
| Mutation m = new Mutation(row); |
| m.put("", "", ""); |
| bw.addMutation(m); |
| } |
| client.tableOperations().flush(tablename, null, null, true); |
| } |
| } |