| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.cassandra.utils; |
| |
| import java.io.IOException; |
| import java.math.BigInteger; |
| import java.nio.ByteBuffer; |
| import java.util.*; |
| |
| import org.junit.Assert; |
| import org.junit.Before; |
| import org.junit.Test; |
| import org.apache.cassandra.config.DatabaseDescriptor; |
| import org.apache.cassandra.db.Digest; |
| import org.apache.cassandra.dht.ByteOrderedPartitioner; |
| import org.apache.cassandra.dht.IPartitioner; |
| import org.apache.cassandra.dht.Murmur3Partitioner; |
| import org.apache.cassandra.dht.RandomPartitioner; |
| import org.apache.cassandra.dht.RandomPartitioner.BigIntegerToken; |
| import org.apache.cassandra.dht.Range; |
| import org.apache.cassandra.dht.Token; |
| import org.apache.cassandra.io.util.DataInputBuffer; |
| import org.apache.cassandra.io.util.DataOutputBuffer; |
| import org.apache.cassandra.net.MessagingService; |
| import org.apache.cassandra.utils.MerkleTree.RowHash; |
| import org.apache.cassandra.utils.MerkleTree.TreeRange; |
| import org.apache.cassandra.utils.MerkleTree.TreeRangeIterator; |
| |
| import static com.google.common.collect.Lists.newArrayList; |
| import static org.apache.cassandra.utils.MerkleTree.RECOMMENDED_DEPTH; |
| import static org.junit.Assert.*; |
| |
| public class MerkleTreeTest |
| { |
| private static final byte[] DUMMY = digest("dummy"); |
| |
| static byte[] digest(String string) |
| { |
| return Digest.forValidator() |
| .update(string.getBytes(), 0, string.getBytes().length) |
| .digest(); |
| } |
| |
| /** |
| * If a test assumes that the tree is 8 units wide, then it should set this value |
| * to 8. |
| */ |
| public static BigInteger TOKEN_SCALE = new BigInteger("8"); |
| |
| protected IPartitioner partitioner; |
| protected MerkleTree mt; |
| |
| private Range<Token> fullRange() |
| { |
| return new Range<>(partitioner.getMinimumToken(), partitioner.getMinimumToken()); |
| } |
| |
| @Before |
| public void setup() |
| { |
| DatabaseDescriptor.clientInitialization(); |
| DatabaseDescriptor.useOffheapMerkleTrees(false); |
| |
| TOKEN_SCALE = new BigInteger("8"); |
| partitioner = RandomPartitioner.instance; |
| // TODO need to trickle TokenSerializer |
| DatabaseDescriptor.setPartitionerUnsafe(partitioner); |
| mt = new MerkleTree(partitioner, fullRange(), RECOMMENDED_DEPTH, Integer.MAX_VALUE); |
| } |
| |
| public static void assertHashEquals(final byte[] left, final byte[] right) |
| { |
| assertHashEquals("", left, right); |
| } |
| |
| public static void assertHashEquals(String message, final byte[] left, final byte[] right) |
| { |
| String lstring = left == null ? "null" : Hex.bytesToHex(left); |
| String rstring = right == null ? "null" : Hex.bytesToHex(right); |
| assertEquals(message, lstring, rstring); |
| } |
| |
| /** |
| * The value returned by this method is affected by TOKEN_SCALE: setting TOKEN_SCALE |
| * to 8 means that passing -1 through 8 for this method will return values mapped |
| * between -1 and Token.MAX_VALUE. |
| */ |
| public static Token tok(int i) |
| { |
| if (i == -1) |
| return new BigIntegerToken(new BigInteger("-1")); |
| BigInteger bint = RandomPartitioner.MAXIMUM.divide(TOKEN_SCALE).multiply(new BigInteger("" + i)); |
| return new BigIntegerToken(bint); |
| } |
| |
| @Test |
| public void testSplit() |
| { |
| // split the range (zero, zero] into: |
| // (zero,four], (four,six], (six,seven] and (seven, zero] |
| mt.split(tok(4)); |
| mt.split(tok(6)); |
| mt.split(tok(7)); |
| |
| assertEquals(4, mt.size()); |
| assertEquals(new Range<>(tok(7), tok(-1)), mt.get(tok(-1))); |
| assertEquals(new Range<>(tok(-1), tok(4)), mt.get(tok(3))); |
| assertEquals(new Range<>(tok(-1), tok(4)), mt.get(tok(4))); |
| assertEquals(new Range<>(tok(4), tok(6)), mt.get(tok(6))); |
| assertEquals(new Range<>(tok(6), tok(7)), mt.get(tok(7))); |
| |
| // check depths |
| assertEquals((byte) 1, mt.get(tok(4)).depth); |
| assertEquals((byte) 2, mt.get(tok(6)).depth); |
| assertEquals((byte) 3, mt.get(tok(7)).depth); |
| assertEquals((byte) 3, mt.get(tok(-1)).depth); |
| |
| try |
| { |
| mt.split(tok(-1)); |
| fail("Shouldn't be able to split outside the initial range."); |
| } |
| catch (AssertionError e) |
| { |
| // pass |
| } |
| } |
| |
| @Test |
| public void testSplitLimitDepth() |
| { |
| mt = new MerkleTree(partitioner, fullRange(), (byte) 2, Integer.MAX_VALUE); |
| |
| assertTrue(mt.split(tok(4))); |
| assertTrue(mt.split(tok(2))); |
| assertEquals(3, mt.size()); |
| |
| // should fail to split below hashdepth |
| assertFalse(mt.split(tok(1))); |
| assertEquals(3, mt.size()); |
| assertEquals(new Range<>(tok(4), tok(-1)), mt.get(tok(-1))); |
| assertEquals(new Range<>(tok(-1), tok(2)), mt.get(tok(2))); |
| assertEquals(new Range<>(tok(2), tok(4)), mt.get(tok(4))); |
| } |
| |
| @Test |
| public void testSplitLimitSize() |
| { |
| mt = new MerkleTree(partitioner, fullRange(), RECOMMENDED_DEPTH, 2); |
| |
| assertTrue(mt.split(tok(4))); |
| assertEquals(2, mt.size()); |
| |
| // should fail to split above maxsize |
| assertFalse(mt.split(tok(2))); |
| assertEquals(2, mt.size()); |
| assertEquals(new Range<>(tok(4), tok(-1)), mt.get(tok(-1))); |
| assertEquals(new Range<>(tok(-1), tok(4)), mt.get(tok(4))); |
| } |
| |
| @Test |
| public void testInvalids() |
| { |
| Iterator<TreeRange> ranges; |
| |
| // (zero, zero] |
| ranges = mt.rangeIterator(); |
| assertEquals(new Range<>(tok(-1), tok(-1)), ranges.next()); |
| assertFalse(ranges.hasNext()); |
| |
| // all invalid |
| mt.split(tok(4)); |
| mt.split(tok(2)); |
| mt.split(tok(6)); |
| mt.split(tok(3)); |
| mt.split(tok(5)); |
| ranges = mt.rangeIterator(); |
| assertEquals(new Range<>(tok(6), tok(-1)), ranges.next()); |
| assertEquals(new Range<>(tok(-1), tok(2)), ranges.next()); |
| assertEquals(new Range<>(tok(2), tok(3)), ranges.next()); |
| assertEquals(new Range<>(tok(3), tok(4)), ranges.next()); |
| assertEquals(new Range<>(tok(4), tok(5)), ranges.next()); |
| assertEquals(new Range<>(tok(5), tok(6)), ranges.next()); |
| assertEquals(new Range<>(tok(6), tok(-1)), ranges.next()); |
| assertFalse(ranges.hasNext()); |
| } |
| |
| |
| @Test |
| public void testHashFull() |
| { |
| byte[] val = DUMMY; |
| Range<Token> range = new Range<>(tok(-1), tok(-1)); |
| |
| // (zero, zero] |
| assertFalse(mt.hashesRange(range)); |
| |
| // validate the range |
| mt.get(tok(-1)).hash(val); |
| |
| assertHashEquals(val, mt.hash(range)); |
| } |
| |
| @Test |
| public void testHashPartial() |
| { |
| byte[] val = DUMMY; |
| byte[] leftval = hashed(val, 1, 1); |
| byte[] partialval = hashed(val, 1); |
| Range<Token> left = new Range<>(tok(-1), tok(4)); |
| Range<Token> partial = new Range<>(tok(2), tok(4)); |
| Range<Token> right = new Range<>(tok(4), tok(-1)); |
| Range<Token> linvalid = new Range<>(tok(1), tok(4)); |
| Range<Token> rinvalid = new Range<>(tok(4), tok(6)); |
| |
| // (zero,two] (two,four] (four, zero] |
| mt.split(tok(4)); |
| mt.split(tok(2)); |
| |
| assertFalse(mt.hashesRange(left)); |
| assertFalse(mt.hashesRange(partial)); |
| assertFalse(mt.hashesRange(right)); |
| assertFalse(mt.hashesRange(linvalid)); |
| assertFalse(mt.hashesRange(rinvalid)); |
| |
| // validate the range |
| mt.get(tok(2)).hash(val); |
| mt.get(tok(4)).hash(val); |
| mt.get(tok(-1)).hash(val); |
| |
| assertHashEquals(leftval, mt.hash(left)); |
| assertHashEquals(partialval, mt.hash(partial)); |
| assertHashEquals(val, mt.hash(right)); |
| assertFalse(mt.hashesRange(linvalid)); |
| assertFalse(mt.hashesRange(rinvalid)); |
| } |
| |
| @Test |
| public void testHashInner() |
| { |
| byte[] val = DUMMY; |
| byte[] lchildval = hashed(val, 3, 3, 2); |
| byte[] rchildval = hashed(val, 2, 2); |
| byte[] fullval = hashed(val, 3, 3, 2, 2, 2); |
| Range<Token> full = new Range<>(tok(-1), tok(-1)); |
| Range<Token> lchild = new Range<>(tok(-1), tok(4)); |
| Range<Token> rchild = new Range<>(tok(4), tok(-1)); |
| Range<Token> invalid = new Range<>(tok(1), tok(-1)); |
| |
| // (zero,one] (one, two] (two,four] (four, six] (six, zero] |
| mt.split(tok(4)); |
| mt.split(tok(2)); |
| mt.split(tok(6)); |
| mt.split(tok(1)); |
| |
| // validate the range |
| mt.get(tok(1)).hash(val); |
| mt.get(tok(2)).hash(val); |
| mt.get(tok(4)).hash(val); |
| mt.get(tok(6)).hash(val); |
| mt.get(tok(-1)).hash(val); |
| |
| assertTrue(mt.hashesRange(full)); |
| assertTrue(mt.hashesRange(lchild)); |
| assertTrue(mt.hashesRange(rchild)); |
| assertFalse(mt.hashesRange(invalid)); |
| |
| assertHashEquals(fullval, mt.hash(full)); |
| assertHashEquals(lchildval, mt.hash(lchild)); |
| assertHashEquals(rchildval, mt.hash(rchild)); |
| } |
| |
| @Test |
| public void testHashDegenerate() |
| { |
| TOKEN_SCALE = new BigInteger("32"); |
| |
| byte[] val = DUMMY; |
| byte[] childfullval = hashed(val, 5, 5, 4); |
| byte[] fullval = hashed(val, 5, 5, 4, 3, 2, 1); |
| Range<Token> childfull = new Range<>(tok(-1), tok(4)); |
| Range<Token> full = new Range<>(tok(-1), tok(-1)); |
| Range<Token> invalid = new Range<>(tok(4), tok(-1)); |
| |
| mt = new MerkleTree(partitioner, fullRange(), RECOMMENDED_DEPTH, Integer.MAX_VALUE); |
| mt.split(tok(16)); |
| mt.split(tok(8)); |
| mt.split(tok(4)); |
| mt.split(tok(2)); |
| mt.split(tok(1)); |
| |
| // validate the range |
| mt.get(tok(1)).hash(val); |
| mt.get(tok(2)).hash(val); |
| mt.get(tok(4)).hash(val); |
| mt.get(tok(8)).hash(val); |
| mt.get(tok(16)).hash(val); |
| mt.get(tok(-1)).hash(val); |
| |
| assertTrue(mt.hashesRange(full)); |
| assertTrue(mt.hashesRange(childfull)); |
| assertFalse(mt.hashesRange(invalid)); |
| |
| assertHashEquals(fullval, mt.hash(full)); |
| assertHashEquals(childfullval, mt.hash(childfull)); |
| } |
| |
| @Test |
| public void testHashRandom() |
| { |
| int max = 1000000; |
| TOKEN_SCALE = new BigInteger("" + max); |
| |
| mt = new MerkleTree(partitioner, fullRange(), RECOMMENDED_DEPTH, 32); |
| Random random = new Random(); |
| while (true) |
| { |
| if (!mt.split(tok(random.nextInt(max)))) |
| break; |
| } |
| |
| // validate the tree |
| TreeRangeIterator ranges = mt.rangeIterator(); |
| for (TreeRange range : ranges) |
| range.addHash(new RowHash(range.right, new byte[0], 0)); |
| |
| assert mt.hash(new Range<>(tok(-1), tok(-1))) != null : |
| "Could not hash tree " + mt; |
| } |
| |
| /** |
| * Generate two trees with different splits, but containing the same keys, and |
| * check that they compare equally. |
| * |
| * The set of keys used in this test is: #{2,4,6,8,12,14,0} |
| */ |
| @Test |
| public void testValidateTree() |
| { |
| TOKEN_SCALE = new BigInteger("16"); // this test needs slightly more resolution |
| |
| Range<Token> full = new Range<>(tok(-1), tok(-1)); |
| Iterator<TreeRange> ranges; |
| MerkleTree mt2 = new MerkleTree(partitioner, fullRange(), RECOMMENDED_DEPTH, Integer.MAX_VALUE); |
| |
| mt.split(tok(8)); |
| mt.split(tok(4)); |
| mt.split(tok(12)); |
| mt.split(tok(6)); |
| mt.split(tok(10)); |
| |
| ranges = mt.rangeIterator(); |
| ranges.next().addAll(new HIterator(2, 4)); // (-1,4]: depth 2 |
| ranges.next().addAll(new HIterator(6)); // (4,6] |
| ranges.next().addAll(new HIterator(8)); // (6,8] |
| ranges.next().addAll(new HIterator(/*empty*/ new int[0])); // (8,10] |
| ranges.next().addAll(new HIterator(12)); // (10,12] |
| ranges.next().addAll(new HIterator(14, -1)); // (12,-1]: depth 2 |
| |
| |
| mt2.split(tok(8)); |
| mt2.split(tok(4)); |
| mt2.split(tok(12)); |
| mt2.split(tok(2)); |
| mt2.split(tok(10)); |
| mt2.split(tok(9)); |
| mt2.split(tok(11)); |
| |
| ranges = mt2.rangeIterator(); |
| ranges.next().addAll(new HIterator(2)); // (-1,2] |
| ranges.next().addAll(new HIterator(4)); // (2,4] |
| ranges.next().addAll(new HIterator(6, 8)); // (4,8]: depth 2 |
| ranges.next().addAll(new HIterator(/*empty*/ new int[0])); // (8,9] |
| ranges.next().addAll(new HIterator(/*empty*/ new int[0])); // (9,10] |
| ranges.next().addAll(new HIterator(/*empty*/ new int[0])); // (10,11]: depth 4 |
| ranges.next().addAll(new HIterator(12)); // (11,12]: depth 4 |
| ranges.next().addAll(new HIterator(14, -1)); // (12,-1]: depth 2 |
| |
| byte[] mthash = mt.hash(full); |
| byte[] mt2hash = mt2.hash(full); |
| assertHashEquals("Tree hashes did not match: " + mt + " && " + mt2, mthash, mt2hash); |
| } |
| |
| @Test |
| public void testSerialization() throws Exception |
| { |
| Range<Token> full = new Range<>(tok(-1), tok(-1)); |
| |
| // populate and validate the tree |
| mt.maxsize(256); |
| mt.init(); |
| for (TreeRange range : mt.rangeIterator()) |
| range.addAll(new HIterator(range.right)); |
| |
| byte[] initialhash = mt.hash(full); |
| |
| DataOutputBuffer out = new DataOutputBuffer(); |
| mt.serialize(out, MessagingService.current_version); |
| byte[] serialized = out.toByteArray(); |
| |
| MerkleTree restoredOnHeap = |
| MerkleTree.deserialize(new DataInputBuffer(serialized), false, MessagingService.current_version); |
| MerkleTree restoredOffHeap = |
| MerkleTree.deserialize(new DataInputBuffer(serialized), true, MessagingService.current_version); |
| MerkleTree movedOffHeap = mt.moveOffHeap(); |
| |
| assertHashEquals(initialhash, restoredOnHeap.hash(full)); |
| assertHashEquals(initialhash, restoredOffHeap.hash(full)); |
| assertHashEquals(initialhash, movedOffHeap.hash(full)); |
| |
| assertEquals(mt, restoredOnHeap); |
| assertEquals(mt, restoredOffHeap); |
| assertEquals(mt, movedOffHeap); |
| |
| assertEquals(restoredOnHeap, restoredOffHeap); |
| assertEquals(restoredOnHeap, movedOffHeap); |
| |
| assertEquals(restoredOffHeap, movedOffHeap); |
| } |
| |
| @Test |
| public void testDifference() |
| { |
| int maxsize = 16; |
| mt.maxsize(maxsize); |
| MerkleTree mt2 = new MerkleTree(partitioner, fullRange(), RECOMMENDED_DEPTH, maxsize); |
| mt.init(); |
| mt2.init(); |
| |
| // add dummy hashes to both trees |
| for (TreeRange range : mt.rangeIterator()) |
| range.addAll(new HIterator(range.right)); |
| for (TreeRange range : mt2.rangeIterator()) |
| range.addAll(new HIterator(range.right)); |
| |
| TreeRange leftmost = null; |
| TreeRange middle = null; |
| |
| mt.maxsize(maxsize + 2); // give some room for splitting |
| |
| // split the leftmost |
| Iterator<TreeRange> ranges = mt.rangeIterator(); |
| leftmost = ranges.next(); |
| mt.split(leftmost.right); |
| |
| // set the hashes for the leaf of the created split |
| middle = mt.get(leftmost.right); |
| middle.hash(digest("arbitrary!")); |
| mt.get(partitioner.midpoint(leftmost.left, leftmost.right)).hash(digest("even more arbitrary!")); |
| |
| // trees should disagree for (leftmost.left, middle.right] |
| List<TreeRange> diffs = MerkleTree.difference(mt, mt2); |
| assertEquals(diffs + " contains wrong number of differences:", 1, diffs.size()); |
| assertTrue(diffs.contains(new Range<>(leftmost.left, middle.right))); |
| } |
| |
| /** |
| * difference should behave as expected, even with extremely small ranges |
| */ |
| @Test |
| public void differenceSmallRange() |
| { |
| Token start = new BigIntegerToken("9"); |
| Token end = new BigIntegerToken("10"); |
| Range<Token> range = new Range<>(start, end); |
| |
| MerkleTree ltree = new MerkleTree(partitioner, range, RECOMMENDED_DEPTH, 16); |
| ltree.init(); |
| MerkleTree rtree = new MerkleTree(partitioner, range, RECOMMENDED_DEPTH, 16); |
| rtree.init(); |
| |
| byte[] h1 = digest("asdf"); |
| byte[] h2 = digest("hjkl"); |
| |
| // add dummy hashes to both trees |
| for (TreeRange tree : ltree.rangeIterator()) |
| { |
| tree.addHash(new RowHash(range.right, h1, h1.length)); |
| } |
| for (TreeRange tree : rtree.rangeIterator()) |
| { |
| tree.addHash(new RowHash(range.right, h2, h2.length)); |
| } |
| |
| List<TreeRange> diffs = MerkleTree.difference(ltree, rtree); |
| assertEquals(newArrayList(range), diffs); |
| assertEquals(MerkleTree.Difference.FULLY_INCONSISTENT, |
| MerkleTree.differenceHelper(ltree, rtree, new ArrayList<>(), new MerkleTree.TreeRange(ltree.fullRange.left, ltree.fullRange.right, (byte)0))); |
| } |
| |
| /** |
| * matching should behave as expected, even with extremely small ranges |
| */ |
| @Test |
| public void matchingSmallRange() |
| { |
| Token start = new BigIntegerToken("9"); |
| Token end = new BigIntegerToken("10"); |
| Range<Token> range = new Range<>(start, end); |
| |
| MerkleTree ltree = new MerkleTree(partitioner, range, RECOMMENDED_DEPTH, 16); |
| ltree.init(); |
| MerkleTree rtree = new MerkleTree(partitioner, range, RECOMMENDED_DEPTH, 16); |
| rtree.init(); |
| |
| byte[] h1 = digest("asdf"); |
| byte[] h2 = digest("asdf"); |
| |
| |
| // add dummy hashes to both trees |
| for (TreeRange tree : ltree.rangeIterator()) |
| { |
| tree.addHash(new RowHash(range.right, h1, h1.length)); |
| } |
| for (TreeRange tree : rtree.rangeIterator()) |
| { |
| tree.addHash(new RowHash(range.right, h2, h2.length)); |
| } |
| |
| // top level difference() should show no differences |
| assertEquals(MerkleTree.difference(ltree, rtree), newArrayList()); |
| } |
| |
| /** |
| * Return the root hash of a binary tree with leaves at the given depths |
| * and with the given hash val in each leaf. |
| */ |
| byte[] hashed(byte[] val, Integer... depths) |
| { |
| ArrayDeque<Integer> dstack = new ArrayDeque<Integer>(); |
| ArrayDeque<byte[]> hstack = new ArrayDeque<byte[]>(); |
| Iterator<Integer> depthiter = Arrays.asList(depths).iterator(); |
| if (depthiter.hasNext()) |
| { |
| dstack.push(depthiter.next()); |
| hstack.push(val); |
| } |
| while (depthiter.hasNext()) |
| { |
| Integer depth = depthiter.next(); |
| byte[] hash = val; |
| while (depth.equals(dstack.peek())) |
| { |
| // consume the stack |
| hash = MerkleTree.xor(hstack.pop(), hash); |
| depth = dstack.pop() - 1; |
| } |
| dstack.push(depth); |
| hstack.push(hash); |
| } |
| assert hstack.size() == 1; |
| return hstack.pop(); |
| } |
| |
| static class HIterator extends AbstractIterator<RowHash> |
| { |
| private Iterator<Token> tokens; |
| |
| public HIterator(int... tokens) |
| { |
| List<Token> tlist = new LinkedList<Token>(); |
| for (int token : tokens) |
| tlist.add(tok(token)); |
| this.tokens = tlist.iterator(); |
| } |
| |
| public HIterator(Token... tokens) |
| { |
| this.tokens = Arrays.asList(tokens).iterator(); |
| } |
| |
| public RowHash computeNext() |
| { |
| if (tokens.hasNext()) |
| return new RowHash(tokens.next(), DUMMY, DUMMY.length); |
| return endOfData(); |
| } |
| } |
| |
| @Test |
| public void testEstimatedSizes() |
| { |
| // With no or negative allowed space we should still get a depth of 1 |
| Assert.assertEquals(1, MerkleTree.estimatedMaxDepthForBytes(Murmur3Partitioner.instance, -20, 32)); |
| Assert.assertEquals(1, MerkleTree.estimatedMaxDepthForBytes(Murmur3Partitioner.instance, 0, 32)); |
| Assert.assertEquals(1, MerkleTree.estimatedMaxDepthForBytes(Murmur3Partitioner.instance, 1, 32)); |
| |
| // The minimum of 1 megabyte split between RF=3 should yield trees of around 10 |
| Assert.assertEquals(10, MerkleTree.estimatedMaxDepthForBytes(Murmur3Partitioner.instance, |
| 1048576 / 3, 32)); |
| |
| // With a single megabyte of space we should get 12 |
| Assert.assertEquals(12, MerkleTree.estimatedMaxDepthForBytes(Murmur3Partitioner.instance, |
| 1048576, 32)); |
| |
| // With 100 megabytes we should get a limit of 19 |
| Assert.assertEquals(19, MerkleTree.estimatedMaxDepthForBytes(Murmur3Partitioner.instance, |
| 100 * 1048576, 32)); |
| |
| // With 300 megabytes we should get the old limit of 20 |
| Assert.assertEquals(20, MerkleTree.estimatedMaxDepthForBytes(Murmur3Partitioner.instance, |
| 300 * 1048576, 32)); |
| Assert.assertEquals(20, MerkleTree.estimatedMaxDepthForBytes(RandomPartitioner.instance, |
| 300 * 1048576, 32)); |
| Assert.assertEquals(20, MerkleTree.estimatedMaxDepthForBytes(ByteOrderedPartitioner.instance, |
| 300 * 1048576, 32)); |
| } |
| |
| @Test |
| public void testEstimatedSizesRealMeasurement() |
| { |
| // Use a fixed source of randomness so that the test does not flake. |
| Random random = new Random(1); |
| checkEstimatedSizes(RandomPartitioner.instance, random); |
| checkEstimatedSizes(Murmur3Partitioner.instance, random); |
| } |
| |
| private void checkEstimatedSizes(IPartitioner partitioner, Random random) |
| { |
| Range<Token> fullRange = new Range<>(partitioner.getMinimumToken(), partitioner.getMinimumToken()); |
| MerkleTree tree = new MerkleTree(partitioner, fullRange, RECOMMENDED_DEPTH, 0); |
| |
| // Test 16 kilobyte -> 16 megabytes |
| for (int i = 14; i < 24; i ++) |
| { |
| long numBytes = 1 << i; |
| int maxDepth = MerkleTree.estimatedMaxDepthForBytes(partitioner, numBytes, 32); |
| long realSizeOfMerkleTree = measureTree(tree, fullRange, maxDepth, random); |
| long biggerTreeSize = measureTree(tree, fullRange, maxDepth + 1, random); |
| |
| Assert.assertTrue(realSizeOfMerkleTree < numBytes); |
| Assert.assertTrue(biggerTreeSize > numBytes); |
| } |
| } |
| |
| private long measureTree(MerkleTree tree, Range<Token> fullRange, int depth, Random random) |
| { |
| tree = new MerkleTree(tree.partitioner(), fullRange, RECOMMENDED_DEPTH, (long) Math.pow(2, depth)); |
| // Initializes it as a fully balanced tree. |
| tree.init(); |
| |
| byte[] key = new byte[128]; |
| // Try to actually allocate some hashes. Note that this is not guaranteed to actually populate the tree, |
| // but we re-use the source of randomness to try to make it reproducible. |
| for (int i = 0; i < tree.maxsize() * 8; i++) |
| { |
| random.nextBytes(key); |
| Token token = tree.partitioner().getToken(ByteBuffer.wrap(key)); |
| tree.get(token).addHash(new RowHash(token, new byte[32], 32)); |
| } |
| |
| tree.hash(fullRange); |
| return ObjectSizes.measureDeep(tree); |
| } |
| |
| @Test |
| public void testEqualTreesSameDepth() throws IOException |
| { |
| int seed = makeSeed(); |
| Trees trees = Trees.make(seed, seed, 3, 3); |
| testDifferences(trees, Collections.emptyList()); |
| } |
| |
| @Test |
| public void testEqualTreesDifferentDepth() throws IOException |
| { |
| int seed = makeSeed(); |
| Trees trees = Trees.make(seed, seed, 2, 3); |
| testDifferences(trees, Collections.emptyList()); |
| } |
| |
| @Test |
| public void testEntirelyDifferentTrees() throws IOException |
| { |
| int seed1 = makeSeed(); |
| int seed2 = seed1 * 32; |
| Trees trees = Trees.make(seed1, seed2, 3, 3); |
| testDifferences(trees, newArrayList(makeTreeRange(0, 16, 0))); |
| } |
| |
| @Test |
| public void testDifferentTrees1SameDepth() throws IOException |
| { |
| int seed = makeSeed(); |
| Trees trees = Trees.make(seed, seed, 3, 3); |
| trees.tree1.get(longToken(1)).addHash(digest("diff_1"), 1); |
| testDifferences(trees, newArrayList(makeTreeRange(0, 2, 3))); |
| } |
| |
| @Test |
| public void testDifferentTrees1DifferentDepth() throws IOException |
| { |
| int seed = makeSeed(); |
| Trees trees = Trees.make(seed, seed, 2, 3); |
| trees.tree1.get(longToken(1)).addHash(digest("diff_1"), 1); |
| testDifferences(trees, newArrayList(makeTreeRange(0, 4, 2))); |
| } |
| |
| @Test |
| public void testDifferentTrees2SameDepth() throws IOException |
| { |
| int seed = makeSeed(); |
| Trees trees = Trees.make(seed, seed, 3, 3); |
| trees.tree1.get(longToken(1)).addHash(digest("diff_1"), 1); |
| trees.tree2.get(longToken(16)).addHash(digest("diff_16"), 1); |
| testDifferences(trees, newArrayList(makeTreeRange(0, 2, 3), |
| makeTreeRange(14, 16, 3))); |
| } |
| |
| @Test |
| public void testDifferentTrees2DifferentDepth() throws IOException |
| { |
| int seed = makeSeed(); |
| Trees trees = Trees.make(seed, seed, 2, 3); |
| trees.tree1.get(longToken(1)).addHash(digest("diff_1"), 1); |
| trees.tree2.get(longToken(16)).addHash(digest("diff_16"), 1); |
| testDifferences(trees, newArrayList(makeTreeRange(0, 4, 2), |
| makeTreeRange(12, 16, 2))); |
| } |
| |
| @Test |
| public void testDifferentTrees3SameDepth() throws IOException |
| { |
| int seed = makeSeed(); |
| Trees trees = Trees.make(seed, seed, 3, 3); |
| trees.tree1.get(longToken(1)).addHash(digest("diff_1"), 1); |
| trees.tree1.get(longToken(3)).addHash(digest("diff_3"), 1); |
| testDifferences(trees, newArrayList(makeTreeRange(0, 4, 2))); |
| } |
| |
| @Test |
| public void testDifferentTrees3Differentepth() throws IOException |
| { |
| int seed = makeSeed(); |
| Trees trees = Trees.make(seed, seed, 2, 3); |
| trees.tree1.get(longToken(1)).addHash(digest("diff_1"), 1); |
| trees.tree1.get(longToken(3)).addHash(digest("diff_3"), 1); |
| testDifferences(trees, newArrayList(makeTreeRange(0, 4, 2))); |
| } |
| |
| @Test |
| public void testDifferentTrees4SameDepth() throws IOException |
| { |
| int seed = makeSeed(); |
| Trees trees = Trees.make(seed, seed, 3, 3); |
| trees.tree1.get(longToken(4)).addHash(digest("diff_4"), 1); |
| trees.tree1.get(longToken(8)).addHash(digest("diff_8"), 1); |
| trees.tree1.get(longToken(12)).addHash(digest("diff_12"), 1); |
| trees.tree1.get(longToken(16)).addHash(digest("diff_16"), 1); |
| testDifferences(trees, newArrayList(makeTreeRange(2, 4, 3), |
| makeTreeRange(6, 8, 3), |
| makeTreeRange(10, 12, 3), |
| makeTreeRange(14, 16, 3))); |
| } |
| |
| @Test |
| public void testDifferentTrees4DifferentDepth() throws IOException |
| { |
| int seed = makeSeed(); |
| Trees trees = Trees.make(seed, seed, 2, 3); |
| trees.tree1.get(longToken(4)).addHash(digest("diff_4"), 1); |
| trees.tree1.get(longToken(8)).addHash(digest("diff_8"), 1); |
| trees.tree1.get(longToken(12)).addHash(digest("diff_12"), 1); |
| trees.tree1.get(longToken(16)).addHash(digest("diff_16"), 1); |
| testDifferences(trees, newArrayList(makeTreeRange(0, 16, 0))); |
| } |
| |
| private static void testDifferences(Trees trees, List<TreeRange> expectedDifference) throws IOException |
| { |
| MerkleTree mt1 = trees.tree1; |
| MerkleTree mt2 = trees.tree2; |
| |
| assertDiffer(mt1, mt2, expectedDifference); |
| assertDiffer(mt1, mt2.moveOffHeap(), expectedDifference); |
| assertDiffer(mt1, cycle(mt2, true), expectedDifference); |
| assertDiffer(mt1, cycle(mt2, false), expectedDifference); |
| assertDiffer(mt1, cycle(mt2.moveOffHeap(), true), expectedDifference); |
| assertDiffer(mt1, cycle(mt2.moveOffHeap(), false), expectedDifference); |
| |
| assertDiffer(mt1.moveOffHeap(), mt2, expectedDifference); |
| assertDiffer(mt1.moveOffHeap(), mt2.moveOffHeap(), expectedDifference); |
| assertDiffer(mt1.moveOffHeap(), cycle(mt2, true), expectedDifference); |
| assertDiffer(mt1.moveOffHeap(), cycle(mt2, false), expectedDifference); |
| assertDiffer(mt1.moveOffHeap(), cycle(mt2.moveOffHeap(), true), expectedDifference); |
| assertDiffer(mt1.moveOffHeap(), cycle(mt2.moveOffHeap(), false), expectedDifference); |
| |
| assertDiffer(cycle(mt1, true), mt2, expectedDifference); |
| assertDiffer(cycle(mt1, true), mt2.moveOffHeap(), expectedDifference); |
| assertDiffer(cycle(mt1, true), cycle(mt2, true), expectedDifference); |
| assertDiffer(cycle(mt1, true), cycle(mt2, false), expectedDifference); |
| assertDiffer(cycle(mt1, true), cycle(mt2.moveOffHeap(), true), expectedDifference); |
| assertDiffer(cycle(mt1, true), cycle(mt2.moveOffHeap(), false), expectedDifference); |
| |
| assertDiffer(cycle(mt1, false), mt2, expectedDifference); |
| assertDiffer(cycle(mt1, false), mt2.moveOffHeap(), expectedDifference); |
| assertDiffer(cycle(mt1, false), cycle(mt2, true), expectedDifference); |
| assertDiffer(cycle(mt1, false), cycle(mt2, false), expectedDifference); |
| assertDiffer(cycle(mt1, false), cycle(mt2.moveOffHeap(), true), expectedDifference); |
| assertDiffer(cycle(mt1, false), cycle(mt2.moveOffHeap(), false), expectedDifference); |
| |
| assertDiffer(cycle(mt1.moveOffHeap(), true), mt2, expectedDifference); |
| assertDiffer(cycle(mt1.moveOffHeap(), true), mt2.moveOffHeap(), expectedDifference); |
| assertDiffer(cycle(mt1.moveOffHeap(), true), cycle(mt2, true), expectedDifference); |
| assertDiffer(cycle(mt1.moveOffHeap(), true), cycle(mt2, false), expectedDifference); |
| assertDiffer(cycle(mt1.moveOffHeap(), true), cycle(mt2.moveOffHeap(), true), expectedDifference); |
| assertDiffer(cycle(mt1.moveOffHeap(), true), cycle(mt2.moveOffHeap(), false), expectedDifference); |
| |
| assertDiffer(cycle(mt1.moveOffHeap(), false), mt2, expectedDifference); |
| assertDiffer(cycle(mt1.moveOffHeap(), false), mt2.moveOffHeap(), expectedDifference); |
| assertDiffer(cycle(mt1.moveOffHeap(), false), cycle(mt2, true), expectedDifference); |
| assertDiffer(cycle(mt1.moveOffHeap(), false), cycle(mt2, false), expectedDifference); |
| assertDiffer(cycle(mt1.moveOffHeap(), false), cycle(mt2.moveOffHeap(), true), expectedDifference); |
| assertDiffer(cycle(mt1.moveOffHeap(), false), cycle(mt2.moveOffHeap(), false), expectedDifference); |
| } |
| |
| private static void assertDiffer(MerkleTree mt1, MerkleTree mt2, List<TreeRange> expectedDifference) |
| { |
| assertEquals(expectedDifference, MerkleTree.difference(mt1, mt2)); |
| assertEquals(expectedDifference, MerkleTree.difference(mt2, mt1)); |
| } |
| |
| private static Range<Token> longTokenRange(long start, long end) |
| { |
| return new Range<>(longToken(start), longToken(end)); |
| } |
| |
| private static Murmur3Partitioner.LongToken longToken(long value) |
| { |
| return new Murmur3Partitioner.LongToken(value); |
| } |
| |
| private static MerkleTree cycle(MerkleTree mt, boolean offHeapRequested) throws IOException |
| { |
| try (DataOutputBuffer output = new DataOutputBuffer()) |
| { |
| mt.serialize(output, MessagingService.current_version); |
| |
| try (DataInputBuffer input = new DataInputBuffer(output.buffer(false), false)) |
| { |
| return MerkleTree.deserialize(input, offHeapRequested, MessagingService.current_version); |
| } |
| } |
| } |
| |
| private static MerkleTree makeTree(long start, long end, int depth) |
| { |
| MerkleTree mt = new MerkleTree(Murmur3Partitioner.instance, longTokenRange(start, end), depth, Long.MAX_VALUE); |
| mt.init(); |
| return mt; |
| } |
| |
| private static TreeRange makeTreeRange(long start, long end, int depth) |
| { |
| return new TreeRange(longToken(start), longToken(end), depth); |
| } |
| |
| private static byte[][] makeHashes(int count, int seed) |
| { |
| Random random = new Random(seed); |
| |
| byte[][] hashes = new byte[count][32]; |
| for (int i = 0; i < count; i++) |
| random.nextBytes(hashes[i]); |
| return hashes; |
| } |
| |
| private static int makeSeed() |
| { |
| int seed = (int) System.currentTimeMillis(); |
| System.out.println("Using seed " + seed); |
| return seed; |
| } |
| |
| private static class Trees |
| { |
| MerkleTree tree1; |
| MerkleTree tree2; |
| |
| Trees(MerkleTree tree1, MerkleTree tree2) |
| { |
| this.tree1 = tree1; |
| this.tree2 = tree2; |
| } |
| |
| static Trees make(int hashes1seed, int hashes2seed, int tree1depth, int tree2depth) |
| { |
| byte[][] hashes1 = makeHashes(16, hashes1seed); |
| byte[][] hashes2 = makeHashes(16, hashes2seed); |
| |
| MerkleTree tree1 = makeTree(0, 16, tree1depth); |
| MerkleTree tree2 = makeTree(0, 16, tree2depth); |
| |
| for (int tok = 1; tok <= 16; tok++) |
| { |
| tree1.get(longToken(tok)).addHash(hashes1[tok - 1], 1); |
| tree2.get(longToken(tok)).addHash(hashes2[tok - 1], 1); |
| } |
| |
| return new Trees(tree1, tree2); |
| } |
| } |
| } |