| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| */ |
| package org.apache.cassandra.db; |
| |
| import java.math.BigInteger; |
| import java.net.InetAddress; |
| import java.nio.ByteBuffer; |
| import java.nio.charset.CharacterCodingException; |
| import java.util.ArrayList; |
| import java.util.Arrays; |
| import java.util.Collections; |
| import java.util.Iterator; |
| import java.util.List; |
| |
| import com.google.common.collect.Iterators; |
| import org.junit.BeforeClass; |
| import org.junit.Test; |
| |
| import static org.apache.cassandra.db.ConsistencyLevel.ONE; |
| import static org.junit.Assert.assertEquals; |
| import static org.junit.Assert.assertTrue; |
| |
| import org.apache.cassandra.*; |
| import org.apache.cassandra.config.CFMetaData; |
| import org.apache.cassandra.config.ColumnDefinition; |
| import org.apache.cassandra.config.DatabaseDescriptor; |
| import org.apache.cassandra.db.marshal.AsciiType; |
| import org.apache.cassandra.db.rows.Row; |
| import org.apache.cassandra.db.marshal.IntegerType; |
| import org.apache.cassandra.db.partitions.*; |
| import org.apache.cassandra.dht.IPartitioner; |
| import org.apache.cassandra.dht.Range; |
| import org.apache.cassandra.dht.Token; |
| import org.apache.cassandra.exceptions.ConfigurationException; |
| import org.apache.cassandra.locator.TokenMetadata; |
| import org.apache.cassandra.schema.KeyspaceParams; |
| import org.apache.cassandra.service.StorageProxy; |
| import org.apache.cassandra.service.StorageService; |
| import org.apache.cassandra.utils.ByteBufferUtil; |
| import org.apache.cassandra.utils.FBUtilities; |
| |
| public class PartitionRangeReadTest |
| { |
| public static final String KEYSPACE1 = "PartitionRangeReadTest1"; |
| public static final String KEYSPACE2 = "PartitionRangeReadTest2"; |
| public static final String CF_STANDARD1 = "Standard1"; |
| public static final String CF_STANDARDINT = "StandardInteger1"; |
| public static final String CF_COMPACT1 = "Compact1"; |
| |
| private static final List<InetAddress> LOCAL = Collections.singletonList(FBUtilities.getBroadcastAddress()); |
| |
| @BeforeClass |
| public static void defineSchema() throws ConfigurationException |
| { |
| SchemaLoader.prepareServer(); |
| SchemaLoader.createKeyspace(KEYSPACE1, |
| KeyspaceParams.simple(1), |
| SchemaLoader.standardCFMD(KEYSPACE1, CF_STANDARD1), |
| SchemaLoader.denseCFMD(KEYSPACE1, CF_STANDARDINT, IntegerType.instance), |
| CFMetaData.Builder.create(KEYSPACE1, CF_COMPACT1, false, false, false) |
| .addPartitionKey("key", AsciiType.instance) |
| .addClusteringColumn("column1", AsciiType.instance) |
| .addRegularColumn("value", AsciiType.instance) |
| .addStaticColumn("val", AsciiType.instance) |
| .build()); |
| SchemaLoader.createKeyspace(KEYSPACE2, |
| KeyspaceParams.simple(1), |
| SchemaLoader.standardCFMD(KEYSPACE2, CF_STANDARD1)); |
| } |
| |
| @Test |
| public void testInclusiveBounds() |
| { |
| ColumnFamilyStore cfs = Keyspace.open(KEYSPACE2).getColumnFamilyStore(CF_STANDARD1); |
| new RowUpdateBuilder(cfs.metadata, 0, ByteBufferUtil.bytes("key1")) |
| .clustering("cc1") |
| .add("val", "asdf").build().applyUnsafe(); |
| new RowUpdateBuilder(cfs.metadata, 0, ByteBufferUtil.bytes("key2")) |
| .clustering("cc2") |
| .add("val", "asdf").build().applyUnsafe(); |
| |
| assertEquals(2, Util.getAll(Util.cmd(cfs).fromIncl("cc1").toIncl("cc2").build()).size()); |
| } |
| |
| @Test |
| public void testCassandra6778() throws CharacterCodingException |
| { |
| String cfname = CF_STANDARDINT; |
| Keyspace keyspace = Keyspace.open(KEYSPACE1); |
| ColumnFamilyStore cfs = keyspace.getColumnFamilyStore(cfname); |
| cfs.truncateBlocking(); |
| |
| ByteBuffer col = ByteBufferUtil.bytes("val"); |
| ColumnDefinition cDef = cfs.metadata.getColumnDefinition(col); |
| |
| // insert two columns that represent the same integer but have different binary forms (the |
| // second one is padded with extra zeros) |
| new RowUpdateBuilder(cfs.metadata, 0, "k1") |
| .clustering(new BigInteger(new byte[]{1})) |
| .add("val", "val1") |
| .build() |
| .applyUnsafe(); |
| cfs.forceBlockingFlush(); |
| |
| new RowUpdateBuilder(cfs.metadata, 1, "k1") |
| .clustering(new BigInteger(new byte[]{0, 0, 1})) |
| .add("val", "val2") |
| .build() |
| .applyUnsafe(); |
| cfs.forceBlockingFlush(); |
| |
| // fetch by the first column name; we should get the second version of the column value |
| Row row = Util.getOnlyRow(Util.cmd(cfs, "k1").includeRow(new BigInteger(new byte[]{1})).build()); |
| assertTrue(row.getCell(cDef).value().equals(ByteBufferUtil.bytes("val2"))); |
| |
| // fetch by the second column name; we should get the second version of the column value |
| row = Util.getOnlyRow(Util.cmd(cfs, "k1").includeRow(new BigInteger(new byte[]{0, 0, 1})).build()); |
| assertTrue(row.getCell(cDef).value().equals(ByteBufferUtil.bytes("val2"))); |
| } |
| |
| @Test |
| public void testLimits() |
| { |
| ColumnFamilyStore cfs = Keyspace.open(KEYSPACE1).getColumnFamilyStore(CF_COMPACT1); |
| for (int i = 0; i < 10; i++) |
| { |
| new RowUpdateBuilder(cfs.metadata, 0, Integer.toString(i)) |
| .add("val", "abcd") |
| .build() |
| .applyUnsafe(); |
| } |
| |
| assertEquals(10, Util.getAll(Util.cmd(cfs).build()).size()); |
| |
| for (int i = 0; i < 10; i++) |
| assertEquals(i, Util.getAll(Util.cmd(cfs).withLimit(i).build()).size()); |
| } |
| |
| @Test |
| public void testRangeSliceInclusionExclusion() throws Throwable |
| { |
| String keyspaceName = KEYSPACE1; |
| String cfName = CF_STANDARD1; |
| Keyspace keyspace = Keyspace.open(keyspaceName); |
| ColumnFamilyStore cfs = keyspace.getColumnFamilyStore(cfName); |
| cfs.clearUnsafe(); |
| |
| for (int i = 0; i < 10; ++i) |
| { |
| RowUpdateBuilder builder = new RowUpdateBuilder(cfs.metadata, 10, String.valueOf(i)); |
| builder.clustering("c"); |
| builder.add("val", String.valueOf(i)); |
| builder.build().applyUnsafe(); |
| } |
| |
| cfs.forceBlockingFlush(); |
| |
| ColumnDefinition cDef = cfs.metadata.getColumnDefinition(ByteBufferUtil.bytes("val")); |
| |
| List<FilteredPartition> partitions; |
| |
| // Start and end inclusive |
| partitions = Util.getAll(Util.cmd(cfs).fromKeyIncl("2").toKeyIncl("7").build()); |
| assertEquals(6, partitions.size()); |
| assertTrue(partitions.get(0).iterator().next().getCell(cDef).value().equals(ByteBufferUtil.bytes("2"))); |
| assertTrue(partitions.get(partitions.size() - 1).iterator().next().getCell(cDef).value().equals(ByteBufferUtil.bytes("7"))); |
| |
| // Start and end excluded |
| partitions = Util.getAll(Util.cmd(cfs).fromKeyExcl("2").toKeyExcl("7").build()); |
| assertEquals(4, partitions.size()); |
| assertTrue(partitions.get(0).iterator().next().getCell(cDef).value().equals(ByteBufferUtil.bytes("3"))); |
| assertTrue(partitions.get(partitions.size() - 1).iterator().next().getCell(cDef).value().equals(ByteBufferUtil.bytes("6"))); |
| |
| // Start excluded, end included |
| partitions = Util.getAll(Util.cmd(cfs).fromKeyExcl("2").toKeyIncl("7").build()); |
| assertEquals(5, partitions.size()); |
| assertTrue(partitions.get(0).iterator().next().getCell(cDef).value().equals(ByteBufferUtil.bytes("3"))); |
| assertTrue(partitions.get(partitions.size() - 1).iterator().next().getCell(cDef).value().equals(ByteBufferUtil.bytes("7"))); |
| |
| // Start included, end excluded |
| partitions = Util.getAll(Util.cmd(cfs).fromKeyIncl("2").toKeyExcl("7").build()); |
| assertEquals(5, partitions.size()); |
| assertTrue(partitions.get(0).iterator().next().getCell(cDef).value().equals(ByteBufferUtil.bytes("2"))); |
| assertTrue(partitions.get(partitions.size() - 1).iterator().next().getCell(cDef).value().equals(ByteBufferUtil.bytes("6"))); |
| } |
| |
| // TODO: Port or remove, depending on what DataLimits.thriftLimits (per cell) looks like |
| // @Test |
| // public void testRangeSliceColumnsLimit() throws Throwable |
| // { |
| // String keyspaceName = KEYSPACE1; |
| // String cfName = CF_STANDARD1; |
| // Keyspace keyspace = Keyspace.open(keyspaceName); |
| // ColumnFamilyStore cfs = keyspace.getColumnFamilyStore(cfName); |
| // cfs.clearUnsafe(); |
| // |
| // Cell[] cols = new Cell[5]; |
| // for (int i = 0; i < 5; i++) |
| // cols[i] = column("c" + i, "value", 1); |
| // |
| // putColsStandard(cfs, Util.dk("a"), cols[0], cols[1], cols[2], cols[3], cols[4]); |
| // putColsStandard(cfs, Util.dk("b"), cols[0], cols[1]); |
| // putColsStandard(cfs, Util.dk("c"), cols[0], cols[1], cols[2], cols[3]); |
| // cfs.forceBlockingFlush(); |
| // |
| // SlicePredicate sp = new SlicePredicate(); |
| // sp.setSlice_range(new SliceRange()); |
| // sp.getSlice_range().setCount(1); |
| // sp.getSlice_range().setStart(ArrayUtils.EMPTY_BYTE_ARRAY); |
| // sp.getSlice_range().setFinish(ArrayUtils.EMPTY_BYTE_ARRAY); |
| // |
| // assertTotalColCount(cfs.getRangeSlice(Util.range("", ""), |
| // null, |
| // ThriftValidation.asIFilter(sp, cfs.metadata, null), |
| // 3, |
| // System.currentTimeMillis(), |
| // true, |
| // false), |
| // 3); |
| // assertTotalColCount(cfs.getRangeSlice(Util.range("", ""), |
| // null, |
| // ThriftValidation.asIFilter(sp, cfs.metadata, null), |
| // 5, |
| // System.currentTimeMillis(), |
| // true, |
| // false), |
| // 5); |
| // assertTotalColCount(cfs.getRangeSlice(Util.range("", ""), |
| // null, |
| // ThriftValidation.asIFilter(sp, cfs.metadata, null), |
| // 8, |
| // System.currentTimeMillis(), |
| // true, |
| // false), |
| // 8); |
| // assertTotalColCount(cfs.getRangeSlice(Util.range("", ""), |
| // null, |
| // ThriftValidation.asIFilter(sp, cfs.metadata, null), |
| // 10, |
| // System.currentTimeMillis(), |
| // true, |
| // false), |
| // 10); |
| // assertTotalColCount(cfs.getRangeSlice(Util.range("", ""), |
| // null, |
| // ThriftValidation.asIFilter(sp, cfs.metadata, null), |
| // 100, |
| // System.currentTimeMillis(), |
| // true, |
| // false), |
| // 11); |
| // |
| // // Check that when querying by name, we always include all names for a |
| // // gien row even if it means returning more columns than requested (this is necesseray for CQL) |
| // sp = new SlicePredicate(); |
| // sp.setColumn_names(Arrays.asList( |
| // ByteBufferUtil.bytes("c0"), |
| // ByteBufferUtil.bytes("c1"), |
| // ByteBufferUtil.bytes("c2") |
| // )); |
| // |
| // assertTotalColCount(cfs.getRangeSlice(Util.range("", ""), |
| // null, |
| // ThriftValidation.asIFilter(sp, cfs.metadata, null), |
| // 1, |
| // System.currentTimeMillis(), |
| // true, |
| // false), |
| // 3); |
| // assertTotalColCount(cfs.getRangeSlice(Util.range("", ""), |
| // null, |
| // ThriftValidation.asIFilter(sp, cfs.metadata, null), |
| // 4, |
| // System.currentTimeMillis(), |
| // true, |
| // false), |
| // 5); |
| // assertTotalColCount(cfs.getRangeSlice(Util.range("", ""), |
| // null, |
| // ThriftValidation.asIFilter(sp, cfs.metadata, null), |
| // 5, |
| // System.currentTimeMillis(), |
| // true, |
| // false), |
| // 5); |
| // assertTotalColCount(cfs.getRangeSlice(Util.range("", ""), |
| // null, |
| // ThriftValidation.asIFilter(sp, cfs.metadata, null), |
| // 6, |
| // System.currentTimeMillis(), |
| // true, |
| // false), |
| // 8); |
| // assertTotalColCount(cfs.getRangeSlice(Util.range("", ""), |
| // null, |
| // ThriftValidation.asIFilter(sp, cfs.metadata, null), |
| // 100, |
| // System.currentTimeMillis(), |
| // true, |
| // false), |
| // 8); |
| // } |
| |
| // TODO: Port or remove, depending on what DataLimits.thriftLimits (per cell) looks like |
| // @Test |
| // public void testRangeSlicePaging() throws Throwable |
| // { |
| // String keyspaceName = KEYSPACE1; |
| // String cfName = CF_STANDARD1; |
| // Keyspace keyspace = Keyspace.open(keyspaceName); |
| // ColumnFamilyStore cfs = keyspace.getColumnFamilyStore(cfName); |
| // cfs.clearUnsafe(); |
| // |
| // Cell[] cols = new Cell[4]; |
| // for (int i = 0; i < 4; i++) |
| // cols[i] = column("c" + i, "value", 1); |
| // |
| // DecoratedKey ka = Util.dk("a"); |
| // DecoratedKey kb = Util.dk("b"); |
| // DecoratedKey kc = Util.dk("c"); |
| // |
| // PartitionPosition min = Util.rp(""); |
| // |
| // putColsStandard(cfs, ka, cols[0], cols[1], cols[2], cols[3]); |
| // putColsStandard(cfs, kb, cols[0], cols[1], cols[2]); |
| // putColsStandard(cfs, kc, cols[0], cols[1], cols[2], cols[3]); |
| // cfs.forceBlockingFlush(); |
| // |
| // SlicePredicate sp = new SlicePredicate(); |
| // sp.setSlice_range(new SliceRange()); |
| // sp.getSlice_range().setCount(1); |
| // sp.getSlice_range().setStart(ArrayUtils.EMPTY_BYTE_ARRAY); |
| // sp.getSlice_range().setFinish(ArrayUtils.EMPTY_BYTE_ARRAY); |
| // |
| // Collection<Row> rows; |
| // Row row, row1, row2; |
| // IDiskAtomFilter filter = ThriftValidation.asIFilter(sp, cfs.metadata, null); |
| // |
| // rows = cfs.getRangeSlice(cfs.makeExtendedFilter(Util.range("", ""), filter, null, 3, true, true, System.currentTimeMillis())); |
| // assert rows.size() == 1 : "Expected 1 row, got " + toString(rows); |
| // row = rows.iterator().next(); |
| // assertColumnNames(row, "c0", "c1", "c2"); |
| // |
| // sp.getSlice_range().setStart(ByteBufferUtil.getArray(ByteBufferUtil.bytes("c2"))); |
| // filter = ThriftValidation.asIFilter(sp, cfs.metadata, null); |
| // rows = cfs.getRangeSlice(cfs.makeExtendedFilter(new Bounds<PartitionPosition>(ka, min), filter, null, 3, true, true, System.currentTimeMillis())); |
| // assert rows.size() == 2 : "Expected 2 rows, got " + toString(rows); |
| // Iterator<Row> iter = rows.iterator(); |
| // row1 = iter.next(); |
| // row2 = iter.next(); |
| // assertColumnNames(row1, "c2", "c3"); |
| // assertColumnNames(row2, "c0"); |
| // |
| // sp.getSlice_range().setStart(ByteBufferUtil.getArray(ByteBufferUtil.bytes("c0"))); |
| // filter = ThriftValidation.asIFilter(sp, cfs.metadata, null); |
| // rows = cfs.getRangeSlice(cfs.makeExtendedFilter(new Bounds<PartitionPosition>(row2.key, min), filter, null, 3, true, true, System.currentTimeMillis())); |
| // assert rows.size() == 1 : "Expected 1 row, got " + toString(rows); |
| // row = rows.iterator().next(); |
| // assertColumnNames(row, "c0", "c1", "c2"); |
| // |
| // sp.getSlice_range().setStart(ByteBufferUtil.getArray(ByteBufferUtil.bytes("c2"))); |
| // filter = ThriftValidation.asIFilter(sp, cfs.metadata, null); |
| // rows = cfs.getRangeSlice(cfs.makeExtendedFilter(new Bounds<PartitionPosition>(row.key, min), filter, null, 3, true, true, System.currentTimeMillis())); |
| // assert rows.size() == 2 : "Expected 2 rows, got " + toString(rows); |
| // iter = rows.iterator(); |
| // row1 = iter.next(); |
| // row2 = iter.next(); |
| // assertColumnNames(row1, "c2"); |
| // assertColumnNames(row2, "c0", "c1"); |
| // |
| // // Paging within bounds |
| // SliceQueryFilter sf = new SliceQueryFilter(cellname("c1"), |
| // cellname("c2"), |
| // false, |
| // 0); |
| // rows = cfs.getRangeSlice(cfs.makeExtendedFilter(new Bounds<PartitionPosition>(ka, kc), sf, cellname("c2"), cellname("c1"), null, 2, true, System.currentTimeMillis())); |
| // assert rows.size() == 2 : "Expected 2 rows, got " + toString(rows); |
| // iter = rows.iterator(); |
| // row1 = iter.next(); |
| // row2 = iter.next(); |
| // assertColumnNames(row1, "c2"); |
| // assertColumnNames(row2, "c1"); |
| // |
| // rows = cfs.getRangeSlice(cfs.makeExtendedFilter(new Bounds<PartitionPosition>(kb, kc), sf, cellname("c1"), cellname("c1"), null, 10, true, System.currentTimeMillis())); |
| // assert rows.size() == 2 : "Expected 2 rows, got " + toString(rows); |
| // iter = rows.iterator(); |
| // row1 = iter.next(); |
| // row2 = iter.next(); |
| // assertColumnNames(row1, "c1", "c2"); |
| // assertColumnNames(row2, "c1"); |
| // } |
| |
| @Test |
| public void testComputeConcurrencyFactor() |
| { |
| int maxConcurrentRangeRequest = 32; |
| |
| // no live row returned, fetch all remaining ranges but hit the max instead |
| int cf = StorageProxy.RangeCommandIterator.computeConcurrencyFactor(100, 30, maxConcurrentRangeRequest, 500, 0); |
| assertEquals(maxConcurrentRangeRequest, cf); // because 100 - 30 = 70 > maxConccurrentRangeRequest |
| |
| // no live row returned, fetch all remaining ranges |
| cf = StorageProxy.RangeCommandIterator.computeConcurrencyFactor(100, 80, maxConcurrentRangeRequest, 500, 0); |
| assertEquals(20, cf); // because 100-80 = 20 < maxConccurrentRangeRequest |
| |
| // returned half rows, fetch rangesQueried again but hit the max instead |
| cf = StorageProxy.RangeCommandIterator.computeConcurrencyFactor(100, 60, maxConcurrentRangeRequest, 480, 240); |
| assertEquals(maxConcurrentRangeRequest, cf); // because 60 > maxConccurrentRangeRequest |
| |
| // returned half rows, fetch rangesQueried again |
| cf = StorageProxy.RangeCommandIterator.computeConcurrencyFactor(100, 30, maxConcurrentRangeRequest, 480, 240); |
| assertEquals(30, cf); // because 30 < maxConccurrentRangeRequest |
| |
| // returned most of rows, 1 more range to fetch |
| cf = StorageProxy.RangeCommandIterator.computeConcurrencyFactor(100, 1, maxConcurrentRangeRequest, 480, 479); |
| assertEquals(1, cf); // because 1 < maxConccurrentRangeRequest |
| } |
| |
| @Test |
| public void testRangeCountWithRangeMerge() |
| { |
| List<Token> tokens = setTokens(Arrays.asList(100, 200, 300, 400)); |
| int vnodeCount = 0; |
| |
| Keyspace keyspace = Keyspace.open(KEYSPACE1); |
| List<StorageProxy.RangeForQuery> ranges = new ArrayList<>(); |
| for (int i = 0; i + 1 < tokens.size(); i++) |
| { |
| Range<PartitionPosition> range = Range.makeRowRange(tokens.get(i), tokens.get(i + 1)); |
| ranges.add(new StorageProxy.RangeForQuery(range, LOCAL, LOCAL, 1)); |
| vnodeCount++; |
| } |
| |
| StorageProxy.RangeMerger merge = new StorageProxy.RangeMerger(ranges.iterator(), keyspace, ONE); |
| StorageProxy.RangeForQuery mergedRange = Iterators.getOnlyElement(merge); |
| // all ranges are merged as test has only one node. |
| assertEquals(vnodeCount, mergedRange.vnodeCount()); |
| } |
| |
| @Test |
| public void testRangeQueried() |
| { |
| List<Token> tokens = setTokens(Arrays.asList(100, 200, 300, 400)); |
| int vnodeCount = tokens.size() + 1; // n tokens divide token ring into n+1 ranges |
| |
| Keyspace keyspace = Keyspace.open(KEYSPACE1); |
| ColumnFamilyStore cfs = keyspace.getColumnFamilyStore(CF_STANDARD1); |
| cfs.clearUnsafe(); |
| |
| int rows = 100; |
| for (int i = 0; i < rows; ++i) |
| { |
| RowUpdateBuilder builder = new RowUpdateBuilder(cfs.metadata, 10, String.valueOf(i)); |
| builder.clustering("c"); |
| builder.add("val", String.valueOf(i)); |
| builder.build().applyUnsafe(); |
| } |
| cfs.forceBlockingFlush(); |
| |
| PartitionRangeReadCommand command = (PartitionRangeReadCommand) Util.cmd(cfs).build(); |
| |
| // without range merger, there will be 2 batches requested: 1st batch with 1 range and 2nd batch with remaining ranges |
| Iterator<StorageProxy.RangeForQuery> ranges = rangeIterator(command, keyspace, false); |
| StorageProxy.RangeCommandIterator data = new StorageProxy.RangeCommandIterator(ranges, command, 1, 1000, vnodeCount, keyspace, ONE); |
| verifyRangeCommandIterator(data, rows, 2, vnodeCount); |
| |
| // without range merger and initial cf=5, there will be 1 batches requested: 5 vnode ranges for 1st batch |
| ranges = rangeIterator(command, keyspace, false); |
| data = new StorageProxy.RangeCommandIterator(ranges, command, vnodeCount, 1000, vnodeCount, keyspace, ONE); |
| verifyRangeCommandIterator(data, rows, 1, vnodeCount); |
| |
| // without range merger and max cf=1, there will be 5 batches requested: 1 vnode range per batch |
| ranges = rangeIterator(command, keyspace, false); |
| data = new StorageProxy.RangeCommandIterator(ranges, command, 1, 1, vnodeCount, keyspace, ONE); |
| verifyRangeCommandIterator(data, rows, vnodeCount, vnodeCount); |
| |
| // with range merger, there will be only 1 batch requested, as all ranges share the same replica - localhost |
| ranges = rangeIterator(command, keyspace, true); |
| data = new StorageProxy.RangeCommandIterator(ranges, command, 1, 1000, vnodeCount, keyspace, ONE); |
| verifyRangeCommandIterator(data, rows, 1, vnodeCount); |
| |
| // with range merger and max cf=1, there will be only 1 batch requested, as all ranges share the same replica - localhost |
| ranges = rangeIterator(command, keyspace, true); |
| data = new StorageProxy.RangeCommandIterator(ranges, command, 1, 1, vnodeCount, keyspace, ONE); |
| verifyRangeCommandIterator(data, rows, 1, vnodeCount); |
| } |
| |
| private Iterator<StorageProxy.RangeForQuery> rangeIterator(PartitionRangeReadCommand command, Keyspace keyspace, boolean withRangeMerger) |
| { |
| Iterator<StorageProxy.RangeForQuery> ranges = new StorageProxy.RangeIterator(command, keyspace, ONE); |
| if (withRangeMerger) |
| ranges = new StorageProxy.RangeMerger(ranges, keyspace, ONE); |
| |
| return ranges; |
| } |
| |
| private void verifyRangeCommandIterator(StorageProxy.RangeCommandIterator data, int rows, int batches, int vnodeCount) |
| { |
| int num = Util.size(data); |
| assertEquals(rows, num); |
| assertEquals(batches, data.batchesRequested()); |
| assertEquals(vnodeCount, data.rangesQueried()); |
| } |
| |
| private List<Token> setTokens(List<Integer> values) |
| { |
| IPartitioner partitioner = DatabaseDescriptor.getPartitioner(); |
| List<Token> tokens = new ArrayList<>(values.size()); |
| for (Integer val : values) |
| tokens.add(partitioner.getToken(ByteBufferUtil.bytes(val))); |
| |
| TokenMetadata tmd = StorageService.instance.getTokenMetadata(); |
| tmd.clearUnsafe(); |
| tmd.updateNormalTokens(tokens, FBUtilities.getBroadcastAddress()); |
| |
| return tokens; |
| } |
| } |
| |