| /* |
| * Licensed to the Apache Software Foundation (ASF) under one or more |
| * contributor license agreements. See the NOTICE file distributed with |
| * this work for additional information regarding copyright ownership. |
| * The ASF licenses this file to You under the Apache License, Version 2.0 |
| * (the "License"); you may not use this file except in compliance with |
| * the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.lucene.codecs.compressing; |
| |
| import com.carrotsearch.randomizedtesting.generators.RandomNumbers; |
| import java.util.Random; |
| import org.apache.lucene.codecs.FilterCodec; |
| import org.apache.lucene.codecs.StoredFieldsFormat; |
| import org.apache.lucene.codecs.TermVectorsFormat; |
| import org.apache.lucene.codecs.compressing.dummy.DummyCompressingCodec; |
| import org.apache.lucene.util.TestUtil; |
| import org.apache.lucene.util.packed.DirectMonotonicWriter; |
| |
| /** |
| * A codec that uses {@link CompressingStoredFieldsFormat} for its stored fields and delegates to |
| * the default codec for everything else. |
| */ |
| public abstract class CompressingCodec extends FilterCodec { |
| |
| /** Create a random instance. */ |
| public static CompressingCodec randomInstance( |
| Random random, |
| int chunkSize, |
| int maxDocsPerChunk, |
| boolean withSegmentSuffix, |
| int blockShift) { |
| switch (random.nextInt(6)) { |
| case 0: |
| return new FastCompressingCodec(chunkSize, maxDocsPerChunk, withSegmentSuffix, blockShift); |
| case 1: |
| return new FastDecompressionCompressingCodec( |
| chunkSize, maxDocsPerChunk, withSegmentSuffix, blockShift); |
| case 2: |
| return new HighCompressionCompressingCodec( |
| chunkSize, maxDocsPerChunk, withSegmentSuffix, blockShift); |
| case 3: |
| return new DummyCompressingCodec(chunkSize, maxDocsPerChunk, withSegmentSuffix, blockShift); |
| case 4: |
| return new DeflateWithPresetCompressingCodec( |
| chunkSize, maxDocsPerChunk, withSegmentSuffix, blockShift); |
| case 5: |
| return new LZ4WithPresetCompressingCodec( |
| chunkSize, maxDocsPerChunk, withSegmentSuffix, blockShift); |
| default: |
| throw new AssertionError(); |
| } |
| } |
| |
| /** Creates a random {@link CompressingCodec} that is using an empty segment suffix */ |
| public static CompressingCodec randomInstance(Random random) { |
| final int chunkSize = |
| random.nextBoolean() |
| ? RandomNumbers.randomIntBetween(random, 10, 100) |
| : RandomNumbers.randomIntBetween(random, 10, 1 << 15); |
| final int chunkDocs = |
| random.nextBoolean() |
| ? RandomNumbers.randomIntBetween(random, 1, 10) |
| : RandomNumbers.randomIntBetween(random, 64, 1024); |
| final int blockSize = |
| random.nextBoolean() |
| ? RandomNumbers.randomIntBetween(random, DirectMonotonicWriter.MIN_BLOCK_SHIFT, 10) |
| : RandomNumbers.randomIntBetween( |
| random, |
| DirectMonotonicWriter.MIN_BLOCK_SHIFT, |
| DirectMonotonicWriter.MAX_BLOCK_SHIFT); |
| return randomInstance(random, chunkSize, chunkDocs, false, blockSize); |
| } |
| |
| /** Creates a random {@link CompressingCodec} with more reasonable parameters for big tests. */ |
| public static CompressingCodec reasonableInstance(Random random) { |
| // e.g. defaults use 2^14 for FAST and ~ 2^16 for HIGH |
| final int chunkSize = TestUtil.nextInt(random, 1 << 13, 1 << 17); |
| // e.g. defaults use 128 for FAST and 512 for HIGH |
| final int chunkDocs = TestUtil.nextInt(random, 1 << 6, 1 << 10); |
| // e.g. defaults use 1024 for both cases |
| final int blockShift = TestUtil.nextInt(random, 8, 12); |
| return randomInstance(random, chunkSize, chunkDocs, false, blockShift); |
| } |
| |
| /** Creates a random {@link CompressingCodec} that is using a segment suffix */ |
| public static CompressingCodec randomInstance(Random random, boolean withSegmentSuffix) { |
| return randomInstance( |
| random, |
| RandomNumbers.randomIntBetween(random, 1, 1 << 15), |
| RandomNumbers.randomIntBetween(random, 64, 1024), |
| withSegmentSuffix, |
| RandomNumbers.randomIntBetween(random, 1, 1024)); |
| } |
| |
| private final CompressingStoredFieldsFormat storedFieldsFormat; |
| private final Lucene90CompressingTermVectorsFormat termVectorsFormat; |
| |
| /** Creates a compressing codec with a given segment suffix */ |
| public CompressingCodec( |
| String name, |
| String segmentSuffix, |
| CompressionMode compressionMode, |
| int chunkSize, |
| int maxDocsPerChunk, |
| int blockShift) { |
| super(name, TestUtil.getDefaultCodec()); |
| this.storedFieldsFormat = |
| new CompressingStoredFieldsFormat( |
| name, segmentSuffix, compressionMode, chunkSize, maxDocsPerChunk, blockShift); |
| this.termVectorsFormat = |
| new Lucene90CompressingTermVectorsFormat( |
| name, segmentSuffix, compressionMode, chunkSize, blockShift); |
| } |
| |
| /** Creates a compressing codec with an empty segment suffix */ |
| public CompressingCodec( |
| String name, |
| CompressionMode compressionMode, |
| int chunkSize, |
| int maxDocsPerChunk, |
| int blockSize) { |
| this(name, "", compressionMode, chunkSize, maxDocsPerChunk, blockSize); |
| } |
| |
| @Override |
| public StoredFieldsFormat storedFieldsFormat() { |
| return storedFieldsFormat; |
| } |
| |
| @Override |
| public TermVectorsFormat termVectorsFormat() { |
| return termVectorsFormat; |
| } |
| |
| @Override |
| public String toString() { |
| return getName() |
| + "(storedFieldsFormat=" |
| + storedFieldsFormat |
| + ", termVectorsFormat=" |
| + termVectorsFormat |
| + ")"; |
| } |
| } |