| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| */ |
| |
| package org.apache.druid.tests.indexer; |
| |
| import com.fasterxml.jackson.core.JsonProcessingException; |
| import org.apache.druid.indexer.partitions.HashedPartitionsSpec; |
| import org.apache.druid.indexer.partitions.PartitionsSpec; |
| import org.apache.druid.indexer.partitions.SingleDimensionPartitionsSpec; |
| import org.apache.druid.java.util.common.StringUtils; |
| import org.apache.druid.testing.guice.DruidTestModuleFactory; |
| import org.apache.druid.tests.TestNGGroup; |
| import org.apache.druid.timeline.partition.HashPartitionFunction; |
| import org.testng.Assert; |
| import org.testng.annotations.DataProvider; |
| import org.testng.annotations.Guice; |
| import org.testng.annotations.Test; |
| |
| import java.io.Closeable; |
| import java.util.function.Function; |
| |
| @Test(groups = TestNGGroup.PERFECT_ROLLUP_PARALLEL_BATCH_INDEX) |
| @Guice(moduleFactory = DruidTestModuleFactory.class) |
| public class ITPerfectRollupParallelIndexTest extends AbstractITBatchIndexTest |
| { |
| // This ingestion spec has a splitHintSpec of maxSplitSize of 1 to test whether or not the task can handle |
| // maxSplitSize of 1 properly. |
| private static final String INDEX_TASK = "/indexer/wikipedia_parallel_index_task.json"; |
| private static final String INDEX_QUERIES_RESOURCE = "/indexer/wikipedia_parallel_index_queries.json"; |
| private static final String INDEX_DATASOURCE = "wikipedia_parallel_index_test"; |
| private static final String INDEX_INGEST_SEGMENT_DATASOURCE = "wikipedia_parallel_ingest_segment_index_test"; |
| private static final String INDEX_INGEST_SEGMENT_TASK = "/indexer/wikipedia_parallel_ingest_segment_index_task.json"; |
| private static final String INDEX_DRUID_INPUT_SOURCE_DATASOURCE = "wikipedia_parallel_druid_input_source_index_test"; |
| private static final String INDEX_DRUID_INPUT_SOURCE_TASK = "/indexer/wikipedia_parallel_druid_input_source_index_task.json"; |
| |
| @DataProvider |
| public static Object[][] resources() |
| { |
| return new Object[][]{ |
| {new HashedPartitionsSpec(null, 2, null, HashPartitionFunction.MURMUR3_32_ABS)}, |
| {new HashedPartitionsSpec(null, 2, null, null)}, |
| {new SingleDimensionPartitionsSpec(2, null, "namespace", false)} |
| }; |
| } |
| |
| @Test(dataProvider = "resources") |
| public void testIndexData(PartitionsSpec partitionsSpec) throws Exception |
| { |
| try ( |
| final Closeable ignored1 = unloader(INDEX_DATASOURCE + config.getExtraDatasourceNameSuffix()); |
| final Closeable ignored2 = unloader(INDEX_INGEST_SEGMENT_DATASOURCE + config.getExtraDatasourceNameSuffix()); |
| final Closeable ignored3 = unloader(INDEX_DRUID_INPUT_SOURCE_DATASOURCE + config.getExtraDatasourceNameSuffix()) |
| ) { |
| boolean forceGuaranteedRollup = partitionsSpec.isForceGuaranteedRollupCompatible(); |
| Assert.assertTrue(forceGuaranteedRollup, "parititionSpec does not support perfect rollup"); |
| |
| final Function<String, String> rollupTransform = spec -> { |
| try { |
| spec = StringUtils.replace( |
| spec, |
| "%%FORCE_GUARANTEED_ROLLUP%%", |
| Boolean.toString(true) |
| ); |
| return StringUtils.replace( |
| spec, |
| "%%PARTITIONS_SPEC%%", |
| jsonMapper.writeValueAsString(partitionsSpec) |
| ); |
| } |
| catch (JsonProcessingException e) { |
| throw new RuntimeException(e); |
| } |
| }; |
| |
| doIndexTest( |
| INDEX_DATASOURCE, |
| INDEX_TASK, |
| rollupTransform, |
| INDEX_QUERIES_RESOURCE, |
| false, |
| true, |
| true |
| ); |
| |
| doReindexTest( |
| INDEX_DATASOURCE, |
| INDEX_INGEST_SEGMENT_DATASOURCE, |
| rollupTransform, |
| INDEX_INGEST_SEGMENT_TASK, |
| INDEX_QUERIES_RESOURCE |
| ); |
| |
| // with DruidInputSource instead of IngestSegmentFirehose |
| doReindexTest( |
| INDEX_DATASOURCE, |
| INDEX_DRUID_INPUT_SOURCE_DATASOURCE, |
| rollupTransform, |
| INDEX_DRUID_INPUT_SOURCE_TASK, |
| INDEX_QUERIES_RESOURCE |
| ); |
| } |
| } |
| } |