blob: e7e85025b698c10a3ab6ea00b8247dfd464757d3 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.drill.exec.store.druid.druid;
import com.fasterxml.jackson.core.JsonProcessingException;
import com.fasterxml.jackson.databind.JsonNode;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.fasterxml.jackson.databind.node.ObjectNode;
import org.apache.drill.common.util.JacksonUtils;
import org.apache.drill.exec.store.druid.common.DruidFilter;
import java.math.BigInteger;
import java.util.ArrayList;
import java.util.List;
import java.util.stream.Collectors;
import java.util.stream.Stream;
import static org.apache.drill.exec.store.druid.common.DruidConstants.INTERVAL_DIMENSION_NAME;
public class ScanQueryBuilder {
private static final ObjectMapper objectMapper = JacksonUtils.createObjectMapper();
public ScanQueryBuilder() {}
public ScanQuery build(String datasource,
List<String> columns,
DruidFilter druidFilter,
BigInteger nextOffset,
int queryThreshold,
String minTime,
String maxTime)
throws JsonProcessingException {
List<JsonNode> userInputIntervals =
druidFilter == null
? new ArrayList<>()
: parseIntervalsFromFilter(druidFilter.toJson());
List<String> queryIntervals = getQueryIntervals(userInputIntervals, minTime, maxTime);
ObjectNode finalFilter =
druidFilter == null
? null
: (ObjectNode) objectMapper.readTree(removeIntervalFilter(druidFilter.toJson(), userInputIntervals));
return new ScanQuery(
datasource,
columns,
finalFilter,
queryIntervals,
nextOffset,
queryThreshold);
}
private List<JsonNode> parseIntervalsFromFilter(String filter)
throws JsonProcessingException {
//if the filter is on the special Interval Dimension, then use it for the interval.
JsonNode filterNode = objectMapper.readTree(filter);
return filterNode.findValues(INTERVAL_DIMENSION_NAME);
}
private List<String> getQueryIntervals(List<JsonNode> userInputIntervals,
String minTime,
String maxTime) {
if (userInputIntervals.isEmpty()) {
String interval = String.format("%s/%s", minTime, maxTime);
return Stream.of(interval).collect(Collectors.toList());
}
JsonNode firstIntervalNode = userInputIntervals.get(0);
String interval = firstIntervalNode.asText();
return Stream.of(interval).collect(Collectors.toList());
}
private String removeIntervalFilter(String filter, List<JsonNode> userInputIntervals) {
for (JsonNode intervalNode : userInputIntervals) {
String interval = intervalNode.asText();
String intervalSubString1 = ",{\"eventInterval\":\"" + interval + "\"}";
String intervalSubString2 = "{\"eventInterval\":\"" + interval + "\"},";
filter = filter.replace(intervalSubString1, "");
filter = filter.replace(intervalSubString2, "");
}
return filter;
}
}