blob: 054051261413c42e80cfda70e544a7fd28aa3f78 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.drill.exec.store.druid;
import com.fasterxml.jackson.annotation.JacksonInject;
import com.fasterxml.jackson.annotation.JsonCreator;
import com.fasterxml.jackson.annotation.JsonProperty;
import com.fasterxml.jackson.annotation.JsonTypeName;
import com.fasterxml.jackson.annotation.JsonIgnore;
import org.apache.drill.common.PlanStringBuilder;
import org.apache.drill.common.expression.SchemaPath;
import org.apache.drill.common.logical.StoragePluginConfig;
import org.apache.drill.exec.physical.base.AbstractBase;
import org.apache.drill.exec.physical.base.PhysicalOperator;
import org.apache.drill.exec.physical.base.PhysicalVisitor;
import org.apache.drill.exec.physical.base.SubScan;
import org.apache.drill.exec.store.StoragePluginRegistry;
import org.apache.drill.exec.store.druid.common.DruidFilter;
import com.google.common.base.Preconditions;
import java.util.Iterator;
import java.util.LinkedList;
import java.util.List;
import static java.util.Collections.emptyIterator;
/**
* A Class containing information to read a single druid data source.
*/
@JsonTypeName("druid-datasource-scan")
public class DruidSubScan extends AbstractBase implements SubScan {
public static final String OPERATOR_TYPE = "DRUID_SUB_SCAN";
@JsonIgnore
private final DruidStoragePlugin druidStoragePlugin;
private final List<DruidSubScanSpec> scanSpec;
private final List<SchemaPath> columns;
private final int maxRecordsToRead;
@JsonCreator
public DruidSubScan(@JacksonInject StoragePluginRegistry registry,
@JsonProperty("userName") String userName,
@JsonProperty("config") StoragePluginConfig config,
@JsonProperty("scanSpec") LinkedList<DruidSubScanSpec> datasourceScanSpecList,
@JsonProperty("columns") List<SchemaPath> columns,
@JsonProperty("maxRecordsToRead") int maxRecordsToRead) {
super(userName);
druidStoragePlugin = registry.resolve(config, DruidStoragePlugin.class);
this.scanSpec = datasourceScanSpecList;
this.columns = columns;
this.maxRecordsToRead = maxRecordsToRead;
}
public DruidSubScan(String userName,
DruidStoragePlugin plugin,
List<DruidSubScanSpec> dataSourceInfoList,
List<SchemaPath> columns,
int maxRecordsToRead) {
super(userName);
this.druidStoragePlugin = plugin;
this.scanSpec = dataSourceInfoList;
this.columns = columns;
this.maxRecordsToRead = maxRecordsToRead;
}
@Override
public <T, X, E extends Throwable> T accept(PhysicalVisitor<T, X, E> physicalVisitor, X value) throws E {
return physicalVisitor.visitSubScan(this, value);
}
@JsonIgnore
public List<DruidSubScanSpec> getScanSpec() {
return scanSpec;
}
public List<SchemaPath> getColumns() {
return columns;
}
public int getMaxRecordsToRead() { return maxRecordsToRead; }
@JsonIgnore
@Override
public boolean isExecutable() {
return false;
}
@JsonIgnore
public DruidStoragePlugin getStorageEngine(){
return druidStoragePlugin;
}
@Override
public PhysicalOperator getNewWithChildren(List<PhysicalOperator> children) {
Preconditions.checkArgument(children.isEmpty());
return new DruidSubScan(getUserName(), druidStoragePlugin, scanSpec, columns, maxRecordsToRead);
}
@JsonIgnore
@Override
public String getOperatorType() {
return OPERATOR_TYPE;
}
@Override
public Iterator<PhysicalOperator> iterator() {
return emptyIterator();
}
public static class DruidSubScanSpec {
protected final String dataSourceName;
protected final DruidFilter filter;
protected final long dataSourceSize;
protected final String maxTime;
protected final String minTime;
@JsonCreator
public DruidSubScanSpec(@JsonProperty("dataSourceName") String dataSourceName,
@JsonProperty("filter") DruidFilter filter,
@JsonProperty("dataSourceSize") long dataSourceSize,
@JsonProperty("minTime") String minTime,
@JsonProperty("maxTime") String maxTime) {
this.dataSourceName = dataSourceName;
this.filter = filter;
this.dataSourceSize = dataSourceSize;
this.minTime = minTime;
this.maxTime = maxTime;
}
public String getDataSourceName() {
return dataSourceName;
}
public DruidFilter getFilter() { return filter; }
public long getDataSourceSize() {
return dataSourceSize;
}
public String getMinTime() {
return minTime;
}
public String getMaxTime() {
return maxTime;
}
@Override
public String toString() {
return new PlanStringBuilder(this)
.field("dataSourceName", dataSourceName)
.field("filter", filter)
.field("dataSourceSize", dataSourceSize)
.field("minTime", minTime)
.field("maxTime", maxTime)
.toString();
}
}
}