| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.drill.exec.store.mongo; |
| |
| import com.fasterxml.jackson.core.type.TypeReference; |
| import com.mongodb.ConnectionString; |
| import com.mongodb.MongoClientSettings; |
| import com.mongodb.client.MongoClient; |
| import com.mongodb.MongoCredential; |
| import com.mongodb.ServerAddress; |
| import com.mongodb.client.MongoClients; |
| import org.apache.calcite.plan.Convention; |
| import org.apache.calcite.plan.RelOptRule; |
| import org.apache.calcite.schema.SchemaPlus; |
| import org.apache.drill.common.JSONOptions; |
| import org.apache.drill.common.exceptions.DrillRuntimeException; |
| import org.apache.drill.exec.ops.OptimizerRulesContext; |
| import org.apache.drill.exec.physical.base.AbstractGroupScan; |
| import org.apache.drill.exec.planner.PlannerPhase; |
| import org.apache.drill.exec.server.DrillbitContext; |
| import org.apache.drill.exec.store.AbstractStoragePlugin; |
| import org.apache.drill.exec.store.PluginRulesProviderImpl; |
| import org.apache.drill.exec.store.SchemaConfig; |
| import org.apache.drill.exec.store.StoragePluginRulesSupplier; |
| import org.apache.drill.exec.store.mongo.plan.MongoPluginImplementor; |
| import org.apache.drill.exec.store.mongo.schema.MongoSchemaFactory; |
| import org.apache.drill.common.logical.security.CredentialsProvider; |
| import org.apache.drill.exec.store.plan.rel.PluginRel; |
| import org.apache.drill.exec.store.security.HadoopCredentialsProvider; |
| import org.apache.drill.common.logical.security.PlainCredentialsProvider; |
| import org.apache.drill.exec.store.security.UsernamePasswordCredentials; |
| import com.google.common.cache.Cache; |
| import com.google.common.cache.CacheBuilder; |
| import com.google.common.cache.RemovalListener; |
| import com.google.common.cache.RemovalNotification; |
| import com.google.common.collect.ImmutableMap; |
| import com.google.common.collect.Lists; |
| import org.slf4j.Logger; |
| import org.slf4j.LoggerFactory; |
| |
| import java.io.IOException; |
| import java.net.URLEncoder; |
| import java.util.Collections; |
| import java.util.List; |
| import java.util.Optional; |
| import java.util.Set; |
| import java.util.concurrent.ExecutionException; |
| import java.util.concurrent.TimeUnit; |
| |
| public class MongoStoragePlugin extends AbstractStoragePlugin { |
| private static final Logger logger = LoggerFactory.getLogger(MongoStoragePlugin.class); |
| |
| private final MongoStoragePluginConfig mongoConfig; |
| private final MongoSchemaFactory schemaFactory; |
| private final Cache<MongoCnxnKey, MongoClient> addressClientMap; |
| private final ConnectionString clientURI; |
| private final StoragePluginRulesSupplier storagePluginRulesSupplier; |
| |
| public MongoStoragePlugin( |
| MongoStoragePluginConfig mongoConfig, |
| DrillbitContext context, |
| String name) { |
| super(context, name); |
| this.mongoConfig = mongoConfig; |
| String connection = addCredentialsFromCredentialsProvider(this.mongoConfig.getConnection(), name); |
| this.clientURI = new ConnectionString(connection); |
| this.addressClientMap = CacheBuilder.newBuilder() |
| .expireAfterAccess(24, TimeUnit.HOURS) |
| .removalListener(new AddressCloser()) |
| .build(); |
| this.schemaFactory = new MongoSchemaFactory(this, name); |
| this.storagePluginRulesSupplier = storagePluginRulesSupplier(name, mongoConfig); |
| } |
| |
| private static StoragePluginRulesSupplier storagePluginRulesSupplier(String name, MongoStoragePluginConfig mongoConfig) { |
| Convention convention = new Convention.Impl("MONGO." + name, PluginRel.class); |
| return StoragePluginRulesSupplier.builder() |
| .rulesProvider(new PluginRulesProviderImpl(convention, MongoPluginImplementor::new)) |
| .supportsProjectPushdown(mongoConfig.getPluginOptimizations().isSupportsProjectPushdown()) |
| .supportsSortPushdown(mongoConfig.getPluginOptimizations().isSupportsSortPushdown()) |
| .supportsAggregatePushdown(mongoConfig.getPluginOptimizations().isSupportsAggregatePushdown()) |
| .supportsFilterPushdown(mongoConfig.getPluginOptimizations().isSupportsFilterPushdown()) |
| .supportsLimitPushdown(mongoConfig.getPluginOptimizations().isSupportsLimitPushdown()) |
| .supportsUnionPushdown(mongoConfig.getPluginOptimizations().isSupportsUnionPushdown()) |
| .convention(convention) |
| .build(); |
| } |
| |
| private String addCredentialsFromCredentialsProvider(String connection, String name) { |
| ConnectionString parsed = new ConnectionString(connection); |
| if (parsed.getCredential() == null) { |
| Optional<UsernamePasswordCredentials> credentials = getUsernamePasswordCredentials(name); |
| try { |
| // The default connection has the name "mongo" but multiple connections can be added; |
| // each will need their own credentials. |
| if (credentials.isPresent()) { |
| String username = URLEncoder.encode(credentials.get().getUsername(), "UTF-8"); |
| String password = URLEncoder.encode(credentials.get().getPassword(), "UTF-8"); |
| return connection.replaceFirst("://", |
| String.format("://%s:%s@", username, password)); |
| } |
| } catch (IOException e) { |
| logger.error("Error fetching mongodb username and password from configuration", e); |
| } |
| } |
| return connection; |
| } |
| |
| private Optional<UsernamePasswordCredentials> getUsernamePasswordCredentials(String name) { |
| CredentialsProvider credentialsProvider = mongoConfig.getCredentialsProvider(); |
| // for the case if empty credentials, tries to obtain credentials using HadoopCredentialsProvider |
| if (credentialsProvider == null || credentialsProvider == PlainCredentialsProvider.EMPTY_CREDENTIALS_PROVIDER) { |
| credentialsProvider = new HadoopCredentialsProvider( |
| ImmutableMap.of( |
| UsernamePasswordCredentials.USERNAME, |
| DrillMongoConstants.STORE_CONFIG_PREFIX + name + DrillMongoConstants.USERNAME_CONFIG_SUFFIX, |
| UsernamePasswordCredentials.PASSWORD, |
| DrillMongoConstants.STORE_CONFIG_PREFIX + name + DrillMongoConstants.PASSWORD_CONFIG_SUFFIX)); |
| } |
| return new UsernamePasswordCredentials.Builder() |
| .setCredentialsProvider(credentialsProvider). |
| build(); |
| } |
| |
| @Override |
| public MongoStoragePluginConfig getConfig() { |
| return mongoConfig; |
| } |
| |
| @Override |
| public void registerSchemas(SchemaConfig schemaConfig, SchemaPlus parent) { |
| schemaFactory.registerSchemas(schemaConfig, parent); |
| } |
| |
| @Override |
| public boolean supportsRead() { |
| return true; |
| } |
| |
| @Override |
| public AbstractGroupScan getPhysicalScan(String userName, JSONOptions selection) throws IOException { |
| MongoScanSpec mongoScanSpec = selection.getListWith(new TypeReference<MongoScanSpec>() { |
| }); |
| return new MongoGroupScan(userName, this, mongoScanSpec, null, false); |
| } |
| |
| @Override |
| public Set<? extends RelOptRule> getOptimizerRules(OptimizerRulesContext optimizerContext, PlannerPhase phase) { |
| switch (phase) { |
| case PHYSICAL: |
| case LOGICAL: |
| return storagePluginRulesSupplier.getOptimizerRules(); |
| case LOGICAL_PRUNE_AND_JOIN: |
| case LOGICAL_PRUNE: |
| case PARTITION_PRUNING: |
| case JOIN_PLANNING: |
| default: |
| return Collections.emptySet(); |
| } |
| } |
| |
| public Convention convention() { |
| return storagePluginRulesSupplier.convention(); |
| } |
| |
| private static class AddressCloser implements |
| RemovalListener<MongoCnxnKey, MongoClient> { |
| @Override |
| public synchronized void onRemoval( |
| RemovalNotification<MongoCnxnKey, MongoClient> removal) { |
| removal.getValue().close(); |
| logger.debug("Closed connection to {}.", removal.getKey().toString()); |
| } |
| } |
| |
| public MongoClient getClient() { |
| List<String> hosts = clientURI.getHosts(); |
| List<ServerAddress> addresses = Lists.newArrayList(); |
| for (String host : hosts) { |
| addresses.add(new ServerAddress(host)); |
| } |
| return getClient(addresses); |
| } |
| |
| public synchronized MongoClient getClient(List<ServerAddress> addresses) { |
| // Take the first replica from the replicated servers |
| ServerAddress serverAddress = addresses.get(0); |
| MongoCredential credential = clientURI.getCredential(); |
| String userName = credential == null ? null : credential.getUserName(); |
| MongoCnxnKey key = new MongoCnxnKey(serverAddress, userName); |
| try { |
| return addressClientMap.get(key, () -> { |
| MongoClientSettings.Builder settings; |
| if (clientURI.isSrvProtocol()) { |
| settings = MongoClientSettings.builder().applyConnectionString(clientURI); |
| logger.info("Created srv protocol connection to {}.", key); |
| } else { |
| settings = MongoClientSettings.builder().applyToClusterSettings(builder -> builder.hosts(addresses)); |
| if (credential != null) { |
| settings.credential(credential); |
| } |
| logger.info("Created connection to {}.", key); |
| } |
| logger.info("Number of open connections {}.", addressClientMap.size() + 1); // include this created |
| return MongoClients.create(settings.build()); |
| }); |
| } catch (ExecutionException e) { |
| throw new DrillRuntimeException(e); |
| } |
| } |
| |
| @Override |
| public void close() { |
| addressClientMap.invalidateAll(); |
| } |
| |
| } |