/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.drill.exec.store.mongo;
import java.io.IOException;
import java.util.Arrays;
import java.util.Collections;
import java.util.List;
import java.util.Set;
import java.util.concurrent.TimeUnit;
import org.apache.calcite.schema.SchemaPlus;
import org.apache.drill.common.JSONOptions;
import org.apache.drill.common.exceptions.ExecutionSetupException;
import org.apache.drill.exec.ops.OptimizerRulesContext;
import org.apache.drill.exec.physical.base.AbstractGroupScan;
import org.apache.drill.exec.server.DrillbitContext;
import org.apache.drill.exec.store.AbstractStoragePlugin;
import org.apache.drill.exec.store.SchemaConfig;
import org.apache.drill.exec.store.StoragePluginOptimizerRule;
import org.apache.drill.exec.store.mongo.schema.MongoSchemaFactory;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import com.fasterxml.jackson.core.type.TypeReference;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.google.common.cache.Cache;
import com.google.common.cache.CacheBuilder;
import com.google.common.cache.RemovalListener;
import com.google.common.cache.RemovalNotification;
import com.google.common.collect.ImmutableSet;
import com.google.common.collect.Lists;
import com.mongodb.MongoClient;
import com.mongodb.MongoClientURI;
import com.mongodb.MongoCredential;
import com.mongodb.ServerAddress;
public class MongoStoragePlugin extends AbstractStoragePlugin {
static final Logger logger = LoggerFactory
.getLogger(MongoStoragePlugin.class);
private final DrillbitContext context;
private final MongoStoragePluginConfig mongoConfig;
private final MongoSchemaFactory schemaFactory;
private final Cache<MongoCnxnKey, MongoClient> addressClientMap;
private final MongoClientURI clientURI;
public MongoStoragePlugin(MongoStoragePluginConfig mongoConfig,
DrillbitContext context, String name) throws IOException,
ExecutionSetupException {
this.context = context;
this.mongoConfig = mongoConfig;
this.clientURI = new MongoClientURI(this.mongoConfig.getConnection());
this.addressClientMap = CacheBuilder.newBuilder()
.expireAfterAccess(24, TimeUnit.HOURS)
.removalListener(new AddressCloser()).build();
this.schemaFactory = new MongoSchemaFactory(this, name);
}
public DrillbitContext getContext() {
return this.context;
}
@Override
public MongoStoragePluginConfig getConfig() {
return mongoConfig;
}
@Override
public void registerSchemas(SchemaConfig schemaConfig, SchemaPlus parent) throws IOException {
schemaFactory.registerSchemas(schemaConfig, parent);
}
@Override
public boolean supportsRead() {
return true;
}
@Override
public AbstractGroupScan getPhysicalScan(String userName, JSONOptions selection) throws IOException {
MongoScanSpec mongoScanSpec = selection.getListWith(new ObjectMapper(), new TypeReference<MongoScanSpec>() {});
return new MongoGroupScan(userName, this, mongoScanSpec, null);
}
@Override
public Set<StoragePluginOptimizerRule> getPhysicalOptimizerRules(OptimizerRulesContext optimizerRulesContext) {
return ImmutableSet.of(MongoPushDownFilterForScan.INSTANCE);
}
private class AddressCloser implements
RemovalListener<MongoCnxnKey, MongoClient> {
@Override
public synchronized void onRemoval(
RemovalNotification<MongoCnxnKey, MongoClient> removal) {
removal.getValue().close();
logger.debug("Closed connection to {}.", removal.getKey().toString());
}
}
public MongoClient getClient(String host) {
return getClient(Collections.singletonList(new ServerAddress(host)));
}
public MongoClient getClient() {
List<String> hosts = clientURI.getHosts();
List<ServerAddress> addresses = Lists.newArrayList();
for (String host : hosts) {
addresses.add(new ServerAddress(host));
}
return getClient(addresses);
}
public synchronized MongoClient getClient(List<ServerAddress> addresses) {
// Take the first replica from the replicated servers
final ServerAddress serverAddress = addresses.get(0);
final MongoCredential credential = clientURI.getCredentials();
String userName = credential == null ? null : credential.getUserName();
MongoCnxnKey key = new MongoCnxnKey(serverAddress, userName);
MongoClient client = addressClientMap.getIfPresent(key);
if (client == null) {
if (credential != null) {
List<MongoCredential> credentialList = Arrays.asList(credential);
client = new MongoClient(addresses, credentialList, clientURI.getOptions());
} else {
client = new MongoClient(addresses, clientURI.getOptions());
}
addressClientMap.put(key, client);
logger.debug("Created connection to {}.", key.toString());
logger.debug("Number of open connections {}.", addressClientMap.size());
}
return client;
}
@Override
public void close() throws Exception {
addressClientMap.invalidateAll();
}
}