blob: f4986839508271541a181e0797dd69e8743b8059 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to you under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.phoenix.end2end;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertNotNull;
import static org.junit.Assert.assertTrue;
import java.io.BufferedReader;
import java.io.File;
import java.io.FileOutputStream;
import java.io.IOException;
import java.io.InputStreamReader;
import java.io.PrintWriter;
import java.io.Reader;
import java.io.Writer;
import java.lang.reflect.Field;
import java.nio.file.Paths;
import java.security.PrivilegedAction;
import java.util.AbstractMap;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.List;
import java.util.Map.Entry;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.Executors;
import java.util.concurrent.TimeUnit;
import java.util.stream.Collectors;
import org.apache.commons.io.FileUtils;
import org.apache.commons.io.IOUtils;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hbase.HBaseTestingUtility;
import org.apache.hadoop.hbase.HConstants;
import org.apache.hadoop.hbase.LocalHBaseCluster;
import org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
import org.apache.hadoop.hbase.security.HBaseKerberosUtils;
import org.apache.hadoop.hbase.security.token.TokenProvider;
import org.apache.hadoop.hbase.util.FSUtils;
import org.apache.hadoop.hdfs.DFSConfigKeys;
import org.apache.hadoop.http.HttpConfig;
import org.apache.hadoop.minikdc.MiniKdc;
import org.apache.hadoop.security.UserGroupInformation;
import org.apache.hadoop.security.authentication.util.KerberosName;
import org.apache.phoenix.query.ConfigurationFactory;
import org.apache.phoenix.queryserver.QueryServerProperties;
import org.apache.phoenix.queryserver.server.QueryServer;
import org.apache.phoenix.util.InstanceResolver;
import org.apache.phoenix.util.ThinClientUtil;
import org.junit.AfterClass;
import org.junit.Assume;
import org.junit.BeforeClass;
import org.junit.Ignore;
import org.junit.Test;
import org.junit.experimental.categories.Category;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
/**
* This integration test stands up a secured PQS and runs Python code against it. See supporting
* files in phoenix-queryserver/src/it/bin.
*/
@Category(NeedsOwnMiniClusterTest.class)
public class SecureQueryServerPhoenixDBIT {
private static enum Kdc {
MIT,
HEIMDAL;
}
private static final Logger LOG = LoggerFactory.getLogger(SecureQueryServerPhoenixDBIT.class);
private static final File TEMP_DIR = new File(getTempDirForClass());
private static final File KEYTAB_DIR = new File(TEMP_DIR, "keytabs");
private static final List<File> USER_KEYTAB_FILES = new ArrayList<>();
private static final String SPNEGO_PRINCIPAL = "HTTP/localhost";
private static final String PQS_PRINCIPAL = "phoenixqs/localhost";
private static final String SERVICE_PRINCIPAL = "securecluster/localhost";
private static File KEYTAB;
private static MiniKdc KDC;
private static HBaseTestingUtility UTIL = new HBaseTestingUtility();
private static LocalHBaseCluster HBASE_CLUSTER;
private static int NUM_CREATED_USERS;
private static ExecutorService PQS_EXECUTOR;
private static QueryServer PQS;
private static int PQS_PORT;
private static String PQS_URL;
private static String getTempDirForClass() {
StringBuilder sb = new StringBuilder(32);
sb.append(System.getProperty("user.dir")).append(File.separator);
sb.append("target").append(File.separator);
sb.append(SecureQueryServerPhoenixDBIT.class.getSimpleName());
return sb.toString();
}
private static void updateDefaultRealm() throws Exception {
// (at least) one other phoenix test triggers the caching of this field before the KDC is up
// which causes principal parsing to fail.
Field f = KerberosName.class.getDeclaredField("defaultRealm");
f.setAccessible(true);
// Default realm for MiniKDC
f.set(null, "EXAMPLE.COM");
}
private static void createUsers(int numUsers) throws Exception {
assertNotNull("KDC is null, was setup method called?", KDC);
NUM_CREATED_USERS = numUsers;
for (int i = 1; i <= numUsers; i++) {
String principal = "user" + i;
File keytabFile = new File(KEYTAB_DIR, principal + ".keytab");
KDC.createPrincipal(keytabFile, principal);
USER_KEYTAB_FILES.add(keytabFile);
}
}
private static Entry<String,File> getUser(int offset) {
if (!(offset > 0 && offset <= NUM_CREATED_USERS)) {
throw new IllegalArgumentException();
}
return new AbstractMap.SimpleImmutableEntry<String,File>("user" + offset, USER_KEYTAB_FILES.get(offset - 1));
}
/**
* Setup the security configuration for hdfs.
*/
private static void setHdfsSecuredConfiguration(Configuration conf) throws Exception {
// Set principal+keytab configuration for HDFS
conf.set(DFSConfigKeys.DFS_NAMENODE_KERBEROS_PRINCIPAL_KEY, SERVICE_PRINCIPAL + "@" + KDC.getRealm());
conf.set(DFSConfigKeys.DFS_NAMENODE_KEYTAB_FILE_KEY, KEYTAB.getAbsolutePath());
conf.set(DFSConfigKeys.DFS_DATANODE_KERBEROS_PRINCIPAL_KEY, SERVICE_PRINCIPAL + "@" + KDC.getRealm());
conf.set(DFSConfigKeys.DFS_DATANODE_KEYTAB_FILE_KEY, KEYTAB.getAbsolutePath());
conf.set(DFSConfigKeys.DFS_WEB_AUTHENTICATION_KERBEROS_PRINCIPAL_KEY, SPNEGO_PRINCIPAL + "@" + KDC.getRealm());
// Enable token access for HDFS blocks
conf.setBoolean(DFSConfigKeys.DFS_BLOCK_ACCESS_TOKEN_ENABLE_KEY, true);
// Only use HTTPS (required because we aren't using "secure" ports)
conf.set(DFSConfigKeys.DFS_HTTP_POLICY_KEY, HttpConfig.Policy.HTTPS_ONLY.name());
// Bind on localhost for spnego to have a chance at working
conf.set(DFSConfigKeys.DFS_NAMENODE_HTTPS_ADDRESS_KEY, "localhost:0");
conf.set(DFSConfigKeys.DFS_DATANODE_HTTPS_ADDRESS_KEY, "localhost:0");
// Generate SSL certs
File keystoresDir = new File(UTIL.getDataTestDir("keystore").toUri().getPath());
keystoresDir.mkdirs();
String sslConfDir = TlsUtil.getClasspathDir(SecureQueryServerPhoenixDBIT.class);
TlsUtil.setupSSLConfig(keystoresDir.getAbsolutePath(), sslConfDir, conf, false);
// Magic flag to tell hdfs to not fail on using ports above 1024
conf.setBoolean("ignore.secure.ports.for.testing", true);
}
private static void ensureIsEmptyDirectory(File f) throws IOException {
if (f.exists()) {
if (f.isDirectory()) {
FileUtils.deleteDirectory(f);
} else {
assertTrue("Failed to delete keytab directory", f.delete());
}
}
assertTrue("Failed to create keytab directory", f.mkdirs());
}
/**
* Verifies that there is a python Executable on the PATH
*/
private static void checkForCommandOnPath(String command) throws Exception {
Process runPythonProcess = new ProcessBuilder(Arrays.asList("which", command)).start();
BufferedReader processOutput = new BufferedReader(new InputStreamReader(runPythonProcess.getInputStream()));
BufferedReader processError = new BufferedReader(new InputStreamReader(runPythonProcess.getErrorStream()));
int exitCode = runPythonProcess.waitFor();
// dump stdout and stderr
while (processOutput.ready()) {
LOG.info(processOutput.readLine());
}
while (processError.ready()) {
LOG.error(processError.readLine());
}
Assume.assumeTrue("Could not find '" + command + "' on the PATH", exitCode == 0);
}
/**
* Setup and start kerberos, hbase
*/
@BeforeClass
public static synchronized void setUp() throws Exception {
checkForCommandOnPath("python");
checkForCommandOnPath("virtualenv");
checkForCommandOnPath("kinit");
final Configuration conf = UTIL.getConfiguration();
// Ensure the dirs we need are created/empty
ensureIsEmptyDirectory(TEMP_DIR);
ensureIsEmptyDirectory(KEYTAB_DIR);
KEYTAB = new File(KEYTAB_DIR, "test.keytab");
// Start a MiniKDC
KDC = UTIL.setupMiniKdc(KEYTAB);
// Create a service principal and spnego principal in one keytab
// NB. Due to some apparent limitations between HDFS and HBase in the same JVM, trying to
// use separate identies for HBase and HDFS results in a GSS initiate error. The quick
// solution is to just use a single "service" principal instead of "hbase" and "hdfs"
// (or "dn" and "nn") per usual.
KDC.createPrincipal(KEYTAB, SPNEGO_PRINCIPAL, PQS_PRINCIPAL, SERVICE_PRINCIPAL);
// Start ZK by hand
UTIL.startMiniZKCluster();
// Create a number of unprivileged users
createUsers(3);
// Set configuration for HBase
HBaseKerberosUtils.setPrincipalForTesting(SERVICE_PRINCIPAL + "@" + KDC.getRealm());
HBaseKerberosUtils.setSecuredConfiguration(conf);
setHdfsSecuredConfiguration(conf);
UserGroupInformation.setConfiguration(conf);
conf.setInt(HConstants.MASTER_PORT, 0);
conf.setInt(HConstants.MASTER_INFO_PORT, 0);
conf.setInt(HConstants.REGIONSERVER_PORT, 0);
conf.setInt(HConstants.REGIONSERVER_INFO_PORT, 0);
conf.setStrings(CoprocessorHost.REGION_COPROCESSOR_CONF_KEY,
TokenProvider.class.getName());
// Secure Phoenix setup
conf.set(QueryServerProperties.QUERY_SERVER_KERBEROS_HTTP_PRINCIPAL_ATTRIB_LEGACY,
SPNEGO_PRINCIPAL + "@" + KDC.getRealm());
conf.set(QueryServerProperties.QUERY_SERVER_HTTP_KEYTAB_FILENAME_ATTRIB,
KEYTAB.getAbsolutePath());
conf.set(QueryServerProperties.QUERY_SERVER_KERBEROS_PRINCIPAL_ATTRIB,
PQS_PRINCIPAL + "@" + KDC.getRealm());
conf.set(QueryServerProperties.QUERY_SERVER_KEYTAB_FILENAME_ATTRIB,
KEYTAB.getAbsolutePath());
conf.setBoolean(QueryServerProperties.QUERY_SERVER_DISABLE_KERBEROS_LOGIN, true);
if(System.getProperty("do.not.randomize.pqs.port") == null) {
conf.setInt(QueryServerProperties.QUERY_SERVER_HTTP_PORT_ATTRIB, 0);
}
// Required so that PQS can impersonate the end-users to HBase
conf.set("hadoop.proxyuser.phoenixqs.groups", "*");
conf.set("hadoop.proxyuser.phoenixqs.hosts", "*");
// Clear the cached singletons so we can inject our own.
InstanceResolver.clearSingletons();
// Make sure the ConnectionInfo doesn't try to pull a default Configuration
InstanceResolver.getSingleton(ConfigurationFactory.class, new ConfigurationFactory() {
@Override
public Configuration getConfiguration() {
return conf;
}
@Override
public Configuration getConfiguration(Configuration confToClone) {
Configuration copy = new Configuration(conf);
copy.addResource(confToClone);
return copy;
}
});
updateDefaultRealm();
// Start HDFS
UTIL.startMiniDFSCluster(1);
// Use LocalHBaseCluster to avoid HBaseTestingUtility from doing something wrong
// NB. I'm not actually sure what HTU does incorrect, but this was pulled from some test
// classes in HBase itself. I couldn't get HTU to work myself (2017/07/06)
Path rootdir = UTIL.getDataTestDirOnTestFS(SecureQueryServerPhoenixDBIT.class.getSimpleName());
FSUtils.setRootDir(conf, rootdir);
HBASE_CLUSTER = new LocalHBaseCluster(conf, 1);
HBASE_CLUSTER.startup();
// Then fork a thread with PQS in it.
startQueryServer();
}
private static void startQueryServer() throws Exception {
PQS = new QueryServer(new String[0], UTIL.getConfiguration());
// Get the PQS ident for PQS to use
final UserGroupInformation ugi = UserGroupInformation.loginUserFromKeytabAndReturnUGI(PQS_PRINCIPAL, KEYTAB.getAbsolutePath());
PQS_EXECUTOR = Executors.newSingleThreadExecutor();
// Launch PQS, doing in the Kerberos login instead of letting PQS do it itself (which would
// break the HBase/HDFS logins also running in the same test case).
PQS_EXECUTOR.submit(new Runnable() {
@Override public void run() {
ugi.doAs(new PrivilegedAction<Void>() {
@Override public Void run() {
PQS.run();
return null;
}
});
}
});
PQS.awaitRunning();
PQS_PORT = PQS.getPort();
PQS_URL = ThinClientUtil.getConnectionUrl("localhost", PQS_PORT) + ";authentication=SPNEGO";
}
@AfterClass
public static synchronized void stopKdc() throws Exception {
// Remove our custom ConfigurationFactory for future tests
InstanceResolver.clearSingletons();
if (PQS_EXECUTOR != null) {
PQS.stop();
PQS_EXECUTOR.shutdown();
if (!PQS_EXECUTOR.awaitTermination(5, TimeUnit.SECONDS)) {
LOG.info("PQS didn't exit in 5 seconds, proceeding anyways.");
}
}
if (HBASE_CLUSTER != null) {
HBASE_CLUSTER.shutdown();
HBASE_CLUSTER.join();
}
if (UTIL != null) {
UTIL.shutdownMiniZKCluster();
}
if (KDC != null) {
KDC.stop();
}
}
@Test
public void testBasicReadWrite() throws Exception {
File file = new File(".");
runShellScript("python", Paths.get(file.getAbsolutePath(), "src", "it", "bin", "test_phoenixdb.py").toString());
}
//This takes about 300s, so we are not running this by default
@Test
public void testFullSuite() throws Exception {
Assume.assumeNotNull(System.getProperty("run.full.python.testsuite"));
File file = new File(".");
runShellScript("python", "-m", "unittest", "discover", "-v", "-s", Paths.get(file.getAbsolutePath(), "..","python-phoenixdb").toString());
}
@Test
//Quick and dirty way start up a local Phoenix+PQS instance for testing against
//When started, this will write a setup script into target/krb_setup.sh
//If you source that file, it should log you into kerberos with the test principal,
//and set the environment so that the python unit tests can run against this instance.
//You'll need to kill the test manually
public void startLocalPQS() throws Exception {
Assume.assumeNotNull(System.getProperty("start.secure.pqs"));
runShellScript("sleep", "86400");
}
public void runShellScript(String ... testCli) throws Exception {
final Entry<String,File> user1 = getUser(1);
String currentDirectory;
File file = new File(".");
currentDirectory = file.getAbsolutePath();
LOG.debug("Current working directory : "+currentDirectory);
LOG.debug("PQS_PORT:" + PQS_PORT);
LOG.debug("PQS_URL: " + PQS_URL);
ArrayList<String> cmdList = new ArrayList<>();
// This assumes the test is being run from phoenix/phoenix-queryserver
cmdList.add(Paths.get(currentDirectory, "src", "it", "bin", "test_phoenixdb.sh").toString());
cmdList.add(Paths.get(currentDirectory, "..", "python-phoenixdb").toString());
cmdList.add(user1.getKey() + "@" + KDC.getRealm());
cmdList.add(user1.getValue().getAbsolutePath());
final String osName = System.getProperty("os.name").toLowerCase();
final Kdc kdcType;
final String kdcImpl = System.getProperty("PHOENIXDB_KDC_IMPL", "");
if (kdcImpl.isEmpty()) {
if (osName.indexOf("mac") >= 0) {
kdcType = Kdc.HEIMDAL;
} else {
kdcType = Kdc.MIT;
}
} else if (kdcImpl.trim().equalsIgnoreCase(Kdc.HEIMDAL.name())) {
kdcType = Kdc.HEIMDAL;
} else {
kdcType = Kdc.MIT;
}
LOG.info("Generating krb5.conf for KDC type:'{}'. OS='{}', PHOENIXDB_KDC_IMPL='{}'", kdcType, osName, kdcImpl);
File krb5ConfFile = null;
switch (kdcType) {
// It appears that we cannot generate a krb5.conf that is compatible with both MIT Kerberos
// and Heimdal Kerberos that works with MiniKdc. MiniKdc forces a choice between either UDP or
// or TCP for the KDC port. If we could have MiniKdc support both UDP and TCP, then we might be
// able to converge on a single krb5.conf for both MIT and Heimdal.
//
// With the below Heimdal configuration, MIT kerberos will fail on a DNS lookup to the hostname
// "tcp/localhost" instead of pulling off the "tcp/" prefix.
case HEIMDAL:
int kdcPort = KDC.getPort();
LOG.info("MINIKDC PORT " + kdcPort);
// Render a Heimdal compatible krb5.conf
// Currently kinit will only try tcp if the KDC is defined as
// kdc = tcp/hostname:port
StringBuilder krb5conf = new StringBuilder();
krb5conf.append("[libdefaults]\n");
krb5conf.append(" default_realm = EXAMPLE.COM\n");
krb5conf.append(" udp_preference_limit = 1\n");
krb5conf.append("\n");
krb5conf.append("[realms]\n");
krb5conf.append(" EXAMPLE.COM = {\n");
krb5conf.append(" kdc = localhost:");
krb5conf.append(kdcPort);
krb5conf.append("\n");
krb5conf.append(" kdc = tcp/localhost:");
krb5conf.append(kdcPort);
krb5conf.append("\n");
krb5conf.append(" }\n");
LOG.info("Writing Heimdal style krb5.conf");
LOG.info(krb5conf.toString());
krb5ConfFile = File.createTempFile("krb5.conf", null);
FileOutputStream fos = new FileOutputStream(krb5ConfFile);
fos.write(krb5conf.toString().getBytes());
fos.close();
LOG.info("krb5.conf written to " + krb5ConfFile.getAbsolutePath());
cmdList.add(krb5ConfFile.getAbsolutePath());
break;
case MIT:
cmdList.add(System.getProperty("java.security.krb5.conf"));
LOG.info("Using miniKDC provided krb5.conf " + KDC.getKrb5conf().getAbsolutePath());
break;
default:
throw new RuntimeException("Unhandled KDC type: " + kdcType);
}
cmdList.add(Integer.toString(PQS_PORT));
cmdList.addAll(Arrays.asList(testCli));
LOG.info("Running command {}", cmdList.stream().collect(Collectors.joining(" ")));
Thread outWriter = null;
Thread errWriter = null;
try {
//This will intersperse that script's output to the maven output, but that's better than
//getting stuck on a full buffer
Process runPythonProcess = new ProcessBuilder(cmdList).start();
BufferedReader processOutput = new BufferedReader(
new InputStreamReader(runPythonProcess.getInputStream()));
outWriter = new Thread(new StreamCopy(processOutput, new PrintWriter(System.out)));
outWriter.start();
BufferedReader processError = new BufferedReader(
new InputStreamReader(runPythonProcess.getErrorStream()));
errWriter = new Thread(new StreamCopy(processError, new PrintWriter(System.err)));
errWriter.start();
int exitCode = runPythonProcess.waitFor();
// Not managed by miniKDC so we have to clean up
if (krb5ConfFile != null)
krb5ConfFile.delete();
assertEquals("Subprocess exited with errors", 0, exitCode);
} finally {
LOG.info("Test exiting");
if (outWriter != null) {
outWriter.stop();
System.out.flush();
}
if (errWriter != null) {
errWriter.stop();
System.err.flush();
}
}
}
private class StreamCopy implements Runnable {
private Reader reader;
private Writer writer;
public StreamCopy(Reader reader, Writer writer) {
this.reader = reader;
this.writer = writer;
}
@Override
public void run() {
try {
while(true) {
IOUtils.copy(reader, writer);
Thread.sleep(10);
}
} catch (Exception e) {
//Just exit
}
}
}
}