blob: 94e44f5e5fe16e840d69007de23b6fca9375cfe1 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.drill.jdbc.test;
import java.nio.file.Paths;
import java.sql.Connection;
import java.sql.ResultSet;
import java.sql.ResultSetMetaData;
import java.sql.SQLException;
import java.sql.Statement;
import java.util.List;
import java.util.concurrent.TimeUnit;
import org.apache.drill.test.TestTools;
import org.apache.drill.jdbc.Driver;
import org.apache.drill.jdbc.JdbcTestBase;
import org.apache.drill.categories.JdbcTest;
import org.junit.Assert;
import org.junit.BeforeClass;
import org.junit.Ignore;
import org.junit.Rule;
import org.junit.Test;
import org.junit.experimental.categories.Category;
import org.junit.rules.TestRule;
import com.google.common.base.Stopwatch;
import com.google.common.collect.Lists;
@Category(JdbcTest.class)
public class TestJdbcDistQuery extends JdbcTestBase {
private static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(TestJdbcDistQuery.class);
// Set a timeout unless we're debugging.
@Rule
public TestRule TIMEOUT = TestTools.getTimeoutRule(50000);
static{
Driver.load();
}
@BeforeClass
public static void setup() {
dirTestWatcher.copyFileToRoot(Paths.get("sample-data"));
}
// TODO: Purge nextUntilEnd(...) and calls when remaining fragment race
// conditions are fixed (not just DRILL-2245 fixes).
/**
* Calls {@link ResultSet#next} on given {@code ResultSet} until it returns
* false. (For TEMPORARY workaround for query cancelation race condition.)
*/
private static void nextUntilEnd(final ResultSet resultSet) throws SQLException {
while (resultSet.next()) {
}
}
@Test
public void testSimpleQuerySingleFile() throws Exception{
testQuery("select R_REGIONKEY, R_NAME from dfs.`sample-data/regionsSF/`");
}
@Test
public void testSimpleQueryMultiFile() throws Exception{
testQuery("select R_REGIONKEY, R_NAME from dfs.`sample-data/regionsMF/`");
}
@Test
public void testWhereOverSFile() throws Exception{
testQuery("select R_REGIONKEY, R_NAME from dfs.`sample-data/regionsSF/` WHERE R_REGIONKEY = 1");
}
@Test
public void testWhereOverMFile() throws Exception{
testQuery("select R_REGIONKEY, R_NAME from dfs.`sample-data/regionsMF/` WHERE R_REGIONKEY = 1");
}
@Test
public void testAggSingleFile() throws Exception{
testQuery("select R_REGIONKEY from dfs.`sample-data/regionsSF/` group by R_REGIONKEY");
}
@Test
public void testAggMultiFile() throws Exception{
testQuery("select R_REGIONKEY from dfs.`sample-data/regionsMF/` group by R_REGIONKEY");
}
@Test
public void testAggOrderByDiffGKeyMultiFile() throws Exception{
testQuery("select R_REGIONKEY, SUM(cast(R_REGIONKEY AS int)) As S "
+ "from dfs.`sample-data/regionsMF/` "
+ "group by R_REGIONKEY ORDER BY S");
}
@Test
public void testAggOrderBySameGKeyMultiFile() throws Exception{
testQuery("select R_REGIONKEY, SUM(cast(R_REGIONKEY AS int)) As S "
+ "from dfs.`sample-data/regionsMF/` "
+ "group by R_REGIONKEY "
+ "ORDER BY R_REGIONKEY");
}
@Ignore
@Test
public void testJoinSingleFile() throws Exception{
testQuery("select T1.R_REGIONKEY "
+ "from dfs.`sample-data/regionsSF/` as T1 "
+ "join dfs.`sample-data/nationsSF/` as T2 "
+ "on T1.R_REGIONKEY = T2.N_REGIONKEY");
}
@Ignore
@Test
public void testJoinMultiFile() throws Exception{
testQuery("select T1.R_REGIONKEY "
+ "from dfs.`sample-data/regionsMF/` as T1 "
+ "join dfs.`sample-data/nationsMF/` as T2 "
+ "on T1.R_REGIONKEY = T2.N_REGIONKEY");
}
@Ignore
@Test
public void testJoinMFileWhere() throws Exception{
testQuery("select T1.R_REGIONKEY, T1.R_NAME "
+ "from dfs.`sample-data/regionsMF/` as T1 "
+ "join dfs.`sample-data/nationsMF/` as T2 "
+ "on T1.R_REGIONKEY = T2.N_REGIONKEY "
+ "WHERE T1.R_REGIONKEY = 3 ");
}
@Test
//NPE at ExternalSortBatch.java : 151
public void testSortSingleFile() throws Exception{
testQuery("select R_REGIONKEY "
+ "from dfs.`sample-data/regionsSF/` "
+ "order by R_REGIONKEY");
}
@Test
//NPE at ExternalSortBatch.java : 151
public void testSortMultiFile() throws Exception{
testQuery("select R_REGIONKEY "
+ "from dfs.`sample-data/regionsMF/` "
+ "order by R_REGIONKEY");
}
@Test
public void testSortMFileWhere() throws Exception{
testQuery("select R_REGIONKEY "
+ "from dfs.`sample-data/regionsMF/` "
+ "WHERE R_REGIONKEY = 1 "
+ "order by R_REGIONKEY");
}
@Ignore
@Test
public void testJoinAggSortWhere() throws Exception{
testQuery("select T1.R_REGIONKEY, COUNT(1) as CNT "
+ "from dfs.`sample-data/regionsMF/` as T1 "
+ "join dfs.`sample-data/nationsMF/` as T2 "
+ "on T1.R_REGIONKEY = T2.N_REGIONKEY "
+ "WHERE T1.R_REGIONKEY = 3 "
+ "GROUP BY T1.R_REGIONKEY "
+ "ORDER BY T1.R_REGIONKEY");
}
@Test
public void testSelectLimit() throws Exception{
testQuery("select R_REGIONKEY, R_NAME "
+ "from dfs.`sample-data/regionsMF/` "
+ "limit 2");
}
private void testQuery(String sql) throws Exception {
final StringBuilder sb = new StringBuilder();
boolean success = false;
try (Connection c = connect()) {
// ???? TODO: What is this currently redundant one-time loop for? (If
// it's kept around to make it easy to switch to looping multiple times
// (e.g., for debugging) then define a constant field or local variable
// for the number of iterations.)
boolean first = true;
for (int x = 0; x < 1; x++) {
Stopwatch watch = Stopwatch.createStarted();
Statement s = c.createStatement();
ResultSet r = s.executeQuery(sql);
ResultSetMetaData md = r.getMetaData();
if (first) {
for (int i = 1; i <= md.getColumnCount(); i++) {
sb.append(md.getColumnName(i));
sb.append('\t');
}
sb.append('\n');
first = false;
}
while (r.next()) {
md = r.getMetaData();
for (int i = 1; i <= md.getColumnCount(); i++) {
sb.append(r.getObject(i));
sb.append('\t');
}
sb.append('\n');
}
sb.append(String.format("Query completed in %d millis.\n", watch.elapsed(TimeUnit.MILLISECONDS)));
}
sb.append("\n\n\n");
success = true;
} finally {
if (!success) {
Thread.sleep(2000);
}
}
logger.info(sb.toString());
}
@Test
public void testSchemaForEmptyResultSet() throws Exception {
String query = "select fullname, occupation, postal_code from cp.`customer.json` where 0 = 1";
try (Connection c = connect()) {
Statement s = c.createStatement();
ResultSet r = s.executeQuery(query);
ResultSetMetaData md = r.getMetaData();
List<String> columns = Lists.newArrayList();
for (int i = 1; i <= md.getColumnCount(); i++) {
columns.add(md.getColumnName(i));
}
String[] expected = {"fullname", "occupation", "postal_code"};
Assert.assertEquals(3, md.getColumnCount());
Assert.assertArrayEquals(expected, columns.toArray());
// TODO: Purge nextUntilEnd(...) and calls when remaining fragment race
// conditions are fixed (not just DRILL-2245 fixes).
nextUntilEnd(r);
}
}
}