| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * <p> |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * <p> |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.atlas.query; |
| |
| import org.apache.atlas.TestModules; |
| import org.apache.atlas.discovery.EntityDiscoveryService; |
| import org.apache.atlas.exception.AtlasBaseException; |
| import org.apache.atlas.model.discovery.AtlasSearchResult; |
| import org.apache.atlas.runner.LocalSolrRunner; |
| import org.testng.annotations.AfterClass; |
| import org.testng.annotations.BeforeClass; |
| import org.testng.annotations.DataProvider; |
| import org.testng.annotations.Guice; |
| import org.testng.annotations.Test; |
| |
| import javax.inject.Inject; |
| |
| import static org.testng.Assert.assertEquals; |
| import static org.testng.Assert.assertNotNull; |
| |
| @Guice(modules = TestModules.TestOnlyModule.class) |
| public class DSLQueriesTest extends BasicTestSetup { |
| @Inject |
| private EntityDiscoveryService discoveryService; |
| |
| @BeforeClass |
| public void setup() throws Exception { |
| LocalSolrRunner.start(); |
| setupTestData(); |
| } |
| |
| @AfterClass |
| public void teardown() throws Exception { |
| LocalSolrRunner.stop(); |
| } |
| |
| @DataProvider(name = "comparisonQueriesProvider") |
| private Object[][] createComparisonQueries() { |
| //create queries the exercise the comparison logic for |
| //all of the different supported data types |
| return new Object[][] { |
| {"Person where (birthday < \"1950-01-01T02:35:58.440Z\" )", 0}, |
| {"Person where (birthday > \"1975-01-01T02:35:58.440Z\" )", 2}, |
| {"Person where (birthday >= \"1975-01-01T02:35:58.440Z\" )", 2}, |
| {"Person where (birthday <= \"1950-01-01T02:35:58.440Z\" )", 0}, |
| {"Person where (birthday = \"1975-01-01T02:35:58.440Z\" )", 0}, |
| {"Person where (birthday != \"1975-01-01T02:35:58.440Z\" )", 4}, |
| |
| {"Person where (hasPets = true)", 2}, |
| {"Person where (hasPets = false)", 2}, |
| {"Person where (hasPets != false)", 2}, |
| {"Person where (hasPets != true)", 2}, |
| |
| {"Person where (numberOfCars > 0)", 2}, |
| {"Person where (numberOfCars > 1)", 1}, |
| {"Person where (numberOfCars >= 1)", 2}, |
| {"Person where (numberOfCars < 2)", 3}, |
| {"Person where (numberOfCars <= 2)", 4}, |
| {"Person where (numberOfCars = 2)", 1}, |
| {"Person where (numberOfCars != 2)", 3}, |
| |
| {"Person where (houseNumber > 0)", 2}, |
| {"Person where (houseNumber > 17)", 1}, |
| {"Person where (houseNumber >= 17)", 2}, |
| {"Person where (houseNumber < 153)", 3}, |
| {"Person where (houseNumber <= 153)", 4}, |
| {"Person where (houseNumber = 17)", 1}, |
| {"Person where (houseNumber != 17)", 3}, |
| |
| {"Person where (carMileage > 0)", 2}, |
| {"Person where (carMileage > 13)", 1}, |
| {"Person where (carMileage >= 13)", 2}, |
| {"Person where (carMileage < 13364)", 3}, |
| {"Person where (carMileage <= 13364)", 4}, |
| {"Person where (carMileage = 13)", 1}, |
| {"Person where (carMileage != 13)", 3}, |
| |
| {"Person where (shares > 0)", 2}, |
| {"Person where (shares > 13)", 2}, |
| {"Person where (shares >= 16000)", 1}, |
| {"Person where (shares < 13364)", 2}, |
| {"Person where (shares <= 15000)", 3}, |
| {"Person where (shares = 15000)", 1}, |
| {"Person where (shares != 1)", 4}, |
| |
| {"Person where (salary > 0)", 2}, |
| {"Person where (salary > 100000)", 2}, |
| {"Person where (salary >= 200000)", 1}, |
| {"Person where (salary < 13364)", 2}, |
| {"Person where (salary <= 150000)", 3}, |
| {"Person where (salary = 12334)", 0}, |
| {"Person where (salary != 12344)", 4}, |
| |
| {"Person where (age > 36)", 1}, |
| {"Person where (age > 49)", 1}, |
| {"Person where (age >= 49)", 1}, |
| {"Person where (age < 50)", 3}, |
| {"Person where (age <= 35)", 2}, |
| {"Person where (age = 35)", 0}, |
| {"Person where (age != 35)", 4} |
| }; |
| } |
| |
| @Test(dataProvider = "comparisonQueriesProvider") |
| public void testComparisonQueries(String query, int expected) throws AtlasBaseException { |
| AtlasSearchResult searchResult = discoveryService.searchUsingDslQuery(query, 25, 0); |
| assertNotNull(searchResult.getEntities()); |
| assertEquals(searchResult.getEntities().size(), expected); |
| } |
| |
| @DataProvider(name = "dslQueriesProvider") |
| private Object[][] createDSLQueries() { |
| return new Object[][]{ |
| {"hive_db as inst where inst.name=\"Reporting\" select inst as id, inst.name", 1}, |
| {"from hive_db as h select h as id", 3}, |
| {"from hive_db", 3}, |
| {"hive_db", 3}, |
| {"hive_db where hive_db.name=\"Reporting\"", 1}, |
| {"hive_db hive_db.name = \"Reporting\"", 1}, |
| {"hive_db where hive_db.name=\"Reporting\" select name, owner", 1}, |
| {"hive_db has name", 3}, |
| {"hive_db, hive_table", 10}, |
| {"View is JdbcAccess", 2}, |
| {"hive_db as db1, hive_table where db1.name = \"Reporting\"", 0}, //Not working - ATLAS-145 |
| // - Final working query -> discoveryService.searchByGremlin("L:{_var_0 = [] as Set;g.V().has(\"__typeName\", \"hive_db\").fill(_var_0);g.V().has(\"__superTypeNames\", \"hive_db\").fill(_var_0);_var_0._().as(\"db1\").in(\"__hive_table.db\").back(\"db1\").and(_().has(\"hive_db.name\", T.eq, \"Reporting\")).toList()}") |
| /* |
| {"hive_db, hive_process has name"}, //Invalid query |
| {"hive_db where hive_db.name=\"Reporting\" and hive_db.createTime < " + System.currentTimeMillis()} |
| */ |
| {"from hive_table", 10}, |
| {"hive_table", 10}, |
| {"hive_table isa Dimension", 3}, |
| {"hive_column where hive_column isa PII", 8}, |
| {"View is Dimension" , 2}, |
| // {"hive_column where hive_column isa PII select hive_column.name", 6}, //Not working - ATLAS-175 |
| {"hive_column select hive_column.name", 37}, |
| {"hive_column select name",37}, |
| {"hive_column where hive_column.name=\"customer_id\"", 6}, |
| {"from hive_table select hive_table.name", 10}, |
| {"hive_db where (name = \"Reporting\")", 1}, |
| {"hive_db where (name = \"Reporting\") select name as _col_0, owner as _col_1", 1}, |
| {"hive_db where hive_db is JdbcAccess", 0}, //Not supposed to work |
| {"hive_db hive_table", 10}, |
| {"hive_db where hive_db has name", 3}, |
| {"hive_db as db1 hive_table where (db1.name = \"Reporting\")", 0}, //Not working -> ATLAS-145 |
| {"hive_db where (name = \"Reporting\") select name as _col_0, (createTime + 1) as _col_1 ", 1}, |
| {"hive_table where (name = \"sales_fact\" and createTime > \"2014-01-01\" ) select name as _col_0, createTime as _col_1 ", 1}, |
| {"hive_table where (name = \"sales_fact\" and createTime >= \"2014-12-11T02:35:58.440Z\" ) select name as _col_0, createTime as _col_1 ", 1}, |
| |
| /* |
| todo: does not work - ATLAS-146 |
| {"hive_db where (name = \"Reporting\") and ((createTime + 1) > 0)"}, |
| {"hive_db as db1 hive_table as tab where ((db1.createTime + 1) > 0) and (db1.name = \"Reporting\") select db1.name |
| as dbName, tab.name as tabName"}, |
| {"hive_db as db1 hive_table as tab where ((db1.createTime + 1) > 0) or (db1.name = \"Reporting\") select db1.name |
| as dbName, tab.name as tabName"}, |
| {"hive_db as db1 hive_table as tab where ((db1.createTime + 1) > 0) and (db1.name = \"Reporting\") or db1 has owner |
| select db1.name as dbName, tab.name as tabName"}, |
| {"hive_db as db1 hive_table as tab where ((db1.createTime + 1) > 0) and (db1.name = \"Reporting\") or db1 has owner |
| select db1.name as dbName, tab.name as tabName"}, |
| */ |
| // trait searches |
| {"Dimension", 5}, |
| {"JdbcAccess", 2}, |
| {"ETL", 5}, |
| {"Metric", 9}, |
| {"PII", 8}, |
| {"`Log Data`", 4}, |
| // Not sure what the expected rows should be, but since we didn't assign or do anything with the created |
| // I assume it'll be zero |
| {"`isa`", 0}, |
| |
| /* Lineage queries are fired through ClosureQuery and are tested through HiveLineageJerseyResourceIt in webapp module. |
| Commenting out the below queries since DSL to Gremlin parsing/translation fails with lineage queries when there are array types |
| used within loop expressions which is the case with DataSet.inputs and outputs.` |
| // Lineage |
| {"Table LoadProcess outputTable"}, {"Table loop (LoadProcess outputTable)"}, |
| {"Table as _loop0 loop (LoadProcess outputTable) withPath"}, |
| {"Table as src loop (LoadProcess outputTable) as dest select src.name as srcTable, dest.name as " |
| + "destTable withPath"}, |
| */ |
| // {"hive_table as t, sd, hive_column as c where t.name=\"sales_fact\" select c.name as colName, c.dataType as " |
| // + "colType", 0}, //Not working - ATLAS-145 and ATLAS-166 |
| |
| {"hive_table where name='sales_fact', db where name='Sales'", 1}, |
| {"hive_table where name='sales_fact', db where name='Reporting'", 0}, |
| {"hive_partition as p where values = ['2015-01-01']", 1}, |
| // {"StorageDesc select cols", 6} //Not working since loading of lists needs to be fixed yet |
| |
| //check supertypeNames |
| {"DataSet where name='sales_fact'", 1}, |
| {"Asset where name='sales_fact'", 1} |
| }; |
| } |
| |
| @Test(dataProvider = "dslQueriesProvider") |
| public void testBasicDSL(String query, int expected) throws AtlasBaseException { |
| AtlasSearchResult searchResult = discoveryService.searchUsingDslQuery(query, 25, 0); |
| assertNotNull(searchResult.getEntities()); |
| assertEquals(searchResult.getEntities().size(), expected); |
| } |
| |
| |
| @DataProvider(name = "dslExplicitLimitQueriesProvider") |
| private Object[][] createDSLQueriesWithExplicitLimit() { |
| return new Object[][]{ |
| {"hive_column", 37, 40, 0},//with higher limit all rows returned |
| {"hive_column limit 10", 10, 50, 0},//lower limit in query |
| {"hive_column select hive_column.name limit 10", 5, 5, 0},//lower limit in query param |
| {"hive_column select hive_column.name withPath", 20, 20, 0},//limit only in params |
| //with offset, only remaining rows returned |
| {"hive_column select hive_column.name limit 40 withPath", 17, 40, 20}, |
| //with higher offset, no rows returned |
| {"hive_column select hive_column.name limit 40 withPath", 0, 40, 40}, |
| //offset used from query |
| {"hive_column select hive_column.name limit 40 offset 10", 27, 40, 0}, |
| //offsets in query and parameter added up |
| {"hive_column select hive_column.name limit 40 offset 10", 17, 40, 10}, |
| //works with where clause |
| {"hive_db where name = 'Reporting' limit 10 offset 0", 1, 40, 0}, |
| //works with joins |
| {"hive_db, hive_table where db.name = 'Reporting' limit 10", 1, 1, 0}, |
| {"hive_column limit 25", 5, 10, 20}, //last page should return records limited by limit in query |
| {"hive_column limit 25", 0, 10, 30}, //offset > limit returns 0 rows |
| }; |
| } |
| |
| @Test(dataProvider = "dslExplicitLimitQueriesProvider") |
| public void testExplicitDSL(String query, int expected, int limit, int offset) throws AtlasBaseException { |
| AtlasSearchResult searchResult = discoveryService.searchUsingDslQuery(query, limit, offset); |
| assertNotNull(searchResult.getEntities()); |
| assertEquals(searchResult.getEntities().size(), expected); |
| } |
| |
| @DataProvider(name = "dslLimitQueriesProvider") |
| private Object[][] createDSLQueriesWithLimit() { |
| return new Object[][]{ |
| {"hive_column limit 10 ", 10}, |
| {"hive_column select hive_column.name limit 10 ", 10}, |
| {"hive_column select hive_column.name withPath", 37}, |
| {"hive_column select hive_column.name limit 10 withPath", 10}, |
| |
| {"from hive_db", 3}, |
| {"from hive_db limit 2", 2}, |
| {"from hive_db limit 2 offset 0", 2}, |
| {"from hive_db limit 2 offset 1", 2}, |
| {"from hive_db limit 3 offset 1", 2}, |
| {"hive_db", 3}, |
| {"hive_db where hive_db.name=\"Reporting\"", 1}, |
| {"hive_db where hive_db.name=\"Reporting\" or hive_db.name=\"Sales\" or hive_db.name=\"Logging\" limit 1 offset 1", 1}, |
| {"hive_db where hive_db.name=\"Reporting\" or hive_db.name=\"Sales\" or hive_db.name=\"Logging\" limit 1 offset 2", 1}, |
| {"hive_db where hive_db.name=\"Reporting\" or hive_db.name=\"Sales\" or hive_db.name=\"Logging\" limit 2 offset 1", 2}, |
| {"hive_db where hive_db.name=\"Reporting\" limit 10 ", 1}, |
| {"hive_db hive_db.name = \"Reporting\"", 1}, |
| {"hive_db where hive_db.name=\"Reporting\" select name, owner", 1}, |
| {"hive_db has name", 3}, |
| {"hive_db has name limit 2 offset 0", 2}, |
| {"hive_db has name limit 2 offset 1", 2}, |
| {"hive_db has name limit 10 offset 1", 2}, |
| {"hive_db has name limit 10 offset 0", 3}, |
| {"hive_db, hive_table", 10}, |
| {"hive_db, hive_table limit 5", 5}, |
| {"hive_db, hive_table limit 5 offset 0", 5}, |
| {"hive_db, hive_table limit 5 offset 5", 5}, |
| |
| {"View is JdbcAccess", 2}, |
| {"View is JdbcAccess limit 1", 1}, |
| {"View is JdbcAccess limit 2 offset 1", 1}, |
| {"hive_db as db1, hive_table where db1.name = \"Reporting\"", 0}, //Not working - ATLAS-145 |
| |
| |
| {"from hive_table", 10}, |
| {"from hive_table limit 5", 5}, |
| {"from hive_table limit 5 offset 5", 5}, |
| |
| {"hive_table", 10}, |
| {"hive_table limit 5", 5}, |
| {"hive_table limit 5 offset 5", 5}, |
| |
| {"hive_table isa Dimension", 3}, |
| {"hive_table isa Dimension limit 2", 2}, |
| {"hive_table isa Dimension limit 2 offset 0", 2}, |
| {"hive_table isa Dimension limit 2 offset 1", 2}, |
| {"hive_table isa Dimension limit 3 offset 1", 2}, |
| |
| {"hive_column where hive_column isa PII", 8}, |
| {"hive_column where hive_column isa PII limit 5", 5}, |
| {"hive_column where hive_column isa PII limit 5 offset 1", 5}, |
| {"hive_column where hive_column isa PII limit 5 offset 5", 3}, |
| |
| |
| {"View is Dimension" , 2}, |
| {"View is Dimension limit 1" , 1}, |
| {"View is Dimension limit 1 offset 1" , 1}, |
| {"View is Dimension limit 10 offset 1" , 1}, |
| |
| {"hive_column select hive_column.name", 37}, |
| {"hive_column select hive_column.name limit 5", 5}, |
| {"hive_column select hive_column.name limit 5 offset 36", 1}, |
| |
| {"hive_column select name", 37}, |
| {"hive_column select name limit 5", 5}, |
| {"hive_column select name limit 5 offset 36 ", 1}, |
| |
| {"hive_column where hive_column.name=\"customer_id\"", 6}, |
| {"hive_column where hive_column.name=\"customer_id\" limit 2", 2}, |
| {"hive_column where hive_column.name=\"customer_id\" limit 2 offset 1", 2}, |
| {"hive_column where hive_column.name=\"customer_id\" limit 10 offset 3", 3}, |
| |
| {"from hive_table select hive_table.name", 10}, |
| {"from hive_table select hive_table.name limit 5", 5}, |
| {"from hive_table select hive_table.name limit 5 offset 5", 5}, |
| |
| {"hive_db where (name = \"Reporting\")", 1}, |
| {"hive_db where (name = \"Reporting\") limit 10", 1}, |
| {"hive_db where (name = \"Reporting\") select name as _col_0, owner as _col_1", 1}, |
| {"hive_db where (name = \"Reporting\") select name as _col_0, owner as _col_1 limit 10", 1}, |
| {"hive_db where hive_db is JdbcAccess", 0}, //Not supposed to work |
| {"hive_db hive_table", 10}, |
| {"hive_db hive_table limit 5", 5}, |
| {"hive_db hive_table limit 5 offset 5", 5}, |
| {"hive_db where hive_db has name", 3}, |
| {"hive_db where hive_db has name limit 5", 3}, |
| {"hive_db where hive_db has name limit 2 offset 0", 2}, |
| {"hive_db where hive_db has name limit 2 offset 1", 2}, |
| |
| {"hive_db as db1 hive_table where (db1.name = \"Reporting\")", 0}, //Not working -> ATLAS-145 |
| {"hive_db where (name = \"Reporting\") select name as _col_0, (createTime + 1) as _col_1 ", 1}, |
| {"hive_db where (name = \"Reporting\") select name as _col_0, (createTime + 1) as _col_1 limit 10", 1}, |
| {"hive_db where (name = \"Reporting\") select name as _col_0, (createTime + 1) as _col_1 limit 10 offset 1", 0}, |
| {"hive_db where (name = \"Reporting\") select name as _col_0, (createTime + 1) as _col_1 limit 10 offset 0", 1}, |
| |
| {"hive_table where (name = \"sales_fact\" and createTime > \"2014-01-01\" ) select name as _col_0, createTime as _col_1 ", 1}, |
| {"hive_table where (name = \"sales_fact\" and createTime > \"2014-01-01\" ) select name as _col_0, createTime as _col_1 limit 10 ", 1}, |
| {"hive_table where (name = \"sales_fact\" and createTime > \"2014-01-01\" ) select name as _col_0, createTime as _col_1 limit 10 offset 0", 1}, |
| {"hive_table where (name = \"sales_fact\" and createTime > \"2014-01-01\" ) select name as _col_0, createTime as _col_1 limit 10 offset 5", 0}, |
| |
| {"hive_table where (name = \"sales_fact\" and createTime >= \"2014-12-11T02:35:58.440Z\" ) select name as _col_0, createTime as _col_1 ", 1}, |
| {"hive_table where (name = \"sales_fact\" and createTime >= \"2014-12-11T02:35:58.440Z\" ) select name as _col_0, createTime as _col_1 limit 10 offset 0", 1}, |
| {"hive_table where (name = \"sales_fact\" and createTime >= \"2014-12-11T02:35:58.440Z\" ) select name as _col_0, createTime as _col_1 limit 10 offset 1", 0}, |
| {"hive_table where (name = \"sales_fact\" and createTime >= \"2014-12-11T02:35:58.440Z\" ) select name as _col_0, createTime as _col_1 limit 10", 1}, |
| {"hive_table where (name = \"sales_fact\" and createTime >= \"2014-12-11T02:35:58.440Z\" ) select name as _col_0, createTime as _col_1 limit 0 offset 1", 0}, |
| |
| // trait searches |
| {"Dimension", 5}, |
| {"Dimension limit 2", 2}, |
| {"Dimension limit 2 offset 1", 2}, |
| {"Dimension limit 5 offset 4", 1}, |
| |
| {"JdbcAccess", 2}, |
| {"JdbcAccess limit 5 offset 0", 2}, |
| {"JdbcAccess limit 2 offset 1", 1}, |
| {"JdbcAccess limit 1", 1}, |
| |
| {"ETL", 5}, |
| {"ETL limit 2", 2}, |
| {"ETL limit 1", 1}, |
| {"ETL limit 1 offset 0", 1}, |
| {"ETL limit 2 offset 1", 2}, |
| |
| {"Metric", 9}, |
| {"Metric limit 10", 9}, |
| {"Metric limit 2", 2}, |
| {"Metric limit 10 offset 1", 8}, |
| |
| |
| |
| {"PII", 8}, |
| {"PII limit 10", 8}, |
| {"PII limit 2", 2}, |
| {"PII limit 10 offset 1", 7}, |
| |
| {"`Log Data`", 4}, |
| {"`Log Data` limit 3", 3}, |
| {"`Log Data` limit 10 offset 2", 2}, |
| |
| |
| {"hive_table where name='sales_fact', db where name='Sales'", 1}, |
| {"hive_table where name='sales_fact', db where name='Sales' limit 10", 1}, |
| {"hive_table where name='sales_fact', db where name='Sales' limit 10 offset 1", 0}, |
| {"hive_table where name='sales_fact', db where name='Reporting'", 0}, |
| {"hive_table where name='sales_fact', db where name='Reporting' limit 10", 0}, |
| {"hive_table where name='sales_fact', db where name='Reporting' limit 10 offset 1", 0}, |
| {"hive_partition as p where values = ['2015-01-01']", 1}, |
| {"hive_partition as p where values = ['2015-01-01'] limit 10", 1}, |
| {"hive_partition as p where values = ['2015-01-01'] limit 10 offset 1", 0}, |
| |
| }; |
| } |
| |
| @Test(dataProvider = "dslLimitQueriesProvider") |
| public void testDSLLimitQueries(String query, int expected) throws AtlasBaseException { |
| AtlasSearchResult searchResult = discoveryService.searchUsingDslQuery(query, 25, 0); |
| assertNotNull(searchResult.getEntities()); |
| assertEquals(searchResult.getEntities().size(), expected); |
| } |
| |
| |
| |
| @DataProvider(name = "dslOrderByQueriesProvider") |
| private Object[][] createDSLQueriesWithOrderBy() { |
| return new Object[][]{ |
| //test with alias |
| // {"from hive_db select hive_db.name as 'o' orderby o limit 3", 3, "name", isAscending}, |
| {"from hive_db as h orderby h.owner limit 3", 3, "owner", true}, |
| {"hive_column as c select c.name orderby hive_column.name ", 37, "c.name", true}, |
| {"hive_column as c select c.name orderby hive_column.name limit 5", 5, "c.name", true}, |
| {"hive_column as c select c.name orderby hive_column.name desc limit 5", 5, "c.name", false}, |
| |
| {"from hive_db orderby hive_db.owner limit 3", 3, "owner", true}, |
| {"hive_column select hive_column.name orderby hive_column.name ", 37, "hive_column.name", true}, |
| {"hive_column select hive_column.name orderby hive_column.name limit 5", 5, "hive_column.name", true}, |
| {"hive_column select hive_column.name orderby hive_column.name desc limit 5", 5, "hive_column.name", false}, |
| |
| {"from hive_db orderby owner limit 3", 3, "owner", true}, |
| {"hive_column select hive_column.name orderby name ", 37, "hive_column.name", true}, |
| {"hive_column select hive_column.name orderby name limit 5", 5, "hive_column.name", true}, |
| {"hive_column select hive_column.name orderby name desc limit 5", 5, "hive_column.name", false}, |
| |
| //Not working, the problem is in server code not figuring out how to sort. not sure if it is valid use case. |
| // {"hive_db hive_table orderby 'hive_db.owner'", 10, "owner", isAscending}, |
| // {"hive_db hive_table orderby 'hive_db.owner' limit 5", 5, "owner", isAscending}, |
| // {"hive_db hive_table orderby 'hive_db.owner' limit 5 offset 5", 3, "owner", isAscending}, |
| |
| {"hive_db select hive_db.description orderby hive_db.description limit 10 withPath", 3, "hive_db.description", true}, |
| {"hive_db select hive_db.description orderby hive_db.description desc limit 10 withPath", 3, "hive_db.description", false}, |
| |
| {"hive_column select hive_column.name orderby hive_column.name limit 10 withPath", 10, "hive_column.name", true}, |
| {"hive_column select hive_column.name orderby hive_column.name asc limit 10 withPath", 10, "hive_column.name", true}, |
| {"hive_column select hive_column.name orderby hive_column.name desc limit 10 withPath", 10, "hive_column.name", false}, |
| {"from hive_db orderby hive_db.owner limit 3", 3, "owner", true}, |
| {"hive_db where hive_db.name=\"Reporting\" orderby 'owner'", 1, "owner", true}, |
| |
| {"hive_db where hive_db.name=\"Reporting\" orderby hive_db.owner limit 10 ", 1, "owner", true}, |
| {"hive_db where hive_db.name=\"Reporting\" select name, owner orderby hive_db.name ", 1, "name", true}, |
| {"hive_db has name orderby hive_db.owner limit 10 offset 0", 3, "owner", true}, |
| |
| {"from hive_table select hive_table.owner orderby hive_table.owner", 10, "hive_table.owner", true}, |
| {"from hive_table select hive_table.owner orderby hive_table.owner limit 8", 8, "hive_table.owner", true}, |
| |
| {"hive_table orderby hive_table.name", 10, "name", true}, |
| |
| {"hive_table orderby hive_table.owner", 10, "owner", true}, |
| {"hive_table orderby hive_table.owner limit 8", 8, "owner", true}, |
| {"hive_table orderby hive_table.owner limit 8 offset 0", 8, "owner", true}, |
| {"hive_table orderby hive_table.owner desc limit 8 offset 0", 8, "owner", false}, |
| |
| //Not working because of existing bug Atlas-175 |
| // {"hive_table isa Dimension orderby hive_table.owner", 3, "hive_table.owner", isAscending},//order not working |
| // {"hive_table isa Dimension orderby hive_table.owner limit 3", 3, "hive_table.owner", isAscending}, |
| // {"hive_table isa Dimension orderby hive_table.owner limit 3 offset 0", 3, "hive_table.owner", isAscending}, |
| // {"hive_table isa Dimension orderby hive_table.owner desc limit 3 offset 0", 3, "hive_table.owner", !isAscending}, |
| // |
| // {"hive_column where hive_column isa PII orderby hive_column.name", 6, "hive_column.name", isAscending}, |
| // {"hive_column where hive_column isa PII orderby hive_column.name limit 5", 5, "hive_column.name", isAscending}, |
| // {"hive_column where hive_column isa PII orderby hive_column.name limit 5 offset 1", 5, "hive_column.name", isAscending}, |
| // {"hive_column where hive_column isa PII orderby hive_column.name desc limit 5 offset 1", 5, "hive_column.name", !isAscending}, |
| |
| {"hive_column select hive_column.name orderby hive_column.name ", 37, "hive_column.name", true}, |
| {"hive_column select hive_column.name orderby hive_column.name limit 5", 5, "hive_column.name", true}, |
| {"hive_column select hive_column.name orderby hive_column.name desc limit 5", 5, "hive_column.name", false}, |
| |
| {"hive_column select hive_column.name orderby hive_column.name limit 5 offset 28", 5, "hive_column.name", true}, |
| |
| {"hive_column select name orderby hive_column.name", 37, "name", true}, |
| {"hive_column select name orderby hive_column.name limit 5", 5, "name", true}, |
| {"hive_column select name orderby hive_column.name desc", 37, "name", false}, |
| |
| {"hive_column where hive_column.name=\"customer_id\" orderby hive_column.name", 6, "name", true}, |
| {"hive_column where hive_column.name=\"customer_id\" orderby hive_column.name limit 2", 2, "name", true}, |
| {"hive_column where hive_column.name=\"customer_id\" orderby hive_column.name limit 2 offset 1", 2, "name", true}, |
| |
| {"from hive_table select owner orderby hive_table.owner",10, "owner", true}, |
| {"from hive_table select owner orderby hive_table.owner limit 5", 5, "owner", true}, |
| {"from hive_table select owner orderby hive_table.owner desc limit 5", 5, "owner", false}, |
| {"from hive_table select owner orderby hive_table.owner limit 5 offset 5", 5, "owner", true}, |
| |
| {"hive_db where (name = \"Reporting\") orderby hive_db.name", 1, "name", true}, |
| {"hive_db where (name = \"Reporting\") orderby hive_db.name limit 10", 1, "name", true}, |
| {"hive_db where hive_db has name orderby hive_db.owner", 3, "owner", true}, |
| {"hive_db where hive_db has name orderby hive_db.owner limit 5", 3, "owner", true}, |
| {"hive_db where hive_db has name orderby hive_db.owner limit 2 offset 0", 2, "owner", true}, |
| {"hive_db where hive_db has name orderby hive_db.owner limit 2 offset 1", 2, "owner", true}, |
| |
| |
| {"hive_db where (name = \"Reporting\") select name as _col_0, (createTime + 1) as _col_1 orderby '_col_1'", 1, "_col_1", true}, |
| {"hive_db where (name = \"Reporting\") select name as _col_0, (createTime + 1) as _col_1 orderby '_col_1' limit 10", 1, "_col_1", true}, |
| {"hive_db where (name = \"Reporting\") select name as _col_0, (createTime + 1) as _col_1 orderby '_col_1' limit 10 offset 1", 0, "_col_1", true}, |
| {"hive_db where (name = \"Reporting\") select name as _col_0, (createTime + 1) as _col_1 orderby '_col_1' limit 10 offset 0", 1, "_col_1", true}, |
| |
| {"hive_table where (name = \"sales_fact\" and createTime > \"2014-01-01\" ) select name as _col_0, createTime as _col_1 orderby '_col_1' ", 1, "_col_1", true}, |
| {"hive_table where (name = \"sales_fact\" and createTime > \"2014-01-01\" ) select name as _col_0, createTime as _col_1 orderby '_col_1' limit 10 ", 1, "_col_1", true}, |
| {"hive_table where (name = \"sales_fact\" and createTime > \"2014-01-01\" ) select name as _col_0, createTime as _col_1 orderby '_col_1' limit 10 offset 0", 1, "_col_1", true}, |
| {"hive_table where (name = \"sales_fact\" and createTime > \"2014-01-01\" ) select name as _col_0, createTime as _col_1 orderby '_col_1' limit 10 offset 5", 0, "_col_1", true}, |
| |
| {"hive_table where (name = \"sales_fact\" and createTime >= \"2014-12-11T02:35:58.440Z\" ) select name as _col_0, createTime as _col_1 orderby '_col_0' ", 1, "_col_0", true}, |
| {"hive_table where (name = \"sales_fact\" and createTime >= \"2014-12-11T02:35:58.440Z\" ) select name as _col_0, createTime as _col_1 orderby '_col_0' limit 10 offset 0", 1, "_col_0", true}, |
| {"hive_table where (name = \"sales_fact\" and createTime >= \"2014-12-11T02:35:58.440Z\" ) select name as _col_0, createTime as _col_1 orderby '_col_0' limit 10 offset 1", 0, "_col_0", true}, |
| {"hive_table where (name = \"sales_fact\" and createTime >= \"2014-12-11T02:35:58.440Z\" ) select name as _col_0, createTime as _col_1 orderby '_col_0' limit 10", 1, "_col_0", true}, |
| {"hive_table where (name = \"sales_fact\" and createTime >= \"2014-12-11T02:35:58.440Z\" ) select name as _col_0, createTime as _col_1 orderby '_col_0' limit 0 offset 1", 0, "_col_0", true}, |
| |
| {"hive_column select hive_column.name orderby hive_column.name limit 10 withPath", 10, "hive_column.name", true}, |
| {"hive_column select hive_column.name orderby hive_column.name limit 10 withPath", 10, "hive_column.name", true}, |
| {"hive_table orderby 'hive_table.owner_notdefined'", 10, null, true}, |
| }; |
| } |
| |
| @Test(dataProvider = "dslOrderByQueriesProvider") |
| public void testOrderByDSL(String query, int expected, String orderBy, boolean ascending) throws AtlasBaseException { |
| AtlasSearchResult searchResult = discoveryService.searchUsingDslQuery(query, 25, 0); |
| assertNotNull(searchResult.getEntities()); |
| assertEquals(searchResult.getEntities().size(), expected); |
| // TODO: Implement order checking here |
| } |
| |
| @DataProvider(name = "dslLikeQueriesProvider") |
| private Object[][] createDslLikeQueries() { |
| return new Object[][]{ |
| {"hive_table where name like \"sa?es*\"", 3}, |
| {"hive_db where name like \"R*\"", 1}, |
| {"hive_db where hive_db.name like \"R???rt?*\" or hive_db.name like \"S?l?s\" or hive_db.name like\"Log*\"", 3}, |
| {"hive_db where hive_db.name like \"R???rt?*\" and hive_db.name like \"S?l?s\" and hive_db.name like\"Log*\"", 0}, |
| {"hive_table where name like 'sales*', db where name like 'Sa?es'", 1}, |
| {"hive_table where name like 'sales*' and db.name like 'Sa?es'", 1}, |
| {"hive_table where db.name like \"Sa*\"", 4}, |
| {"hive_table where db.name like \"Sa*\" and name like \"*dim\"", 3}, |
| }; |
| } |
| |
| @Test(dataProvider = "comparisonQueriesProvider") |
| public void testLikeQueries(String query, int expected) throws AtlasBaseException { |
| AtlasSearchResult searchResult = discoveryService.searchUsingDslQuery(query, 25, 0); |
| assertNotNull(searchResult.getEntities()); |
| assertEquals(searchResult.getEntities().size(), expected); |
| } |
| |
| |
| |
| // TODO: Implement FieldValidator with new Data types |
| // @DataProvider(name = "dslGroupByQueriesProvider") |
| // private Object[][] createDSLGroupByQueries() { |
| // return new Object[][]{ |
| // { "from Person as p, mentor as m groupby(m.name) select m.name, count()", |
| // new FieldValueValidator().withFieldNames("m.name", "count()").withExpectedValues("Max", 1) |
| // .withExpectedValues("Julius", 1) }, |
| // |
| // // This variant of this query is currently failing. See OMS-335 for details. |
| // { "from Person as p, mentor groupby(mentor.name) select mentor.name, count()", |
| // new FieldValueValidator().withFieldNames("mentor.name", "count()").withExpectedValues("Max", 1) |
| // .withExpectedValues("Julius", 1) }, |
| // |
| // { "from Person, mentor groupby(mentor.name) select mentor.name, count()", |
| // new FieldValueValidator().withFieldNames("mentor.name", "count()").withExpectedValues("Max", 1) |
| // .withExpectedValues("Julius", 1) }, |
| // |
| // { "from Person, mentor as m groupby(m.name) select m.name, count()", |
| // new FieldValueValidator().withFieldNames("m.name", "count()").withExpectedValues("Max", 1) |
| // .withExpectedValues("Julius", 1) }, |
| // |
| // { "from Person groupby (isOrganDonor) select count()", |
| // new FieldValueValidator().withFieldNames("count()").withExpectedValues(2) |
| // .withExpectedValues(2) }, |
| // { "from Person groupby (isOrganDonor) select Person.isOrganDonor, count()", |
| // new FieldValueValidator().withFieldNames("Person.isOrganDonor", "count()") |
| // .withExpectedValues(true, 2).withExpectedValues(false, 2) }, |
| // |
| // { "from Person groupby (isOrganDonor) select Person.isOrganDonor as 'organDonor', count() as 'count', max(Person.age) as 'max', min(Person.age) as 'min'", |
| // new FieldValueValidator().withFieldNames("organDonor", "max", "min", "count") |
| // .withExpectedValues(true, 50, 36, 2).withExpectedValues(false, 0, 0, 2) }, |
| // |
| // { "from hive_db groupby (owner, name) select count() ", new FieldValueValidator() |
| // .withFieldNames("count()").withExpectedValues(1).withExpectedValues(1).withExpectedValues(1) }, |
| // |
| // { "from hive_db groupby (owner, name) select hive_db.owner, hive_db.name, count() ", |
| // new FieldValueValidator().withFieldNames("hive_db.owner", "hive_db.name", "count()") |
| // .withExpectedValues("Jane BI", "Reporting", 1) |
| // .withExpectedValues("Tim ETL", "Logging", 1) |
| // .withExpectedValues("John ETL", "Sales", 1) }, |
| // |
| // { "from hive_db groupby (owner) select count() ", |
| // new FieldValueValidator().withFieldNames("count()").withExpectedValues(1).withExpectedValues(1) |
| // .withExpectedValues(1) }, |
| // |
| // { "from hive_db groupby (owner) select hive_db.owner, count() ", |
| // new FieldValueValidator().withFieldNames("hive_db.owner", "count()") |
| // .withExpectedValues("Jane BI", 1).withExpectedValues("Tim ETL", 1) |
| // .withExpectedValues("John ETL", 1) }, |
| // |
| // { "from hive_db groupby (owner) select hive_db.owner, max(hive_db.name) ", |
| // new FieldValueValidator().withFieldNames("hive_db.owner", "max(hive_db.name)") |
| // .withExpectedValues("Tim ETL", "Logging").withExpectedValues("Jane BI", "Reporting") |
| // .withExpectedValues("John ETL", "Sales") }, |
| // |
| // { "from hive_db groupby (owner) select max(hive_db.name) ", |
| // new FieldValueValidator().withFieldNames("max(hive_db.name)").withExpectedValues("Logging") |
| // .withExpectedValues("Reporting").withExpectedValues("Sales") }, |
| // |
| // { "from hive_db groupby (owner) select owner, hive_db.name, min(hive_db.name) ", |
| // new FieldValueValidator().withFieldNames("owner", "hive_db.name", "min(hive_db.name)") |
| // .withExpectedValues("Tim ETL", "Logging", "Logging") |
| // .withExpectedValues("Jane BI", "Reporting", "Reporting") |
| // .withExpectedValues("John ETL", "Sales", "Sales") }, |
| // |
| // { "from hive_db groupby (owner) select owner, min(hive_db.name) ", |
| // new FieldValueValidator().withFieldNames("owner", "min(hive_db.name)") |
| // .withExpectedValues("Tim ETL", "Logging").withExpectedValues("Jane BI", "Reporting") |
| // .withExpectedValues("John ETL", "Sales") }, |
| // |
| // { "from hive_db groupby (owner) select min(name) ", |
| // new FieldValueValidator().withFieldNames("min(name)") |
| // .withExpectedValues("Reporting").withExpectedValues("Logging") |
| // .withExpectedValues("Sales") }, |
| // |
| // { "from hive_db groupby (owner) select min('name') ", |
| // new FieldValueValidator().withFieldNames("min(\"name\")").withExpectedValues("name") |
| // .withExpectedValues("name").withExpectedValues("name") }, //finding the minimum of a constant literal expression... |
| // |
| // { "from hive_db groupby (owner) select name ", |
| // new FieldValueValidator().withFieldNames("name").withExpectedValues("Reporting") |
| // .withExpectedValues("Sales").withExpectedValues("Logging") }, |
| // |
| // //implied group by |
| // { "from hive_db select count() ", |
| // new FieldValueValidator().withFieldNames("count()").withExpectedValues(3) }, |
| // //implied group by |
| // { "from Person select count() as 'count', max(Person.age) as 'max', min(Person.age) as 'min'", |
| // new FieldValueValidator().withFieldNames("max", "min", "count").withExpectedValues(50, 0, 4) }, |
| // //Sum |
| // { "from Person groupby (isOrganDonor) select count() as 'count', sum(Person.age) as 'sum'", |
| // new FieldValueValidator().withFieldNames("count", "sum").withExpectedValues(2, 0) |
| // .withExpectedValues(2, 86) }, |
| // { "from Person groupby (isOrganDonor) select Person.isOrganDonor as 'organDonor', count() as 'count', sum(Person.age) as 'sum'", |
| // new FieldValueValidator().withFieldNames("organDonor", "count", "sum").withExpectedValues(false, 2, 0) |
| // .withExpectedValues(true, 2, 86) }, |
| // { "from Person select count() as 'count', sum(Person.age) as 'sum'", |
| // new FieldValueValidator().withFieldNames("count", "sum").withExpectedValues(4, 86) }, |
| // // tests to ensure that group by works with order by and limit |
| // { "from hive_db groupby (owner) select min(name) orderby name limit 2 ", |
| // new FieldValueValidator().withFieldNames("min(name)") |
| // .withExpectedValues("Logging").withExpectedValues("Reporting") |
| // }, |
| // |
| // { "from hive_db groupby (owner) select min(name) orderby name desc limit 2 ", |
| // new FieldValueValidator().withFieldNames("min(name)") |
| // .withExpectedValues("Reporting").withExpectedValues("Sales") |
| // }, |
| // }; |
| // } |
| // |
| // @DataProvider(name = "dslObjectQueriesReturnIdProvider") |
| // private Object[][] createDSLObjectIdQueries() { |
| // return new Object[][] { { |
| // "from hive_db as h select h as id", |
| // new FieldValueValidator().withFieldNames("id") |
| // .withExpectedValues(idType).withExpectedValues(idType) |
| // .withExpectedValues(idType) } |
| // }; |
| // } |
| |
| } |