| // @@@ START COPYRIGHT @@@ |
| // |
| // Licensed to the Apache Software Foundation (ASF) under one |
| // or more contributor license agreements. See the NOTICE file |
| // distributed with this work for additional information |
| // regarding copyright ownership. The ASF licenses this file |
| // to you under the Apache License, Version 2.0 (the |
| // "License"); you may not use this file except in compliance |
| // with the License. You may obtain a copy of the License at |
| // |
| // http://www.apache.org/licenses/LICENSE-2.0 |
| // |
| // Unless required by applicable law or agreed to in writing, |
| // software distributed under the License is distributed on an |
| // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| // KIND, either express or implied. See the License for the |
| // specific language governing permissions and limitations |
| // under the License. |
| // |
| // @@@ END COPYRIGHT @@@ |
| |
| package org.trafodion.sql; |
| |
| import java.io.IOException; |
| import java.util.HashMap; |
| import java.util.ArrayList; |
| import java.util.List; |
| import java.lang.reflect.Field; |
| |
| import org.apache.log4j.PropertyConfigurator; |
| import org.apache.log4j.Logger; |
| import org.apache.thrift.TException; |
| |
| import org.apache.hadoop.util.StringUtils; |
| |
| import org.apache.hadoop.hive.metastore.HiveMetaStoreClient; |
| import org.apache.hadoop.hive.metastore.api.Table; |
| import org.apache.hadoop.hive.metastore.api.Database; |
| import org.apache.hadoop.hive.metastore.api.MetaException; |
| import org.apache.hadoop.hive.metastore.api.NoSuchObjectException; |
| import org.apache.hadoop.hive.metastore.api.UnknownDBException; |
| // These are needed for the DDL_TIME constant. This class is different in Hive 0.10. |
| // We use Java reflection instead of importing the class statically. |
| // For Hive 0.9 or lower |
| // import org.apache.hadoop.hive.metastore.api.Constants; |
| // For Hive 0.10 or higher |
| // import org.apache.hadoop.hive.metastore.api.hive_metastoreConstants; |
| import org.apache.hadoop.hive.conf.HiveConf; |
| import org.apache.hadoop.fs.Path; |
| import org.apache.hadoop.fs.FileSystem; |
| |
| import java.sql.SQLException; |
| import java.sql.Connection; |
| import java.sql.ResultSet; |
| import java.sql.Statement; |
| import java.sql.DriverManager; |
| |
| |
| public class HiveClient { |
| private static Logger logger = Logger.getLogger(HiveClient.class.getName()); |
| private final String lockPath="/trafodion/traflock"; |
| |
| private static HiveConf hiveConf = null; |
| private static HiveMetaStoreClient hmsClient ; |
| private static String ddlTimeConst = null; |
| |
| private static Statement stmt = null; |
| |
| static { |
| String confFile = System.getProperty("trafodion.log4j.configFile"); |
| System.setProperty("trafodion.root", System.getenv("TRAF_HOME")); |
| if (confFile == null) |
| confFile = System.getenv("TRAF_CONF") + "/log4j.sql.config"; |
| PropertyConfigurator.configure(confFile); |
| hiveConf = new HiveConf(); |
| try { |
| hmsClient = new HiveMetaStoreClient(hiveConf, null); |
| ddlTimeConst = getDDLTimeConstant(); |
| } catch (MetaException me) |
| { |
| throw new RuntimeException("Checked MetaException from HiveClient static block"); |
| } |
| } |
| |
| public static boolean close() |
| { |
| hmsClient.close(); |
| return true; |
| } |
| |
| public static boolean exists(String schName, String tblName) |
| throws MetaException, TException, UnknownDBException |
| { |
| if (logger.isDebugEnabled()) logger.debug("HiveClient.exists(" + schName + " , " + tblName + ") called."); |
| boolean result = hmsClient.tableExists(schName, tblName); |
| return result; |
| } |
| |
| public static String getHiveTableString(String schName, String tblName) |
| throws MetaException, TException |
| { |
| Table table; |
| |
| if (logger.isDebugEnabled()) logger.debug("HiveClient.getHiveTableString(" + schName + " , " + |
| tblName + ") called."); |
| try { |
| table = hmsClient.getTable(schName, tblName); |
| } |
| catch (NoSuchObjectException x) { |
| if (logger.isDebugEnabled()) logger.debug("HiveTable not found"); |
| return new String(""); |
| } |
| if (logger.isDebugEnabled()) logger.debug("HiveTable is " + table.toString()); |
| return table.toString(); |
| } |
| |
| public static String getHiveTableParameters(String schName, String tblName) |
| throws MetaException, TException |
| { |
| Table table; |
| if (logger.isDebugEnabled()) logger.debug("HiveClient.getHiveTableParameters(" + schName + " , " + |
| tblName + ") called."); |
| try { |
| table = hmsClient.getTable(schName, tblName); |
| } |
| catch (NoSuchObjectException x) { |
| if (logger.isDebugEnabled()) logger.debug("HiveTable not found"); |
| return new String(""); |
| } |
| String tableParams = new String(); |
| return tableParams.toString(); |
| } |
| |
| public static long getRedefTime(String schName, String tblName) |
| throws MetaException, TException, ClassCastException, NullPointerException, NumberFormatException |
| { |
| Table table; |
| if (logger.isDebugEnabled()) logger.debug("HiveClient.getRedefTime(" + schName + " , " + |
| tblName + ") called."); |
| try { |
| table = hmsClient.getTable(schName, tblName); |
| if (logger.isDebugEnabled()) logger.debug("getTable returns null for " + schName + "." + tblName + "."); |
| if (table == null) |
| return 0; |
| } |
| catch (NoSuchObjectException x) { |
| if (logger.isDebugEnabled()) logger.debug("Hive table no longer exists."); |
| return 0; |
| } |
| |
| long redefTime = table.getCreateTime(); |
| if (table.getParameters() != null){ |
| // those would be used without reflection |
| //String rfTime = table.getParameters().get(Constants.DDL_TIME); |
| //String rfTime = table.getParameters().get(hive_metastoreConstants.DDL_TIME); |
| // determing the constant using reflection instead |
| String rfTime = table.getParameters().get(ddlTimeConst); |
| if (rfTime != null) |
| redefTime = Long.parseLong(rfTime); |
| } |
| if (logger.isDebugEnabled()) logger.debug("RedefTime is " + redefTime); |
| return redefTime ; |
| } |
| |
| public static Object[] getAllSchemas() throws MetaException |
| { |
| List<String> schemaList = (hmsClient.getAllDatabases()); |
| if (schemaList != null) |
| return schemaList.toArray(); |
| else |
| return null; |
| } |
| |
| public static Object[] getAllTables(String schName) |
| throws MetaException, TException |
| { |
| try { |
| Database db = hmsClient.getDatabase(schName); |
| if (db == null) |
| return null; |
| |
| List<String> tableList = hmsClient.getAllTables(schName); |
| if (tableList != null) |
| return tableList.toArray(); |
| else |
| return null; |
| } catch (NoSuchObjectException e) { |
| return null; |
| } |
| } |
| |
| // Because Hive changed the name of the class containing internal constants changed |
| // in Hive 0.10, we are using Java Reflection to get the value of the DDL_TIME constant. |
| public static String getDDLTimeConstant() |
| throws MetaException |
| { |
| |
| Class constsClass = null; |
| Object constsFromReflection = null; |
| Field ddlTimeField = null; |
| Object fieldVal = null; |
| |
| // Using the class loader, try to load either class by name. |
| // Note that both classes have a default constructor and both have a static |
| // String field DDL_TIME, so the rest of the code is the same for both. |
| try { |
| try { |
| constsClass = Class.forName( |
| // Name in Hive 0.10 and higher |
| "org.apache.hadoop.hive.metastore.api.hive_metastoreConstants"); |
| } catch (ClassNotFoundException e) { |
| // probably not found because we are using Hive 0.10 or later |
| constsClass = null; |
| } |
| if (constsClass == null) { |
| constsClass = Class.forName( |
| // Name in Hive 0.9 and lower |
| "org.apache.hadoop.hive.metastore.api.Constants"); |
| } |
| |
| // Make a new object for this class, using the default constructor |
| constsFromReflection = constsClass.newInstance(); |
| } catch (InstantiationException e) { |
| throw new MetaException("Instantiation error for metastore constants class"); |
| } catch (IllegalAccessException e) { |
| throw new MetaException("Illegal access exception"); |
| } catch (ClassNotFoundException e) { |
| throw new MetaException("Could not find Hive Metastore constants class"); |
| } |
| // Using Java reflection, get a reference to the DDL_TIME field |
| try { |
| ddlTimeField = constsClass.getField("DDL_TIME"); |
| } catch (NoSuchFieldException e) { |
| throw new MetaException("Could not find DDL_TIME constant field"); |
| } |
| // get the String object that represents the value of this field |
| try { |
| fieldVal = ddlTimeField.get(constsFromReflection); |
| |
| } catch (IllegalAccessException e) { |
| throw new MetaException("Could not get value for DDL_TIME constant field"); |
| } |
| |
| return fieldVal.toString(); |
| } |
| |
| public static void executeHiveSQL(String ddl) |
| throws ClassNotFoundException, SQLException |
| { |
| if (stmt == null) { |
| Class.forName("org.apache.hive.jdbc.HiveDriver"); |
| Connection con = null; |
| String isSecureHadoop = System.getenv("SECURE_HADOOP"); |
| //If Kerberos is enabled, then we need to connect to remote hiveserver2 using hive principal |
| if(isSecureHadoop != null && isSecureHadoop.equalsIgnoreCase("Y")){ |
| String hiveServer2Url = System.getenv("HIVESERVER2_URL"); |
| if(hiveServer2Url == null || hiveServer2Url.isEmpty()){ |
| hiveServer2Url = "localhost:10000"; |
| } |
| String hivePrincipal = System.getenv("HIVE_PRINCIPAL"); |
| con = DriverManager.getConnection("jdbc:hive2://" + hiveServer2Url+"/;principal=" + hivePrincipal, "hive", ""); |
| }else{ |
| con = DriverManager.getConnection("jdbc:hive2://", "hive", ""); |
| } |
| stmt = con.createStatement(); |
| } |
| |
| try { |
| |
| stmt.execute(ddl); |
| } catch (SQLException s) { |
| throw s; |
| } |
| } |
| |
| } |