| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| */ |
| package org.apache.hcatalog.mapreduce; |
| |
| import org.apache.hadoop.classification.InterfaceAudience; |
| import org.apache.hadoop.classification.InterfaceStability; |
| import org.apache.hadoop.hive.metastore.MetaStoreUtils; |
| |
| import java.io.IOException; |
| import java.io.ObjectInputStream; |
| import java.io.ObjectOutputStream; |
| import java.io.Serializable; |
| import java.util.List; |
| import java.util.Properties; |
| import java.util.zip.Deflater; |
| import java.util.zip.DeflaterOutputStream; |
| import java.util.zip.InflaterInputStream; |
| |
| /** |
| * Container for metadata read from the metadata server. |
| * Prior to release 0.5, InputJobInfo was a key part of the public API, exposed directly |
| * to end-users as an argument to |
| * {@link HCatInputFormat#setInput(org.apache.hadoop.mapreduce.Job, InputJobInfo)}. |
| * Going forward, we plan on treating InputJobInfo as an implementation detail and no longer |
| * expose to end-users. Should you have a need to use InputJobInfo outside HCatalog itself, |
| * please contact the developer mailing list before depending on this class. |
| */ |
| @InterfaceAudience.Private |
| @InterfaceStability.Evolving |
| public class InputJobInfo implements Serializable { |
| |
| /** The serialization version */ |
| private static final long serialVersionUID = 1L; |
| |
| /** The db and table names. */ |
| private final String databaseName; |
| private final String tableName; |
| |
| /** meta information of the table to be read from */ |
| private HCatTableInfo tableInfo; |
| |
| /** The partition filter */ |
| private String filter; |
| |
| /** The list of partitions matching the filter. */ |
| transient private List<PartInfo> partitions; |
| |
| /** implementation specific job properties */ |
| private Properties properties; |
| |
| /** |
| * Initializes a new InputJobInfo |
| * for reading data from a table. |
| * @param databaseName the db name |
| * @param tableName the table name |
| * @param filter the partition filter |
| */ |
| public static InputJobInfo create(String databaseName, |
| String tableName, |
| String filter, |
| Properties properties) { |
| return new InputJobInfo(databaseName, tableName, filter, properties); |
| } |
| |
| private InputJobInfo(String databaseName, |
| String tableName, |
| String filter, |
| Properties properties) { |
| this.databaseName = (databaseName == null) ? |
| MetaStoreUtils.DEFAULT_DATABASE_NAME : databaseName; |
| this.tableName = tableName; |
| this.filter = filter; |
| this.properties = properties == null ? new Properties() : properties; |
| } |
| |
| /** |
| * Gets the value of databaseName |
| * @return the databaseName |
| */ |
| public String getDatabaseName() { |
| return databaseName; |
| } |
| |
| /** |
| * Gets the value of tableName |
| * @return the tableName |
| */ |
| public String getTableName() { |
| return tableName; |
| } |
| |
| /** |
| * Gets the table's meta information |
| * @return the HCatTableInfo |
| */ |
| public HCatTableInfo getTableInfo() { |
| return tableInfo; |
| } |
| |
| /** |
| * set the tablInfo instance |
| * this should be the same instance |
| * determined by this object's DatabaseName and TableName |
| * @param tableInfo |
| */ |
| void setTableInfo(HCatTableInfo tableInfo) { |
| this.tableInfo = tableInfo; |
| } |
| |
| /** |
| * Gets the value of partition filter |
| * @return the filter string |
| */ |
| public String getFilter() { |
| return filter; |
| } |
| |
| /** |
| * @return partition info |
| */ |
| public List<PartInfo> getPartitions() { |
| return partitions; |
| } |
| |
| /** |
| * @return partition info list |
| */ |
| void setPartitions(List<PartInfo> partitions) { |
| this.partitions = partitions; |
| } |
| |
| /** |
| * Set/Get Property information to be passed down to *StorageHandler implementation |
| * put implementation specific storage handler configurations here |
| * @return the implementation specific job properties |
| */ |
| public Properties getProperties() { |
| return properties; |
| } |
| |
| /** |
| * Serialize this object, compressing the partitions which can exceed the |
| * allowed jobConf size. |
| * @see <a href="https://issues.apache.org/jira/browse/HCATALOG-453">HCATALOG-453</a> |
| */ |
| private void writeObject(ObjectOutputStream oos) |
| throws IOException { |
| oos.defaultWriteObject(); |
| Deflater def = new Deflater(Deflater.BEST_COMPRESSION); |
| ObjectOutputStream partInfoWriter = |
| new ObjectOutputStream(new DeflaterOutputStream(oos, def)); |
| partInfoWriter.writeObject(partitions); |
| partInfoWriter.close(); |
| } |
| |
| /** |
| * Deserialize this object, decompressing the partitions which can exceed the |
| * allowed jobConf size. |
| * @see <a href="https://issues.apache.org/jira/browse/HCATALOG-453">HCATALOG-453</a> |
| */ |
| @SuppressWarnings("unchecked") |
| private void readObject(ObjectInputStream ois) |
| throws IOException, ClassNotFoundException { |
| ois.defaultReadObject(); |
| ObjectInputStream partInfoReader = |
| new ObjectInputStream(new InflaterInputStream(ois)); |
| partitions = (List<PartInfo>)partInfoReader.readObject(); |
| } |
| } |