blob: 3da97d3ef92698d13d0f0cdf33cda02b0e14ea6a [file] [log] [blame]
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.falcon.oozie.feed;
import org.apache.falcon.FalconException;
import org.apache.falcon.Tag;
import org.apache.falcon.entity.ClusterHelper;
import org.apache.falcon.entity.EntityUtil;
import org.apache.falcon.entity.HiveUtil;
import org.apache.falcon.entity.v0.cluster.Cluster;
import org.apache.falcon.entity.v0.feed.Feed;
import org.apache.falcon.oozie.workflow.ACTION;
import org.apache.falcon.oozie.workflow.WORKFLOWAPP;
import org.apache.falcon.oozie.workflow.CONFIGURATION;
import org.apache.falcon.util.OozieUtils;
import org.apache.falcon.workflow.WorkflowExecutionArgs;
import javax.xml.bind.JAXBElement;
import java.util.Arrays;
import java.util.Map;
import java.util.Properties;
/**
* Builds replication workflow for hcat based feed.
*/
public class HCatReplicationWorkflowBuilder extends FeedReplicationWorkflowBuilder {
private static final String EXPORT_ACTION_TEMPLATE = "/action/feed/table-export.xml";
private static final String IMPORT_ACTION_TEMPLATE = "/action/feed/table-import.xml";
private static final String CLEANUP_ACTION_TEMPLATE = "/action/feed/table-cleanup.xml";
private static final String SOURCE_HIVE_CREDENTIAL_NAME = "falconSourceHiveAuth";
private static final String TARGET_HIVE_CREDENTIAL_NAME = "falconTargetHiveAuth";
public static final String EXPORT_ACTION_NAME = "table-export";
public static final String IMPORT_ACTION_NAME = "table-import";
private static final String CLEANUP_ACTION_NAME = "cleanup-table-staging-dir";
public HCatReplicationWorkflowBuilder(Feed entity) {
super(entity);
}
@Override protected WORKFLOWAPP getWorkflow(Cluster src, Cluster target) throws FalconException {
WORKFLOWAPP workflow = new WORKFLOWAPP();
String wfName = EntityUtil.getWorkflowName(Tag.REPLICATION, Arrays.asList(src.getName()), entity).toString();
String start = EXPORT_ACTION_NAME;
//Add pre-processing
if (shouldPreProcess()) {
ACTION action = getPreProcessingAction(false, Tag.REPLICATION);
Properties hiveConf = HiveUtil.getHiveCredentials(src);
for (Map.Entry<Object, Object> e : hiveConf.entrySet()) {
CONFIGURATION.Property prop = new CONFIGURATION.Property();
prop.setName((String) e.getKey());
prop.setValue((String) e.getValue());
LOG.info("Adding config to replication hive preprocessing action : key = {} value = {}",
e.getKey(), e.getValue());
action.getJava().getConfiguration().getProperty().add(prop);
}
addHDFSServersConfig(action, src, target);
addTransition(action, EXPORT_ACTION_NAME, FAIL_POSTPROCESS_ACTION_NAME);
workflow.getDecisionOrForkOrJoin().add(action);
start = PREPROCESS_ACTION_NAME;
}
//Add export action
ACTION export = unmarshalAction(EXPORT_ACTION_TEMPLATE);
JAXBElement<org.apache.falcon.oozie.hive.ACTION> exportActionJaxbElement =
OozieUtils.unMarshalHiveAction(export);
org.apache.falcon.oozie.hive.ACTION hiveExportAction = exportActionJaxbElement.getValue();
addHDFSServersConfig(hiveExportAction, src, target);
Properties hiveConf = HiveUtil.getHiveCredentials(src);
for (Map.Entry<Object, Object> e : hiveConf.entrySet()) {
org.apache.falcon.oozie.hive.CONFIGURATION.Property prop =
new org.apache.falcon.oozie.hive.CONFIGURATION.Property();
prop.setName((String) e.getKey());
prop.setValue((String) e.getValue());
LOG.info("Adding config to replication hive export action : key = {} value = {}",
e.getKey(), e.getValue());
hiveExportAction.getConfiguration().getProperty().add(prop);
}
OozieUtils.marshalHiveAction(export, exportActionJaxbElement);
addTransition(export, REPLICATION_ACTION_NAME, FAIL_POSTPROCESS_ACTION_NAME);
workflow.getDecisionOrForkOrJoin().add(export);
//Add replication
ACTION replication = unmarshalAction(REPLICATION_ACTION_TEMPLATE);
addHDFSServersConfig(replication, src, target);
addAdditionalReplicationProperties(replication);
addTransition(replication, IMPORT_ACTION_NAME, FAIL_POSTPROCESS_ACTION_NAME);
workflow.getDecisionOrForkOrJoin().add(replication);
//Add import action
ACTION importAction = unmarshalAction(IMPORT_ACTION_TEMPLATE);
JAXBElement<org.apache.falcon.oozie.hive.ACTION> importActionJaxbElement =
OozieUtils.unMarshalHiveAction(importAction);
org.apache.falcon.oozie.hive.ACTION hiveImportAction = importActionJaxbElement.getValue();
addHDFSServersConfig(hiveImportAction, src, target);
Properties hiveConf2 = HiveUtil.getHiveCredentials(target);
for (Map.Entry<Object, Object> e : hiveConf2.entrySet()) {
org.apache.falcon.oozie.hive.CONFIGURATION.Property prop =
new org.apache.falcon.oozie.hive.CONFIGURATION.Property();
prop.setName((String) e.getKey());
prop.setValue((String) e.getValue());
LOG.info("Adding config to replication hive import action : key = {} value = {}",
e.getKey(), e.getValue());
hiveImportAction.getConfiguration().getProperty().add(prop);
}
OozieUtils.marshalHiveAction(importAction, importActionJaxbElement);
addTransition(importAction, CLEANUP_ACTION_NAME, FAIL_POSTPROCESS_ACTION_NAME);
workflow.getDecisionOrForkOrJoin().add(importAction);
//Add cleanup action
ACTION cleanup = unmarshalAction(CLEANUP_ACTION_TEMPLATE);
addTransition(cleanup, SUCCESS_POSTPROCESS_ACTION_NAME, FAIL_POSTPROCESS_ACTION_NAME);
workflow.getDecisionOrForkOrJoin().add(cleanup);
//Add post-processing actions
ACTION success = getSuccessPostProcessAction();
addHDFSServersConfig(success, src, target);
addTransition(success, OK_ACTION_NAME, FAIL_ACTION_NAME);
workflow.getDecisionOrForkOrJoin().add(success);
ACTION fail = getFailPostProcessAction();
addHDFSServersConfig(fail, src, target);
addTransition(fail, FAIL_ACTION_NAME, FAIL_ACTION_NAME);
workflow.getDecisionOrForkOrJoin().add(fail);
decorateWorkflow(workflow, wfName, start);
setupHiveCredentials(src, target, workflow);
return workflow;
}
private void setupHiveCredentials(Cluster sourceCluster, Cluster targetCluster, WORKFLOWAPP workflowApp) {
if (isSecurityEnabled) {
// add hcatalog credentials for secure mode and add a reference to each action
addHCatalogCredentials(workflowApp, sourceCluster, SOURCE_HIVE_CREDENTIAL_NAME);
addHCatalogCredentials(workflowApp, targetCluster, TARGET_HIVE_CREDENTIAL_NAME);
}
// hive-site.xml file is created later in coordinator initialization but
// actions are set to point to that here
for (Object object : workflowApp.getDecisionOrForkOrJoin()) {
if (!(object instanceof org.apache.falcon.oozie.workflow.ACTION)) {
continue;
}
org.apache.falcon.oozie.workflow.ACTION action =
(org.apache.falcon.oozie.workflow.ACTION) object;
String actionName = action.getName();
if (PREPROCESS_ACTION_NAME.equals(actionName)) {
// add reference to hive-site conf to each action
if (isSecurityEnabled) { // add a reference to credential in the action
action.setCred(SOURCE_HIVE_CREDENTIAL_NAME);
}
} else if (EXPORT_ACTION_NAME.equals(actionName)) {
if (isSecurityEnabled) { // add a reference to credential in the action
action.setCred(SOURCE_HIVE_CREDENTIAL_NAME);
}
} else if (IMPORT_ACTION_NAME.equals(actionName)) {
if (isSecurityEnabled) { // add a reference to credential in the action
action.setCred(TARGET_HIVE_CREDENTIAL_NAME);
}
}
}
}
protected Properties getWorkflowProperties(Feed feed) throws FalconException {
Properties props = super.getWorkflowProperties(feed);
props.put("availabilityFlag", "NA");
props.put(WorkflowExecutionArgs.DATASOURCE_NAME.getName(), "NA");
return props;
}
private org.apache.falcon.oozie.hive.ACTION addHDFSServersConfig(org.apache.falcon.oozie.hive.ACTION action,
Cluster sourceCluster, Cluster targetCluster) {
if (isSecurityEnabled) {
// this is to ensure that the delegation tokens are checked out for both clusters
org.apache.falcon.oozie.hive.CONFIGURATION.Property hiveProperty = new org.apache.falcon.oozie.hive
.CONFIGURATION.Property();
hiveProperty.setName("oozie.launcher.mapreduce.job.hdfs-servers");
hiveProperty.setValue(ClusterHelper.getReadOnlyStorageUrl(sourceCluster)
+ "," + ClusterHelper.getStorageUrl(targetCluster));
action.getConfiguration().getProperty().add(hiveProperty);
}
return action;
}
}