blob: a5291d93db00223f8bb3df1b1e91e76330a86a08 [file] [log] [blame]
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.oozie.action.hadoop;
import org.apache.pig.Main;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import java.io.FileNotFoundException;
import java.io.OutputStream;
import java.io.FileOutputStream;
import java.io.BufferedReader;
import java.io.FileReader;
import java.io.File;
import java.io.IOException;
import java.util.HashSet;
import java.util.Map;
import java.util.List;
import java.util.ArrayList;
import java.util.Properties;
import java.util.Set;
import java.net.URL;
public class PigMainWithOldAPI extends LauncherMain {
private static final Set<String> DISALLOWED_PIG_OPTIONS = new HashSet<String>();
static {
DISALLOWED_PIG_OPTIONS.add("-4");
DISALLOWED_PIG_OPTIONS.add("-log4jconf");
DISALLOWED_PIG_OPTIONS.add("-e");
DISALLOWED_PIG_OPTIONS.add("-execute");
DISALLOWED_PIG_OPTIONS.add("-f");
DISALLOWED_PIG_OPTIONS.add("-file");
DISALLOWED_PIG_OPTIONS.add("-l");
DISALLOWED_PIG_OPTIONS.add("-logfile");
DISALLOWED_PIG_OPTIONS.add("-r");
DISALLOWED_PIG_OPTIONS.add("-dryrun");
DISALLOWED_PIG_OPTIONS.add("-x");
DISALLOWED_PIG_OPTIONS.add("-exectype");
DISALLOWED_PIG_OPTIONS.add("-P");
DISALLOWED_PIG_OPTIONS.add("-propertyFile");
}
public static void main(String[] args) throws Exception {
run(PigMainWithOldAPI.class, args);
}
protected void run(String[] args) throws Exception {
System.out.println();
System.out.println("Oozie Pig action configuration");
System.out.println("=================================================================");
// loading action conf prepared by Oozie
Configuration actionConf = new Configuration(false);
String actionXml = System.getProperty("oozie.action.conf.xml");
if (actionXml == null) {
throw new RuntimeException("Missing Java System Property [oozie.action.conf.xml]");
}
if (!new File(actionXml).exists()) {
throw new RuntimeException("Action Configuration XML file [" + actionXml + "] does not exist");
}
actionConf.addResource(new Path("file:///", actionXml));
Properties pigProperties = new Properties();
for (Map.Entry<String, String> entry : actionConf) {
pigProperties.setProperty(entry.getKey(), entry.getValue());
}
//propagate delegation related props from launcher job to Pig job
String jobTokenFile = getFilePathFromEnv("HADOOP_TOKEN_FILE_LOCATION");
if (jobTokenFile != null) {
pigProperties.setProperty("mapreduce.job.credentials.binary", jobTokenFile);
System.out.println("------------------------");
System.out.println("Setting env property for mapreduce.job.credentials.binary to:"
+ jobTokenFile);
System.out.println("------------------------");
System.setProperty("mapreduce.job.credentials.binary", jobTokenFile);
}
else {
System.out.println("Non-kerberoes execution");
}
OutputStream os = new FileOutputStream("pig.properties");
pigProperties.store(os, "");
os.close();
System.out.println();
System.out.println("pig.properties content:");
System.out.println("------------------------");
pigProperties.store(System.out, "");
System.out.flush();
System.out.println("------------------------");
System.out.println();
List<String> arguments = new ArrayList<String>();
String script = actionConf.get("oozie.pig.script");
if (script == null) {
throw new RuntimeException("Action Configuration does not have [oozie.pig.script] property");
}
if (!new File(script).exists()) {
throw new RuntimeException("Error: Pig script file [" + script + "] does not exist");
}
System.out.println("Pig script [" + script + "] content: ");
System.out.println("------------------------");
BufferedReader br = new BufferedReader(new FileReader(script));
String line = br.readLine();
while (line != null) {
System.out.println(line);
line = br.readLine();
}
br.close();
System.out.println("------------------------");
System.out.println();
arguments.add("-file");
arguments.add(script);
String[] params = MapReduceMain.getStrings(actionConf, "oozie.pig.params");
for (String param : params) {
arguments.add("-param");
arguments.add(param);
}
String hadoopJobId = System.getProperty("oozie.launcher.job.id");
if (hadoopJobId == null) {
throw new RuntimeException("Launcher Hadoop Job ID system property not set");
}
String logFile = new File("pig-oozie-" + hadoopJobId + ".log").getAbsolutePath();
URL log4jFile = Thread.currentThread().getContextClassLoader().getResource("log4j.properties");
if (log4jFile != null) {
String pigLogLevel = actionConf.get("oozie.pig.log.level", "INFO");
String rootLogLevel = actionConf.get("oozie.action." + LauncherMapper.ROOT_LOGGER_LEVEL, "INFO");
// append required PIG properties to the default hadoop log4j file
Properties hadoopProps = new Properties();
hadoopProps.load(log4jFile.openStream());
hadoopProps.setProperty("log4j.rootLogger", rootLogLevel + ", A, B");
hadoopProps.setProperty("log4j.logger.org.apache.pig", pigLogLevel + ", A, B");
hadoopProps.setProperty("log4j.appender.A", "org.apache.log4j.ConsoleAppender");
hadoopProps.setProperty("log4j.appender.A.layout", "org.apache.log4j.PatternLayout");
hadoopProps.setProperty("log4j.appender.A.layout.ConversionPattern", "%d [%t] %-5p %c %x - %m%n");
hadoopProps.setProperty("log4j.appender.B", "org.apache.log4j.FileAppender");
hadoopProps.setProperty("log4j.appender.B.file", logFile);
hadoopProps.setProperty("log4j.appender.B.layout", "org.apache.log4j.PatternLayout");
hadoopProps.setProperty("log4j.appender.B.layout.ConversionPattern", "%d [%t] %-5p %c %x - %m%n");
String localProps = new File("piglog4j.properties").getAbsolutePath();
OutputStream os1 = new FileOutputStream(localProps);
hadoopProps.store(os1, "");
os1.close();
arguments.add("-log4jconf");
arguments.add(localProps);
// print out current directory
File localDir = new File(localProps).getParentFile();
System.out.println("Current (local) dir = " + localDir.getAbsolutePath());
}
else {
System.out.println("log4jfile is null");
}
String pigLog = "pig-" + hadoopJobId + ".log";
arguments.add("-logfile");
arguments.add(pigLog);
String[] pigArgs = MapReduceMain.getStrings(actionConf, "oozie.pig.args");
for (String pigArg : pigArgs) {
if (DISALLOWED_PIG_OPTIONS.contains(pigArg)) {
throw new RuntimeException("Error: Pig argument " + pigArg + " is not supported");
}
arguments.add(pigArg);
}
System.out.println("Pig command arguments :");
for (String arg : arguments) {
System.out.println(" " + arg);
}
System.out.println("=================================================================");
System.out.println();
System.out.println(">>> Invoking Pig command line now >>>");
System.out.println();
System.out.flush();
try {
System.out.println();
runPigJob(new String[] { "-version" });
}
catch (SecurityException ex) {
LauncherSecurityManager.reset();
}
System.out.println();
System.out.flush();
try {
runPigJob(arguments.toArray(new String[arguments.size()]));
}
catch (SecurityException ex) {
if (LauncherSecurityManager.getExitInvoked()) {
if (LauncherSecurityManager.getExitCode() != 0) {
System.err.println();
System.err.println("Pig logfile dump:");
System.err.println();
try {
BufferedReader reader = new BufferedReader(new FileReader(pigLog));
line = reader.readLine();
while (line != null) {
System.err.println(line);
line = reader.readLine();
}
reader.close();
}
catch (FileNotFoundException e) {
System.err.println("pig log file: " + pigLog + " not found.");
}
throw ex;
}
}
}
System.out.println();
System.out.println("<<< Invocation of Pig command completed <<<");
System.out.println();
// harvesting and recording Hadoop Job IDs
Properties jobIds = getHadoopJobIds(logFile);
File file = new File(System.getProperty(LauncherMapper.ACTION_PREFIX + LauncherMapper.ACTION_DATA_OUTPUT_PROPS));
os = new FileOutputStream(file);
jobIds.store(os, "");
os.close();
System.out.println(" Hadoop Job IDs executed by Pig: " + jobIds.getProperty(HADOOP_JOBS));
System.out.println();
}
protected void runPigJob(String[] args) throws Exception {
// running as from the command line
Main.main(args);
}
private static final String JOB_ID_LOG_PREFIX = "HadoopJobId: ";
protected Properties getHadoopJobIds(String logFile) throws IOException {
int jobCount = 0;
Properties props = new Properties();
StringBuffer sb = new StringBuffer(100);
if (new File(logFile).exists() == false) {
System.err.println("pig log file: " + logFile + " not present. Therefore no Hadoop jobids found");
props.setProperty(HADOOP_JOBS, "");
}
else {
BufferedReader br = new BufferedReader(new FileReader(logFile));
String line = br.readLine();
String separator = "";
while (line != null) {
if (line.contains(JOB_ID_LOG_PREFIX)) {
int jobIdStarts = line.indexOf(JOB_ID_LOG_PREFIX) + JOB_ID_LOG_PREFIX.length();
String jobId = line.substring(jobIdStarts);
int jobIdEnds = jobId.indexOf(" ");
if (jobIdEnds > -1) {
jobId = jobId.substring(0, jobId.indexOf(" "));
}
sb.append(separator).append(jobId);
separator = ",";
}
line = br.readLine();
}
br.close();
props.setProperty(HADOOP_JOBS, sb.toString());
}
return props;
}
}