blob: 73e9eaf7cf8630afd2c98014e9b1cb8be76422fe [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.pig.scripting;
import java.io.BufferedReader;
import java.io.File;
import java.io.FileInputStream;
import java.io.IOException;
import java.io.InputStreamReader;
import java.io.Reader;
import java.util.HashMap;
import java.util.Map;
import java.util.UUID;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.pig.ExecType;
import org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MRConfiguration;
import org.apache.pig.impl.PigImplConstants;
import org.apache.pig.impl.util.UDFContext;
import com.google.common.base.Charsets;
/**
* This class helps a scripting UDF capture user output by managing when to capture output and where
* the output is written to.
*
* For illustrate, we will only capture output for the last run (with the final set of data) and
* we need to keep track of the file containing that output for returning w/ the illustrate results.
* For runs, all standard output is written to the user logs.
*/
public class ScriptingOutputCapturer {
private static Log log = LogFactory.getLog(ScriptingOutputCapturer.class);
private static Map<String, String> outputFileNames = new HashMap<String, String>();
private static String runId = UUID.randomUUID().toString(); //Unique ID for this run to ensure udf output files aren't corrupted from previous runs.
//Illustrate will set the static flag telling udf to start capturing its output. It's up to each
//instance to react to it and set its own flag.
private static boolean captureOutput = false;
private boolean instancedCapturingOutput = false;
private ExecType execType;
public ScriptingOutputCapturer(ExecType execType) {
this.execType = execType;
}
public String getStandardOutputRootWriteLocation() throws IOException {
Configuration conf = UDFContext.getUDFContext().getJobConf();
String jobId = conf.get(MRConfiguration.JOB_ID);
String taskId = conf.get(MRConfiguration.TASK_ID);
String hadoopLogDir = System.getProperty("yarn.app.container.log.dir");
if (hadoopLogDir == null) {
hadoopLogDir = conf.get("yarn.app.container.log.dir");
}
if (hadoopLogDir == null) {
hadoopLogDir = System.getProperty("hadoop.log.dir");
}
if (hadoopLogDir == null) {
hadoopLogDir = conf.get("hadoop.log.dir");
}
String tmpDir = conf.get("hadoop.tmp.dir");
boolean fallbackToTmp = (hadoopLogDir == null);
if (!fallbackToTmp) {
try {
if (!(new File(hadoopLogDir).canWrite())) {
fallbackToTmp = true;
}
}
catch (SecurityException e) {
fallbackToTmp = true;
}
finally {
if (fallbackToTmp)
log.warn(String.format("Insufficient permission to write into %s. Change path to: %s", hadoopLogDir, tmpDir));
}
}
if (fallbackToTmp) {
hadoopLogDir = tmpDir;
}
log.debug("JobId: " + jobId);
log.debug("TaskId: " + taskId);
log.debug("hadoopLogDir: " + hadoopLogDir);
if (execType.isLocal() || conf.getBoolean(PigImplConstants.CONVERTED_TO_FETCH, false)) {
String logDir = System.getProperty("pig.udf.scripting.log.dir");
if (logDir == null)
logDir = ".";
return logDir + "/" + (taskId == null ? "" : (taskId + "_"));
} else {
String taskLogDir = getTaskLogDir(jobId, taskId, hadoopLogDir);
return taskLogDir + "/";
}
}
public String getTaskLogDir(String jobId, String taskId, String hadoopLogDir) throws IOException {
String taskLogDir;
String defaultUserLogDir = hadoopLogDir + File.separator + "userlogs";
if ( new File(defaultUserLogDir + File.separator + jobId).exists() ) {
taskLogDir = defaultUserLogDir + File.separator + jobId + File.separator + taskId;
} else if ( new File(defaultUserLogDir + File.separator + taskId).exists() ) {
taskLogDir = defaultUserLogDir + File.separator + taskId;
} else if ( new File(defaultUserLogDir).exists() ){
taskLogDir = defaultUserLogDir;
} else {
taskLogDir = hadoopLogDir + File.separator + "udfOutput";
File dir = new File(taskLogDir);
dir.mkdirs();
if (!dir.exists()) {
throw new IOException("Could not create directory: " + taskLogDir);
}
}
return taskLogDir;
}
public static void startCapturingOutput() {
ScriptingOutputCapturer.captureOutput = true;
}
public static Map<String, String> getUdfOutput() throws IOException {
Map<String, String> udfFuncNameToOutput = new HashMap<String,String>();
for (Map.Entry<String, String> funcToOutputFileName : outputFileNames.entrySet()) {
StringBuffer udfOutput = new StringBuffer();
FileInputStream fis = new FileInputStream(funcToOutputFileName.getValue());
Reader fr = new InputStreamReader(fis, Charsets.UTF_8);
BufferedReader br = new BufferedReader(fr);
try {
String line = br.readLine();
while (line != null) {
udfOutput.append("\t" + line + "\n");
line = br.readLine();
}
} finally {
br.close();
}
udfFuncNameToOutput.put(funcToOutputFileName.getKey(), udfOutput.toString());
}
return udfFuncNameToOutput;
}
public void registerOutputLocation(String functionName, String fileName) {
outputFileNames.put(functionName, fileName);
}
public static String getRunId() {
return runId;
}
public static boolean isClassCapturingOutput() {
return ScriptingOutputCapturer.captureOutput;
}
public boolean isInstanceCapturingOutput() {
return this.instancedCapturingOutput;
}
public void setInstanceCapturingOutput(boolean instanceCapturingOutput) {
this.instancedCapturingOutput = instanceCapturingOutput;
}
}