blob: 2a474a03f01f347e5abce31aefc9452a445a711d [file] [log] [blame]
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.mapred;
import java.io.File;
import java.io.FileOutputStream;
import java.io.IOException;
import java.io.PrintWriter;
import java.util.List;
import java.util.ArrayList;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.permission.FsAction;
import org.apache.hadoop.fs.permission.FsPermission;
import org.apache.hadoop.hdfs.MiniDFSCluster;
import org.apache.hadoop.mapreduce.filecache.TestTrackerDistributedCacheManager;
import org.apache.hadoop.mapreduce.MRConfig;
import org.apache.hadoop.mapreduce.server.jobtracker.JTConfig;
import org.apache.hadoop.mapreduce.server.tasktracker.TTConfig;
import org.apache.hadoop.security.UserGroupInformation;
import org.apache.hadoop.util.StringUtils;
import junit.framework.TestCase;
/**
* The base class which starts up a cluster with LinuxTaskController as the task
* controller.
*
* In order to run test cases utilizing LinuxTaskController please follow the
* following steps:
* <ol>
* <li>Build LinuxTaskController by not passing any
* <code>-Dhadoop.conf.dir</code></li>
* <li>Change ownership of the built binary to root:group1, where group1 is
* a secondary group of the test runner.</li>
* <li>Change permissions on the binary so that <em>others</em> component does
* not have any permissions on binary</li>
* <li>Make the built binary to setuid and setgid executable</li>
* <li>Execute following targets:
* <code>ant test -Dcompile.c++=true -Dtaskcontroller-path=<em>path to built binary</em>
* -Dtaskcontroller-ugi=<em>user,group</em></code>
* <br/>(Note that "path to built binary" means the directory containing task-controller -
* not the actual complete path of the binary itself. This path must end in ".../bin")
* </li>
* </ol>
*
*/
public class ClusterWithLinuxTaskController extends TestCase {
private static final Log LOG =
LogFactory.getLog(ClusterWithLinuxTaskController.class);
/**
* The wrapper class around LinuxTaskController which allows modification of
* the custom path to task-controller which we can use for task management.
*
**/
public static class MyLinuxTaskController extends LinuxTaskController {
String taskControllerExePath = System.getProperty(TASKCONTROLLER_PATH)
+ "/task-controller";
@Override
public void setup() throws IOException {
getConf().set(TTConfig.TT_GROUP, taskTrackerSpecialGroup);
// write configuration file
configurationFile = createTaskControllerConf(System
.getProperty(TASKCONTROLLER_PATH), getConf());
super.setup();
}
@Override
protected String getTaskControllerExecutablePath() {
return taskControllerExePath;
}
void setTaskControllerExe(String execPath) {
this.taskControllerExePath = execPath;
}
volatile static int attemptedSigQuits = 0;
volatile static int failedSigQuits = 0;
/** Work like LinuxTaskController, but also count the number of
* attempted and failed SIGQUIT sends via the task-controller
* executable.
*/
@Override
void dumpTaskStack(TaskControllerContext context) {
attemptedSigQuits++;
try {
signalTask(context, TaskControllerCommands.SIGQUIT_TASK_JVM);
} catch (Exception e) {
LOG.warn("Execution sending SIGQUIT: " + StringUtils.stringifyException(e));
failedSigQuits++;
}
}
}
// cluster instances which sub classes can use
protected MiniMRCluster mrCluster = null;
protected MiniDFSCluster dfsCluster = null;
private JobConf clusterConf = null;
protected Path homeDirectory;
/** changing this to a larger number needs more work for creating
* taskcontroller.cfg.
* see {@link #startCluster()} and
* {@link #createTaskControllerConf(String, Configuration)}
*/
private static final int NUMBER_OF_NODES = 1;
static final String TASKCONTROLLER_PATH = "taskcontroller-path";
static final String TASKCONTROLLER_UGI = "taskcontroller-ugi";
private static File configurationFile = null;
protected UserGroupInformation jobOwner;
protected static String taskTrackerSpecialGroup = null;
/**
* Primary group of the tasktracker - i.e. the user running the
* test.
*/
protected static String taskTrackerPrimaryGroup = null;
static {
if (isTaskExecPathPassed()) {
try {
taskTrackerSpecialGroup = FileSystem.getLocal(new Configuration())
.getFileStatus(
new Path(System.getProperty(TASKCONTROLLER_PATH),
"task-controller")).getGroup();
} catch (IOException e) {
LOG.warn("Could not get group of the binary", e);
fail("Could not get group of the binary");
}
try {
taskTrackerPrimaryGroup =
UserGroupInformation.getCurrentUser().getGroupNames()[0];
} catch (IOException ioe) {
LOG.warn("Could not get primary group of the current user", ioe);
fail("Could not get primary group of the current user");
}
}
}
/*
* Utility method which subclasses use to start and configure the MR Cluster
* so they can directly submit a job.
*/
protected void startCluster()
throws IOException, InterruptedException {
JobConf conf = new JobConf();
dfsCluster = new MiniDFSCluster(conf, NUMBER_OF_NODES, true, null);
conf.set(TTConfig.TT_TASK_CONTROLLER,
MyLinuxTaskController.class.getName());
conf.setBoolean(JTConfig.JT_PERSIST_JOBSTATUS, false);
mrCluster =
new MiniMRCluster(NUMBER_OF_NODES, dfsCluster.getFileSystem().getUri()
.toString(), 4, null, null, conf);
clusterConf = mrCluster.createJobConf();
String ugi = System.getProperty(TASKCONTROLLER_UGI);
String[] splits = ugi.split(",");
jobOwner = UserGroupInformation.createUserForTesting(splits[0],
new String[]{splits[1]});
createHomeAndStagingDirectory(clusterConf);
}
private void createHomeAndStagingDirectory(JobConf conf)
throws IOException {
FileSystem fs = dfsCluster.getFileSystem();
String path = "/user/" + jobOwner.getUserName();
homeDirectory = new Path(path);
LOG.info("Creating Home directory : " + homeDirectory);
fs.mkdirs(homeDirectory);
changePermission(fs);
Path stagingArea = new Path(conf.get(JTConfig.JT_STAGING_AREA_ROOT));
LOG.info("Creating Staging root directory : " + stagingArea);
fs.mkdirs(stagingArea);
fs.setPermission(stagingArea, new FsPermission((short)0777));
}
private void changePermission(FileSystem fs)
throws IOException {
fs.setOwner(homeDirectory, jobOwner.getUserName(),
jobOwner.getGroupNames()[0]);
}
static File getTaskControllerConfFile(String path) {
File confDirectory = new File(path, "../conf");
return new File(confDirectory, "taskcontroller.cfg");
}
/**
* Create taskcontroller.cfg.
*
* @param path Path to the taskcontroller binary.
* @param conf TaskTracker's configuration
* @return the created conf file
* @throws IOException
*/
static File createTaskControllerConf(String path,
Configuration conf) throws IOException {
File confDirectory = new File(path, "../conf");
if (!confDirectory.exists()) {
confDirectory.mkdirs();
}
File configurationFile = new File(confDirectory, "taskcontroller.cfg");
PrintWriter writer =
new PrintWriter(new FileOutputStream(configurationFile));
writer.println(String.format(MRConfig.LOCAL_DIR + "=%s", conf
.get(MRConfig.LOCAL_DIR)));
writer
.println(String.format("hadoop.log.dir=%s", TaskLog.getBaseLogDir()));
writer.println(String.format(TTConfig.TT_GROUP + "=%s",
conf.get(TTConfig.TT_GROUP)));
writer.flush();
writer.close();
return configurationFile;
}
/**
* Can we run the tests with LinuxTaskController?
*
* @return boolean
*/
protected static boolean shouldRun() {
if (!isTaskExecPathPassed() || !isUserPassed()) {
LOG.info("Not running test.");
return false;
}
return true;
}
static boolean isTaskExecPathPassed() {
String path = System.getProperty(TASKCONTROLLER_PATH);
if (path == null || path.isEmpty()
|| path.equals("${" + TASKCONTROLLER_PATH + "}")) {
LOG.info("Invalid taskcontroller-path : " + path);
return false;
}
return true;
}
private static boolean isUserPassed() {
String ugi = System.getProperty(TASKCONTROLLER_UGI);
if (ugi != null && !(ugi.equals("${" + TASKCONTROLLER_UGI + "}"))
&& !ugi.isEmpty()) {
if (ugi.indexOf(",") > 1) {
return true;
}
LOG.info("Invalid taskcontroller-ugi : " + ugi);
return false;
}
LOG.info("Invalid taskcontroller-ugi : " + ugi);
return false;
}
protected JobConf getClusterConf() {
return new JobConf(clusterConf);
}
@Override
protected void tearDown()
throws Exception {
if (mrCluster != null) {
mrCluster.shutdown();
}
if (dfsCluster != null) {
dfsCluster.shutdown();
}
if (configurationFile != null) {
configurationFile.delete();
}
super.tearDown();
}
/**
* Assert that the job is actually run by the specified user by verifying the
* permissions of the output part-files.
*
* @param outDir
* @throws IOException
*/
protected void assertOwnerShip(Path outDir)
throws IOException {
FileSystem fs = outDir.getFileSystem(clusterConf);
assertOwnerShip(outDir, fs);
}
/**
* Assert that the job is actually run by the specified user by verifying the
* permissions of the output part-files.
*
* @param outDir
* @param fs
* @throws IOException
*/
protected void assertOwnerShip(Path outDir, FileSystem fs)
throws IOException {
for (FileStatus status : fs.listStatus(outDir,
new Utils.OutputFileUtils
.OutputFilesFilter())) {
String owner = status.getOwner();
String group = status.getGroup();
LOG.info("Ownership of the file is " + status.getPath() + " is " + owner
+ "," + group);
assertTrue("Output part-file's owner is not correct. Expected : "
+ jobOwner.getUserName() + " Found : " + owner, owner
.equals(jobOwner.getUserName()));
assertTrue("Output part-file's group is not correct. Expected : "
+ jobOwner.getGroupNames()[0] + " Found : " + group, group
.equals(jobOwner.getGroupNames()[0]));
}
}
/**
* Validates permissions of private distcache dir and its contents fully
*/
public static void checkPermissionsOnPrivateDistCache(String[] localDirs,
String user, String taskTrackerUser, String groupOwner)
throws IOException {
// user-dir, jobcache and distcache will have
// 2770 permissions if jobOwner is same as tt_user
// 2570 permissions for any other user
String expectedDirPerms = taskTrackerUser.equals(user)
? "drwxrws---"
: "dr-xrws---";
String expectedFilePerms = taskTrackerUser.equals(user)
? "-rwxrwx---"
: "-r-xrwx---";
for (String localDir : localDirs) {
File distCacheDir = new File(localDir,
TaskTracker.getPrivateDistributedCacheDir(user));
if (distCacheDir.exists()) {
checkPermissionsOnDir(distCacheDir, user, groupOwner, expectedDirPerms,
expectedFilePerms);
}
}
}
/**
* Check that files expected to be localized in distributed cache for a user
* are present.
* @param localDirs List of mapred local directories.
* @param user User against which localization is happening
* @param expectedFileNames List of files expected to be localized
* @throws IOException
*/
public static void checkPresenceOfPrivateDistCacheFiles(String[] localDirs,
String user, String[] expectedFileNames) throws IOException {
FileGatherer gatherer = new FileGatherer();
for (String localDir : localDirs) {
File distCacheDir = new File(localDir,
TaskTracker.getPrivateDistributedCacheDir(user));
findExpectedFiles(expectedFileNames, distCacheDir, gatherer);
}
assertEquals("Files expected in private distributed cache were not found",
expectedFileNames.length, gatherer.getCount());
}
/**
* Validates permissions and ownership of public distcache dir and its
* contents fully in all local dirs
*/
public static void checkPermissionsOnPublicDistCache(FileSystem localFS,
String[] localDirs, String owner, String group) throws IOException {
for (String localDir : localDirs) {
File distCacheDir = new File(localDir,
TaskTracker.getPublicDistributedCacheDir());
if (distCacheDir.exists()) {
checkPublicFilePermissions(localFS, distCacheDir, owner, group);
}
}
}
/**
* Checks that files expected to be localized in the public distributed
* cache are present
* @param localDirs List of mapred local directories
* @param expectedFileNames List of expected file names.
* @throws IOException
*/
public static void checkPresenceOfPublicDistCacheFiles(String[] localDirs,
String[] expectedFileNames) throws IOException {
FileGatherer gatherer = new FileGatherer();
for (String localDir : localDirs) {
File distCacheDir = new File(localDir,
TaskTracker.getPublicDistributedCacheDir());
findExpectedFiles(expectedFileNames, distCacheDir, gatherer);
}
assertEquals("Files expected in public distributed cache were not found",
expectedFileNames.length, gatherer.getCount());
}
/**
* Validates permissions and ownership on the public distributed cache files
*/
private static void checkPublicFilePermissions(FileSystem localFS, File dir,
String owner, String group)
throws IOException {
Path dirPath = new Path(dir.getAbsolutePath());
TestTrackerDistributedCacheManager.checkPublicFilePermissions(localFS,
new Path[] {dirPath});
TestTrackerDistributedCacheManager.checkPublicFileOwnership(localFS,
new Path[] {dirPath}, owner, group);
if (dir.isDirectory()) {
File[] files = dir.listFiles();
for (File file : files) {
checkPublicFilePermissions(localFS, file, owner, group);
}
}
}
/**
* Validates permissions of given dir and its contents fully(i.e. recursively)
*/
private static void checkPermissionsOnDir(File dir, String user,
String groupOwner, String expectedDirPermissions,
String expectedFilePermissions) throws IOException {
TestTaskTrackerLocalization.checkFilePermissions(dir.toString(),
expectedDirPermissions, user, groupOwner);
File[] files = dir.listFiles();
for (File file : files) {
if (file.isDirectory()) {
checkPermissionsOnDir(file, user, groupOwner, expectedDirPermissions,
expectedFilePermissions);
} else {
TestTaskTrackerLocalization.checkFilePermissions(file.toString(),
expectedFilePermissions, user, groupOwner);
}
}
}
// Check which files among those expected are present in the rootDir
// Add those present to the FileGatherer.
private static void findExpectedFiles(String[] expectedFileNames,
File rootDir, FileGatherer gatherer) {
File[] files = rootDir.listFiles();
if (files == null) {
return;
}
for (File file : files) {
if (file.isDirectory()) {
findExpectedFiles(expectedFileNames, file, gatherer);
} else {
if (isFilePresent(expectedFileNames, file)) {
gatherer.addFileName(file.getName());
}
}
}
}
// Test if the passed file is present in the expected list of files.
private static boolean isFilePresent(String[] expectedFileNames, File file) {
boolean foundFileName = false;
for (String name : expectedFileNames) {
if (name.equals(file.getName())) {
foundFileName = true;
break;
}
}
return foundFileName;
}
// Helper class to collect a list of file names across multiple
// method calls. Wrapper around a collection defined for clarity
private static class FileGatherer {
List<String> foundFileNames = new ArrayList<String>();
void addFileName(String fileName) {
foundFileNames.add(fileName);
}
int getCount() {
return foundFileNames.size();
}
}
}