| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.hadoop.mapred.gridmix; |
| |
| import org.apache.commons.logging.LogFactory; |
| import org.apache.commons.logging.Log; |
| import org.apache.hadoop.conf.Configuration; |
| import org.apache.hadoop.fs.Path; |
| import org.apache.hadoop.mapreduce.test.system.MRCluster; |
| import org.apache.hadoop.mapreduce.test.system.JTProtocol; |
| import org.apache.hadoop.mapreduce.test.system.JTClient; |
| import org.apache.hadoop.mapred.JobConf; |
| import org.apache.hadoop.mapred.gridmix.test.system.GridmixJobSubmission; |
| import org.apache.hadoop.mapred.gridmix.test.system.GridmixJobVerification; |
| import org.apache.hadoop.mapred.gridmix.test.system.GridMixRunMode; |
| import org.apache.hadoop.mapred.gridmix.test.system.GridMixConfig; |
| import org.apache.hadoop.mapred.gridmix.test.system.UtilsForGridmix; |
| import org.apache.hadoop.mapred.gridmix.test.system.GridmixJobStory; |
| import org.apache.hadoop.tools.rumen.ZombieJob; |
| import org.apache.hadoop.security.UserGroupInformation; |
| import org.apache.hadoop.mapreduce.JobID; |
| import org.junit.AfterClass; |
| import org.junit.BeforeClass; |
| |
| import java.util.Iterator; |
| import java.util.Map; |
| import java.util.List; |
| import java.util.Set; |
| import java.io.IOException; |
| import org.junit.Assert; |
| |
| /** |
| * Run and verify the Gridmix jobs for given a trace. |
| */ |
| public class GridmixSystemTestCase { |
| private static final Log LOG = |
| LogFactory.getLog(GridmixSystemTestCase.class); |
| protected static Configuration conf = new Configuration(); |
| protected static MRCluster cluster; |
| protected static int cSize; |
| protected static JTClient jtClient; |
| protected static JTProtocol rtClient; |
| protected static Path gridmixDir; |
| protected static Map<String, String> map; |
| protected static GridmixJobSubmission gridmixJS; |
| protected static GridmixJobVerification gridmixJV; |
| protected static List<JobID> jobids; |
| |
| @BeforeClass |
| public static void before() throws Exception { |
| String [] excludeExpList = {"java.net.ConnectException", |
| "java.io.IOException"}; |
| cluster = MRCluster.createCluster(conf); |
| cluster.setExcludeExpList(excludeExpList); |
| cluster.setUp(); |
| cSize = cluster.getTTClients().size(); |
| jtClient = cluster.getJTClient(); |
| rtClient = jtClient.getProxy(); |
| gridmixDir = new Path("herriot-gridmix"); |
| UtilsForGridmix.createDirs(gridmixDir, rtClient.getDaemonConf()); |
| map = UtilsForGridmix.getMRTraces(rtClient.getDaemonConf()); |
| } |
| |
| @AfterClass |
| public static void after() throws Exception { |
| UtilsForGridmix.cleanup(gridmixDir, rtClient.getDaemonConf()); |
| org.apache.hadoop.fs.FileUtil.fullyDelete(new java.io.File(System. |
| getProperty("java.io.tmpdir") + "/gridmix-st/")); |
| cluster.tearDown(); |
| |
| /* Clean up the proxy user directories if gridmix run with |
| RoundRobinUserResovler mode.*/ |
| if (gridmixJV != null |
| && gridmixJV.getJobUserResolver().contains("RoundRobin")) { |
| List<String> proxyUsers = |
| UtilsForGridmix.listProxyUsers(gridmixJS.getJobConf(), |
| UserGroupInformation.getLoginUser().getShortUserName()); |
| for(int index = 0; index < proxyUsers.size(); index++){ |
| UtilsForGridmix.cleanup(new Path("hdfs:///user/" + |
| proxyUsers.get(index)), |
| rtClient.getDaemonConf()); |
| } |
| } |
| } |
| |
| /** |
| * Run the gridmix with specified runtime parameters and |
| * verify the jobs the after completion of execution. |
| * @param runtimeValues - common runtime arguments for gridmix. |
| * @param otherValues - test specific runtime arguments for gridmix. |
| * @param tracePath - path of a trace file. |
| * @throws Exception - if an exception occurs. |
| */ |
| public static void runGridmixAndVerify(String[] runtimeValues, |
| String [] otherValues, String tracePath) throws Exception { |
| runGridmixAndVerify(runtimeValues, otherValues, tracePath , |
| GridMixRunMode.DATA_GENERATION_AND_RUN_GRIDMIX.getValue()); |
| } |
| |
| /** |
| * Run the gridmix with specified runtime parameters and |
| * verify the jobs the after completion of execution. |
| * @param runtimeValues - common runtime arguments for gridmix. |
| * @param otherValues - test specific runtime arguments for gridmix. |
| * @param tracePath - path of a trace file. |
| * @param mode - 1 for data generation, 2 for run the gridmix and 3 for |
| * data generation and run the gridmix. |
| * @throws Exception - if an exception occurs. |
| */ |
| public static void runGridmixAndVerify(String [] runtimeValues, |
| String [] otherValues, String tracePath, int mode) throws Exception { |
| List<JobID> jobids = runGridmix(runtimeValues, otherValues, mode); |
| gridmixJV = new GridmixJobVerification(new Path(tracePath), |
| gridmixJS.getJobConf(), jtClient); |
| gridmixJV.verifyGridmixJobsWithJobStories(jobids); |
| } |
| |
| /** |
| * Run the gridmix with user specified mode. |
| * @param runtimeValues - common runtime parameters for gridmix. |
| * @param otherValues - test specifix runtime parameters for gridmix. |
| * @param mode - 1 for data generation, 2 for run the gridmix and 3 for |
| * data generation and run the gridmix. |
| * @return - list of gridmix job ids. |
| * @throws Exception - if an exception occurs. |
| */ |
| public static List<JobID> runGridmix(String[] runtimeValues, |
| String[] otherValues, int mode) throws Exception { |
| gridmixJS = new GridmixJobSubmission(rtClient.getDaemonConf(), |
| jtClient, gridmixDir); |
| gridmixJS.submitJobs(runtimeValues, otherValues, mode); |
| List<JobID> jobids = |
| UtilsForGridmix.listGridmixJobIDs(jtClient.getClient(), |
| gridmixJS.getGridmixJobCount()); |
| return jobids; |
| } |
| |
| /** |
| * get the trace file based on given regular expression. |
| * @param regExp - trace file file pattern. |
| * @return - trace file as string. |
| * @throws IOException - if an I/O error occurs. |
| */ |
| public static String getTraceFile(String regExp) throws IOException { |
| List<String> listTraces = UtilsForGridmix.listMRTraces( |
| rtClient.getDaemonConf()); |
| Iterator<String> ite = listTraces.iterator(); |
| while(ite.hasNext()) { |
| String traceFile = ite.next(); |
| if (traceFile.indexOf(regExp)>=0) { |
| return traceFile; |
| } |
| } |
| return null; |
| } |
| |
| /** |
| * Validate the task memory parameters. |
| * @param tracePath - trace file. |
| * @param isTraceHasHighRamJobs - true if trace has high ram job(s) |
| * otherwise its false |
| */ |
| @SuppressWarnings("deprecation") |
| public static void validateTaskMemoryParamters(String tracePath, |
| boolean isTraceHasHighRamJobs) throws IOException { |
| if (isTraceHasHighRamJobs) { |
| GridmixJobStory gjs = new GridmixJobStory(new Path(tracePath), |
| rtClient.getDaemonConf()); |
| Set<JobID> jobids = gjs.getZombieJobs().keySet(); |
| boolean isHighRamFlag = false; |
| for (JobID jobid :jobids) { |
| ZombieJob zombieJob = gjs.getZombieJobs().get(jobid); |
| JobConf origJobConf = zombieJob.getJobConf(); |
| int origMapFactor = |
| GridmixJobVerification.getMapFactor(origJobConf); |
| int origReduceFactor = |
| GridmixJobVerification.getReduceFactor(origJobConf); |
| if (origMapFactor >= 2 || origReduceFactor >= 2) { |
| isHighRamFlag = true; |
| long TaskMapMemInMB = |
| GridmixJobVerification.getScaledTaskMemInMB( |
| GridMixConfig.JOB_MAP_MEMORY_MB, |
| GridMixConfig.CLUSTER_MAP_MEMORY, |
| origJobConf, rtClient.getDaemonConf()); |
| |
| long TaskReduceMemInMB = |
| GridmixJobVerification.getScaledTaskMemInMB( |
| GridMixConfig.JOB_REDUCE_MEMORY_MB, |
| GridMixConfig.CLUSTER_REDUCE_MEMORY, |
| origJobConf, rtClient.getDaemonConf()); |
| long taskMapLimitInMB = |
| conf.getLong(GridMixConfig.CLUSTER_MAX_MAP_MEMORY, |
| JobConf.DISABLED_MEMORY_LIMIT); |
| |
| long taskReduceLimitInMB = |
| conf.getLong(GridMixConfig.CLUSTER_MAX_REDUCE_MEMORY, |
| JobConf.DISABLED_MEMORY_LIMIT); |
| |
| GridmixJobVerification.verifyMemoryLimits(TaskMapMemInMB, |
| taskMapLimitInMB); |
| GridmixJobVerification.verifyMemoryLimits(TaskReduceMemInMB, |
| taskReduceLimitInMB); |
| } |
| } |
| Assert.assertTrue("Trace doesn't have atleast one high ram job.", |
| isHighRamFlag); |
| } |
| } |
| |
| public static boolean isLocalDistCache(String fileName, String userName, |
| boolean visibility) { |
| return DistributedCacheEmulator.isLocalDistCacheFile(fileName, |
| userName, visibility); |
| } |
| } |
| |