| package org.apache.helix.integration.task; |
| |
| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| */ |
| |
| import java.util.Arrays; |
| import java.util.HashMap; |
| import java.util.List; |
| import java.util.Map; |
| import java.util.Set; |
| |
| import com.google.common.collect.ImmutableMap; |
| import com.google.common.collect.Sets; |
| import org.apache.helix.TestHelper; |
| import org.apache.helix.integration.manager.ClusterControllerManager; |
| import org.apache.helix.integration.manager.MockParticipantManager; |
| import org.apache.helix.model.ExternalView; |
| import org.apache.helix.model.MasterSlaveSMD; |
| import org.apache.helix.task.JobConfig; |
| import org.apache.helix.task.JobContext; |
| import org.apache.helix.task.TaskPartitionState; |
| import org.apache.helix.task.TaskResult; |
| import org.apache.helix.task.TaskState; |
| import org.apache.helix.task.TaskSynchronizedTestBase; |
| import org.apache.helix.task.TaskUtil; |
| import org.apache.helix.task.Workflow; |
| import org.apache.helix.util.TestInputLoader; |
| import org.testng.Assert; |
| import org.testng.annotations.BeforeClass; |
| import org.testng.annotations.DataProvider; |
| import org.testng.annotations.Test; |
| |
| public final class TestJobFailure extends TaskSynchronizedTestBase { |
| private final String DB_NAME = WorkflowGenerator.DEFAULT_TGT_DB; |
| |
| @BeforeClass |
| public void beforeClass() throws Exception { |
| _participants = new MockParticipantManager[_numNodes]; |
| _numNodes = 2; |
| _numPartitions = 2; |
| _numReplicas = 1; // only Master, no Slave |
| _numDbs = 1; |
| |
| _gSetupTool.addCluster(CLUSTER_NAME, true); |
| setupParticipants(); |
| setupDBs(); |
| startParticipants(); |
| createManagers(); |
| _controller = new ClusterControllerManager(ZK_ADDR, CLUSTER_NAME, CONTROLLER_PREFIX); |
| _controller.syncStart(); |
| |
| Thread.sleep(1000L); // Wait for cluster to setup. |
| } |
| |
| private static final String EXPECTED_ENDING_STATE = "ExpectedEndingState"; |
| private static int testNum = 0; |
| |
| @Test(dataProvider = "testJobFailureInput") |
| public void testNormalJobFailure(String comment, List<String> taskStates, |
| List<String> expectedTaskEndingStates, String expectedJobEndingStates, |
| String expectedWorkflowEndingStates) throws Exception { |
| final String JOB_NAME = "test_job"; |
| final String WORKFLOW_NAME = TestHelper.getTestMethodName() + testNum++; |
| System.out.println("Test case comment: " + comment); |
| |
| Map<String, Map<String, String>> targetPartitionConfigs = |
| createPartitionConfig(taskStates, expectedTaskEndingStates); |
| |
| JobConfig.Builder firstJobBuilder = |
| new JobConfig.Builder().setWorkflow(WORKFLOW_NAME).setTargetResource(DB_NAME) |
| .setTargetPartitionStates(Sets.newHashSet(MasterSlaveSMD.States.MASTER.name())) |
| .setCommand(MockTask.TASK_COMMAND) |
| .setJobCommandConfigMap(ImmutableMap.of(MockTask.TARGET_PARTITION_CONFIG, |
| MockTask.serializeTargetPartitionConfig(targetPartitionConfigs))); |
| |
| Workflow.Builder workflowBuilder = |
| new Workflow.Builder(WORKFLOW_NAME).addJob(JOB_NAME, firstJobBuilder); |
| |
| _driver.start(workflowBuilder.build()); |
| _driver.pollForJobState(WORKFLOW_NAME, TaskUtil.getNamespacedJobName(WORKFLOW_NAME, JOB_NAME), |
| TaskState.valueOf(expectedJobEndingStates)); |
| _driver.pollForWorkflowState(WORKFLOW_NAME, TaskState.valueOf(expectedWorkflowEndingStates)); |
| |
| JobContext jobContext = |
| _driver.getJobContext(TaskUtil.getNamespacedJobName(WORKFLOW_NAME, JOB_NAME)); |
| for (int pId : jobContext.getPartitionSet()) { |
| Map<String, String> targetPartitionConfig = |
| targetPartitionConfigs.get(jobContext.getTargetForPartition(pId)); |
| Assert |
| .assertTrue(Arrays.asList(targetPartitionConfig.get(EXPECTED_ENDING_STATE).split("\\s+")) |
| .contains(jobContext.getPartitionState(pId).name())); |
| } |
| } |
| |
| @DataProvider(name = "testJobFailureInput") |
| public Object[][] loadtestJobFailureInput() { |
| String[] params = { |
| "comment", "taskStates", "expectedTaskEndingStates", "expectedJobEndingStates", |
| "expectedWorkflowEndingStates" |
| }; |
| return TestInputLoader.loadTestInputs("TestJobFailure.json", params); |
| } |
| |
| private Map<String, Map<String, String>> createPartitionConfig(List<String> taskStates, |
| List<String> expectedTaskEndingStates) throws Exception { |
| Map<String, Map<String, String>> targetPartitionConfigs = new HashMap<>(); |
| // Make sure external view has been created for the resource |
| Assert.assertTrue(TestHelper.verify(() -> { |
| ExternalView externalView = |
| _manager.getClusterManagmentTool().getResourceExternalView(CLUSTER_NAME, DB_NAME); |
| return externalView != null; |
| }, TestHelper.WAIT_DURATION)); |
| |
| ExternalView externalView = |
| _manager.getClusterManagmentTool().getResourceExternalView(CLUSTER_NAME, DB_NAME); |
| Set<String> partitionSet = externalView.getPartitionSet(); |
| if (taskStates.size() != partitionSet.size()) { |
| throw new IllegalArgumentException( |
| "Input size does not match number of partitions for target resource: " + DB_NAME); |
| } |
| int i = 0; |
| // Set job command configs for target partitions(order doesn't matter) according to specified |
| // task states. |
| for (String partition : partitionSet) { |
| Map<String, String> config = new HashMap<>(); |
| if (taskStates.get(i).equals(TaskPartitionState.COMPLETED.name())) { |
| config.put(MockTask.TASK_RESULT_STATUS, TaskResult.Status.COMPLETED.name()); |
| } else if (taskStates.get(i).equals(TaskPartitionState.TASK_ERROR.name())) { |
| config.put(MockTask.TASK_RESULT_STATUS, TaskResult.Status.FAILED.name()); |
| } else if (taskStates.get(i).equals(TaskPartitionState.TASK_ABORTED.name())) { |
| config.put(MockTask.TASK_RESULT_STATUS, TaskResult.Status.FATAL_FAILED.name()); |
| } else if (taskStates.get(i).equals(TaskPartitionState.RUNNING.name())) { |
| config.put(MockTask.JOB_DELAY, "99999999"); |
| } else { |
| throw new IllegalArgumentException("Invalid taskStates input: " + taskStates.get(i)); |
| } |
| config.put(EXPECTED_ENDING_STATE, expectedTaskEndingStates.get(i)); |
| targetPartitionConfigs.put(partition, config); |
| i++; |
| } |
| return targetPartitionConfigs; |
| } |
| } |