| package org.apache.helix.webapp.resources; |
| |
| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| */ |
| |
| import java.util.Map; |
| |
| import org.apache.helix.HelixDataAccessor; |
| import org.apache.helix.HelixException; |
| import org.apache.helix.PropertyKey; |
| import org.apache.helix.zookeeper.datamodel.ZNRecord; |
| import org.apache.helix.zookeeper.impl.client.ZkClient; |
| import org.apache.helix.model.ResourceConfig; |
| import org.apache.helix.task.JobConfig; |
| import org.apache.helix.task.TaskDriver; |
| import org.apache.helix.task.TaskUtil; |
| import org.apache.helix.task.Workflow; |
| import org.apache.helix.task.WorkflowContext; |
| import org.restlet.data.Form; |
| import org.restlet.data.MediaType; |
| import org.restlet.data.Status; |
| import org.restlet.representation.Representation; |
| import org.restlet.representation.StringRepresentation; |
| import org.restlet.representation.Variant; |
| import org.restlet.resource.ServerResource; |
| import org.slf4j.Logger; |
| import org.slf4j.LoggerFactory; |
| |
| /** |
| * Class for server-side resource at <code>"/clusters/{clusterName}/jobQueues/{jobQueue}" |
| * <p> |
| * <li>GET list job queue info |
| * <li>POST start a new job in a job queue, or stop/resume/persistDataChanges/delete a job queue |
| */ |
| public class JobQueueResource extends ServerResource { |
| private final static Logger LOG = LoggerFactory.getLogger(JobQueueResource.class); |
| |
| public JobQueueResource() { |
| getVariants().add(new Variant(MediaType.TEXT_PLAIN)); |
| getVariants().add(new Variant(MediaType.APPLICATION_JSON)); |
| setNegotiated(false); |
| } |
| |
| /** |
| * List job queue info |
| * <p> |
| * Usage: <code>curl http://{host:port}/clusters/{clusterName}/jobQueues/{jobQueue} |
| */ |
| @Override |
| public Representation get() { |
| StringRepresentation presentation; |
| try { |
| String clusterName = |
| ResourceUtil.getAttributeFromRequest(getRequest(), ResourceUtil.RequestKey.CLUSTER_NAME); |
| String jobQueueName = |
| ResourceUtil.getAttributeFromRequest(getRequest(), ResourceUtil.RequestKey.JOB_QUEUE); |
| presentation = getHostedEntitiesRepresentation(clusterName, jobQueueName); |
| } catch (Exception e) { |
| String error = ClusterRepresentationUtil.getErrorAsJsonStringFromException(e); |
| presentation = new StringRepresentation(error, MediaType.APPLICATION_JSON); |
| LOG.error("Fail to get job queue", e); |
| } |
| return presentation; |
| } |
| |
| StringRepresentation getHostedEntitiesRepresentation(String clusterName, String jobQueueName) |
| throws Exception { |
| ZkClient zkClient = |
| ResourceUtil.getAttributeFromCtx(getContext(), ResourceUtil.ContextKey.ZKCLIENT); |
| HelixDataAccessor accessor = |
| ClusterRepresentationUtil.getClusterDataAccessor(zkClient, clusterName); |
| PropertyKey.Builder keyBuilder = accessor.keyBuilder(); |
| |
| TaskDriver taskDriver = new TaskDriver(zkClient, clusterName); |
| |
| // Get job queue config |
| // TODO: fix this to use workflowConfig. |
| ResourceConfig jobQueueConfig = accessor.getProperty(keyBuilder.resourceConfig(jobQueueName)); |
| |
| // Get job queue context |
| WorkflowContext ctx = taskDriver.getWorkflowContext(jobQueueName); |
| |
| // Create the result |
| ZNRecord hostedEntitiesRecord = new ZNRecord(jobQueueName); |
| if (jobQueueConfig != null) { |
| hostedEntitiesRecord.merge(jobQueueConfig.getRecord()); |
| } |
| if (ctx != null) { |
| hostedEntitiesRecord.merge(ctx.getRecord()); |
| } |
| |
| StringRepresentation representation = |
| new StringRepresentation(ClusterRepresentationUtil.ZNRecordToJson(hostedEntitiesRecord), |
| MediaType.APPLICATION_JSON); |
| |
| return representation; |
| } |
| |
| /** |
| * Start a new job in a job queue, or stop/resume/persistDataChanges/delete a job queue |
| * <p> |
| * Usage: |
| * <p> |
| * <li>Start a new job in a job queue: |
| * <code>curl -d @'./{input.txt}' -H 'Content-Type: application/json' |
| * http://{host:port}/clusters/{clusterName}/jobQueues/{jobQueue} |
| * <p> |
| * input.txt: <code>jsonParameters={"command":"start"}&newJob={newJobConfig.yaml} |
| * <p> |
| * For newJobConfig.yaml, see {@link Workflow#parse(String)} |
| * <li>Stop/resume/persistDataChanges/delete a job queue: |
| * <code>curl -d 'jsonParameters={"command":"{stop/resume/persistDataChanges/delete}"}' |
| * -H "Content-Type: application/json" http://{host:port}/clusters/{clusterName}/jobQueues/{jobQueue} |
| */ |
| @Override |
| public Representation post(Representation entity) { |
| String clusterName = |
| ResourceUtil.getAttributeFromRequest(getRequest(), ResourceUtil.RequestKey.CLUSTER_NAME); |
| String jobQueueName = |
| ResourceUtil.getAttributeFromRequest(getRequest(), ResourceUtil.RequestKey.JOB_QUEUE); |
| ZkClient zkClient = |
| ResourceUtil.getAttributeFromCtx(getContext(), ResourceUtil.ContextKey.ZKCLIENT); |
| try { |
| TaskDriver driver = new TaskDriver(zkClient, clusterName); |
| |
| Form form = new Form(entity); |
| JsonParameters jsonParameters = new JsonParameters(form); |
| |
| TaskDriver.DriverCommand cmd = TaskDriver.DriverCommand.valueOf(jsonParameters.getCommand()); |
| switch (cmd) { |
| case start: { |
| // Get the job queue and submit it |
| String yamlPayload = |
| ResourceUtil.getYamlParameters(form, ResourceUtil.YamlParamKey.NEW_JOB); |
| if (yamlPayload == null) { |
| throw new HelixException("Yaml job config is required!"); |
| } |
| Workflow workflow = Workflow.parse(yamlPayload); |
| |
| for (String jobName : workflow.getJobConfigs().keySet()) { |
| Map<String, String> jobCfgMap = workflow.getJobConfigs().get(jobName); |
| JobConfig.Builder jobCfgBuilder = JobConfig.Builder.fromMap(jobCfgMap); |
| if (workflow.getTaskConfigs() != null && workflow.getTaskConfigs().containsKey(jobName)) { |
| jobCfgBuilder.addTaskConfigs(workflow.getTaskConfigs().get(jobName)); |
| } |
| driver.enqueueJob(jobQueueName, TaskUtil.getDenamespacedJobName(jobQueueName, jobName), |
| jobCfgBuilder); |
| } |
| break; |
| } |
| case stop: { |
| driver.stop(jobQueueName); |
| break; |
| } |
| case resume: { |
| driver.resume(jobQueueName); |
| break; |
| } |
| case flush: { |
| driver.flushQueue(jobQueueName); |
| break; |
| } |
| case delete: { |
| driver.delete(jobQueueName); |
| break; |
| } |
| case clean: { |
| driver.cleanupQueue(jobQueueName); |
| break; |
| } |
| default: |
| throw new HelixException("Unsupported job queue command: " + cmd); |
| } |
| getResponse().setEntity(getHostedEntitiesRepresentation(clusterName, jobQueueName)); |
| getResponse().setStatus(Status.SUCCESS_OK); |
| } catch (Exception e) { |
| getResponse().setEntity(ClusterRepresentationUtil.getErrorAsJsonStringFromException(e), |
| MediaType.APPLICATION_JSON); |
| getResponse().setStatus(Status.SUCCESS_OK); |
| LOG.error("Error in posting job queue: " + entity, e); |
| } |
| return null; |
| } |
| } |