| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package org.apache.falcon.resource; |
| |
| import org.apache.commons.lang.StringUtils; |
| import org.apache.falcon.FalconException; |
| import org.apache.falcon.FalconWebException; |
| import org.apache.falcon.Pair; |
| import org.apache.falcon.entity.EntityUtil; |
| import org.apache.falcon.entity.v0.Entity; |
| import org.apache.falcon.entity.v0.EntityType; |
| import org.apache.falcon.entity.v0.SchemaHelper; |
| import org.apache.falcon.entity.v0.UnschedulableEntityException; |
| import org.apache.falcon.entity.v0.cluster.Cluster; |
| import org.apache.falcon.monitors.Dimension; |
| import org.apache.hadoop.security.authorize.AuthorizationException; |
| import org.slf4j.Logger; |
| import org.slf4j.LoggerFactory; |
| |
| import javax.servlet.http.HttpServletRequest; |
| import javax.ws.rs.PathParam; |
| import javax.ws.rs.core.Context; |
| import javax.ws.rs.core.Response; |
| import java.util.*; |
| |
| /** |
| * REST resource of allowed actions on Schedulable Entities, Only Process and |
| * Feed can have schedulable actions. |
| */ |
| public abstract class AbstractSchedulableEntityManager extends AbstractInstanceManager { |
| |
| private static final Logger LOG = LoggerFactory.getLogger(AbstractSchedulableEntityManager.class); |
| |
| /** |
| * Schedules an submitted entity immediately. |
| * |
| * @param type entity type |
| * @param entity entity name |
| * @return APIResult |
| */ |
| public APIResult schedule( |
| @Context HttpServletRequest request, @Dimension("entityType") @PathParam("type") String type, |
| @Dimension("entityName") @PathParam("entity") String entity, |
| @Dimension("colo") @PathParam("colo") String colo) { |
| checkColo(colo); |
| try { |
| audit(request, entity, type, "SCHEDULED"); |
| scheduleInternal(type, entity); |
| return new APIResult(APIResult.Status.SUCCEEDED, entity + "(" + type + ") scheduled successfully"); |
| } catch (Throwable e) { |
| LOG.error("Unable to schedule workflow", e); |
| throw FalconWebException.newException(e, Response.Status.BAD_REQUEST); |
| } |
| } |
| |
| private synchronized void scheduleInternal(String type, String entity) |
| throws FalconException, AuthorizationException { |
| |
| checkSchedulableEntity(type); |
| Entity entityObj = EntityUtil.getEntity(type, entity); |
| getWorkflowEngine().schedule(entityObj); |
| } |
| |
| /** |
| * Submits a new entity and schedules it immediately. |
| * |
| * @param type entity type |
| * @return APIResult |
| */ |
| public APIResult submitAndSchedule( |
| @Context HttpServletRequest request, @Dimension("entityType") @PathParam("type") String type, |
| @Dimension("colo") @PathParam("colo") String colo) { |
| checkColo(colo); |
| try { |
| checkSchedulableEntity(type); |
| audit(request, "STREAMED_DATA", type, "SUBMIT_AND_SCHEDULE"); |
| Entity entity = submitInternal(request, type); |
| scheduleInternal(type, entity.getName()); |
| return new APIResult(APIResult.Status.SUCCEEDED, |
| entity.getName() + "(" + type + ") scheduled successfully"); |
| } catch (Throwable e) { |
| LOG.error("Unable to submit and schedule ", e); |
| throw FalconWebException.newException(e, Response.Status.BAD_REQUEST); |
| } |
| } |
| |
| /** |
| * Suspends a running entity. |
| * |
| * @param type entity type |
| * @param entity entity name |
| * @return APIResult |
| */ |
| public APIResult suspend( |
| @Context HttpServletRequest request, @Dimension("entityType") @PathParam("type") String type, |
| @Dimension("entityName") @PathParam("entity") String entity, |
| @Dimension("entityName") @PathParam("entity") String colo) { |
| checkColo(colo); |
| try { |
| checkSchedulableEntity(type); |
| audit(request, entity, type, "SUSPEND"); |
| Entity entityObj = EntityUtil.getEntity(type, entity); |
| if (getWorkflowEngine().isActive(entityObj)) { |
| getWorkflowEngine().suspend(entityObj); |
| } else { |
| throw new FalconException(entity + "(" + type + ") is not scheduled"); |
| } |
| return new APIResult(APIResult.Status.SUCCEEDED, entity + "(" + type + ") suspended successfully"); |
| } catch (Throwable e) { |
| LOG.error("Unable to suspend entity", e); |
| throw FalconWebException.newException(e, Response.Status.BAD_REQUEST); |
| } |
| } |
| |
| /** |
| * Resumes a suspended entity. |
| * |
| * @param type entity type |
| * @param entity entity name |
| * @return APIResult |
| */ |
| public APIResult resume( |
| @Context HttpServletRequest request, @Dimension("entityType") @PathParam("type") String type, |
| @Dimension("entityName") @PathParam("entity") String entity, |
| @Dimension("colo") @PathParam("colo") String colo) { |
| |
| checkColo(colo); |
| try { |
| checkSchedulableEntity(type); |
| audit(request, entity, type, "RESUME"); |
| Entity entityObj = EntityUtil.getEntity(type, entity); |
| if (getWorkflowEngine().isActive(entityObj)) { |
| getWorkflowEngine().resume(entityObj); |
| } else { |
| throw new FalconException(entity + "(" + type + ") is not scheduled"); |
| } |
| return new APIResult(APIResult.Status.SUCCEEDED, entity + "(" + type + ") resumed successfully"); |
| } catch (Throwable e) { |
| LOG.error("Unable to resume entity", e); |
| throw FalconWebException.newException(e, Response.Status.BAD_REQUEST); |
| } |
| } |
| |
| //SUSPEND CHECKSTYLE CHECK ParameterNumberCheck |
| /** |
| * Returns summary of most recent N instances of an entity, filtered by cluster. |
| * |
| * @param type Only return entities of this type. |
| * @param startDate For each entity, show instances after startDate. |
| * @param endDate For each entity, show instances before endDate. |
| * @param cluster Return entities for specific cluster. |
| * @param fields fields that the query is interested in, separated by comma |
| * @param filterBy filter by a specific field. |
| * @param filterTags filter by these tags. |
| * @param orderBy order result by these fields. |
| * @param offset Pagination offset. |
| * @param resultsPerPage Number of results that should be returned starting at the offset. |
| * @param numInstances Number of instance summaries to show per entity |
| * @return EntitySummaryResult |
| */ |
| public EntitySummaryResult getEntitySummary(String type, String cluster, String startDate, String endDate, |
| String fields, String filterBy, String filterTags, |
| String orderBy, String sortOrder, Integer offset, |
| Integer resultsPerPage, Integer numInstances) { |
| HashSet<String> fieldSet = new HashSet<String>(Arrays.asList(fields.toLowerCase().split(","))); |
| Pair<Date, Date> startAndEndDates = getStartEndDatesForSummary(startDate, endDate); |
| |
| List<Entity> entities; |
| String colo; |
| try { |
| entities = getEntities(type, |
| SchemaHelper.getDateFormat().format(startAndEndDates.first), |
| SchemaHelper.getDateFormat().format(startAndEndDates.second), |
| cluster, filterBy, filterTags, orderBy, sortOrder, offset, resultsPerPage); |
| colo = ((Cluster) configStore.get(EntityType.CLUSTER, cluster)).getColo(); |
| } catch (Exception e) { |
| LOG.error("Failed to get entities", e); |
| throw FalconWebException.newException(e, Response.Status.BAD_REQUEST); |
| } |
| |
| List<EntitySummaryResult.EntitySummary> entitySummaries = new ArrayList<EntitySummaryResult.EntitySummary>(); |
| for (Entity entity : entities) { |
| InstancesResult instancesResult = getInstances(entity.getEntityType().name(), entity.getName(), |
| SchemaHelper.getDateFormat().format(startAndEndDates.first), |
| SchemaHelper.getDateFormat().format(startAndEndDates.second), |
| colo, null, "", "", "", 0, numInstances); |
| |
| /* ToDo - Use oozie bulk API after FALCON-591 is implemented |
| * getBulkInstances(entity, cluster, |
| * startAndEndDates.first, startAndEndDates.second, colo, "starttime", 0, numInstances); |
| */ |
| List<EntitySummaryResult.Instance> entitySummaryInstances = |
| getElementsFromInstanceResult(instancesResult); |
| |
| List<String> pipelines = new ArrayList<String>(); |
| List<String> tags = new ArrayList<String>(); |
| if (fieldSet.contains("pipelines")) { pipelines = EntityUtil.getPipelines(entity); } |
| if (fieldSet.contains("tags")) { tags = EntityUtil.getTags(entity); } |
| |
| EntitySummaryResult.EntitySummary entitySummary = |
| new EntitySummaryResult.EntitySummary(entity.getName(), entity.getEntityType().toString(), |
| getStatusString(entity), |
| tags.toArray(new String[tags.size()]), |
| pipelines.toArray(new String[pipelines.size()]), |
| entitySummaryInstances.toArray( |
| new EntitySummaryResult.Instance[entitySummaryInstances.size()])); |
| entitySummaries.add(entitySummary); |
| } |
| return new EntitySummaryResult("Entity Summary Result", |
| entitySummaries.toArray(new EntitySummaryResult.EntitySummary[entitySummaries.size()])); |
| } |
| //RESUME CHECKSTYLE CHECK ParameterNumberCheck |
| |
| private Pair<Date, Date> getStartEndDatesForSummary(String startDate, String endDate) { |
| Date end = (StringUtils.isEmpty(endDate)) ? new Date() : SchemaHelper.parseDateUTC(endDate); |
| |
| long startMillisecs = end.getTime() - (2* DAY_IN_MILLIS); // default - 2 days before end |
| Date start = (StringUtils.isEmpty(startDate)) |
| ? new Date(startMillisecs) : SchemaHelper.parseDateUTC(startDate); |
| |
| return new Pair<Date, Date>(start, end); |
| } |
| |
| private List<EntitySummaryResult.Instance> getElementsFromInstanceResult(InstancesResult instancesResult) { |
| ArrayList<EntitySummaryResult.Instance> elemInstanceList = |
| new ArrayList<EntitySummaryResult.Instance>(); |
| InstancesResult.Instance[] instances = instancesResult.getInstances(); |
| if (instances != null && instances.length > 0) { |
| for (InstancesResult.Instance rawInstance : instances) { |
| EntitySummaryResult.Instance instance = new EntitySummaryResult.Instance(rawInstance.getCluster(), |
| rawInstance.getInstance(), |
| EntitySummaryResult.WorkflowStatus.valueOf(rawInstance.getStatus().toString())); |
| instance.logFile = rawInstance.getLogFile(); |
| instance.sourceCluster = rawInstance.sourceCluster; |
| instance.startTime = rawInstance.startTime; |
| instance.endTime = rawInstance.endTime; |
| elemInstanceList.add(instance); |
| } |
| } |
| |
| return elemInstanceList; |
| } |
| |
| private void checkSchedulableEntity(String type) throws UnschedulableEntityException { |
| EntityType entityType = EntityType.valueOf(type.toUpperCase()); |
| if (!entityType.isSchedulable()) { |
| throw new UnschedulableEntityException( |
| "Entity type (" + type + ") " + " cannot be Scheduled/Suspended/Resumed"); |
| } |
| } |
| } |