| /* |
| * Licensed to the Apache Software Foundation (ASF) under one or more |
| * contributor license agreements. See the NOTICE file distributed with |
| * this work for additional information regarding copyright ownership. |
| * The ASF licenses this file to You under the Apache License, Version 2.0 |
| * (the "License"); you may not use this file except in compliance with |
| * the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.nifi.controller; |
| |
| import org.apache.commons.collections4.CollectionUtils; |
| import org.apache.nifi.authorization.AbstractPolicyBasedAuthorizer; |
| import org.apache.nifi.authorization.Authorizer; |
| import org.apache.nifi.cluster.protocol.DataFlow; |
| import org.apache.nifi.cluster.protocol.StandardDataFlow; |
| import org.apache.nifi.connectable.Connectable; |
| import org.apache.nifi.connectable.ConnectableType; |
| import org.apache.nifi.connectable.Connection; |
| import org.apache.nifi.connectable.Funnel; |
| import org.apache.nifi.connectable.Port; |
| import org.apache.nifi.connectable.Position; |
| import org.apache.nifi.connectable.Size; |
| import org.apache.nifi.controller.exception.ProcessorInstantiationException; |
| import org.apache.nifi.controller.label.Label; |
| import org.apache.nifi.controller.reporting.ReportingTaskInstantiationException; |
| import org.apache.nifi.controller.reporting.StandardReportingInitializationContext; |
| import org.apache.nifi.controller.serialization.FlowEncodingVersion; |
| import org.apache.nifi.controller.serialization.FlowFromDOMFactory; |
| import org.apache.nifi.controller.serialization.FlowSerializationException; |
| import org.apache.nifi.controller.serialization.FlowSynchronizationException; |
| import org.apache.nifi.controller.serialization.FlowSynchronizer; |
| import org.apache.nifi.controller.serialization.StandardFlowSerializer; |
| import org.apache.nifi.controller.service.ControllerServiceLoader; |
| import org.apache.nifi.controller.service.ControllerServiceNode; |
| import org.apache.nifi.controller.service.ControllerServiceState; |
| import org.apache.nifi.encrypt.StringEncryptor; |
| import org.apache.nifi.events.BulletinFactory; |
| import org.apache.nifi.fingerprint.FingerprintException; |
| import org.apache.nifi.fingerprint.FingerprintFactory; |
| import org.apache.nifi.flowfile.FlowFilePrioritizer; |
| import org.apache.nifi.groups.ProcessGroup; |
| import org.apache.nifi.groups.RemoteProcessGroup; |
| import org.apache.nifi.groups.RemoteProcessGroupPortDescriptor; |
| import org.apache.nifi.logging.ComponentLog; |
| import org.apache.nifi.logging.LogLevel; |
| import org.apache.nifi.processor.Relationship; |
| import org.apache.nifi.processor.SimpleProcessLogger; |
| import org.apache.nifi.remote.RemoteGroupPort; |
| import org.apache.nifi.remote.RootGroupPort; |
| import org.apache.nifi.remote.protocol.SiteToSiteTransportProtocol; |
| import org.apache.nifi.reporting.InitializationException; |
| import org.apache.nifi.reporting.ReportingInitializationContext; |
| import org.apache.nifi.reporting.Severity; |
| import org.apache.nifi.scheduling.SchedulingStrategy; |
| import org.apache.nifi.util.DomUtils; |
| import org.apache.nifi.util.NiFiProperties; |
| import org.apache.nifi.util.file.FileUtils; |
| import org.apache.nifi.web.api.dto.ConnectableDTO; |
| import org.apache.nifi.web.api.dto.ConnectionDTO; |
| import org.apache.nifi.web.api.dto.ControllerServiceDTO; |
| import org.apache.nifi.web.api.dto.FlowSnippetDTO; |
| import org.apache.nifi.web.api.dto.FunnelDTO; |
| import org.apache.nifi.web.api.dto.LabelDTO; |
| import org.apache.nifi.web.api.dto.PortDTO; |
| import org.apache.nifi.web.api.dto.PositionDTO; |
| import org.apache.nifi.web.api.dto.ProcessGroupDTO; |
| import org.apache.nifi.web.api.dto.ProcessorConfigDTO; |
| import org.apache.nifi.web.api.dto.ProcessorDTO; |
| import org.apache.nifi.web.api.dto.RemoteProcessGroupDTO; |
| import org.apache.nifi.web.api.dto.ReportingTaskDTO; |
| import org.apache.nifi.web.api.dto.TemplateDTO; |
| import org.slf4j.Logger; |
| import org.slf4j.LoggerFactory; |
| import org.w3c.dom.Document; |
| import org.w3c.dom.Element; |
| import org.w3c.dom.Node; |
| import org.w3c.dom.NodeList; |
| import org.xml.sax.SAXException; |
| |
| import javax.xml.XMLConstants; |
| import javax.xml.parsers.DocumentBuilder; |
| import javax.xml.parsers.DocumentBuilderFactory; |
| import javax.xml.parsers.ParserConfigurationException; |
| import javax.xml.validation.Schema; |
| import javax.xml.validation.SchemaFactory; |
| import java.io.ByteArrayInputStream; |
| import java.io.ByteArrayOutputStream; |
| import java.io.IOException; |
| import java.io.InputStream; |
| import java.net.URL; |
| import java.nio.charset.StandardCharsets; |
| import java.nio.file.Files; |
| import java.nio.file.Path; |
| import java.nio.file.StandardOpenOption; |
| import java.util.ArrayList; |
| import java.util.Arrays; |
| import java.util.Collections; |
| import java.util.HashSet; |
| import java.util.List; |
| import java.util.Map; |
| import java.util.Set; |
| import java.util.concurrent.TimeUnit; |
| import java.util.zip.GZIPInputStream; |
| |
| /** |
| */ |
| public class StandardFlowSynchronizer implements FlowSynchronizer { |
| |
| private static final Logger logger = LoggerFactory.getLogger(StandardFlowSynchronizer.class); |
| public static final URL FLOW_XSD_RESOURCE = StandardFlowSynchronizer.class.getResource("/FlowConfiguration.xsd"); |
| private final StringEncryptor encryptor; |
| private final boolean autoResumeState; |
| |
| public StandardFlowSynchronizer(final StringEncryptor encryptor) { |
| this.encryptor = encryptor; |
| autoResumeState = NiFiProperties.getInstance().getAutoResumeState(); |
| } |
| |
| public static boolean isEmpty(final DataFlow dataFlow) { |
| if (dataFlow == null || dataFlow.getFlow() == null || dataFlow.getFlow().length == 0) { |
| return true; |
| } |
| |
| final Document document = parseFlowBytes(dataFlow.getFlow()); |
| final Element rootElement = document.getDocumentElement(); |
| |
| final Element rootGroupElement = (Element) rootElement.getElementsByTagName("rootGroup").item(0); |
| final FlowEncodingVersion encodingVersion = FlowEncodingVersion.parse(rootGroupElement); |
| |
| final ProcessGroupDTO rootGroupDto = FlowFromDOMFactory.getProcessGroup(null, rootGroupElement, null, encodingVersion); |
| return isEmpty(rootGroupDto); |
| } |
| |
| @Override |
| public void sync(final FlowController controller, final DataFlow proposedFlow, final StringEncryptor encryptor) |
| throws FlowSerializationException, UninheritableFlowException, FlowSynchronizationException { |
| |
| // handle corner cases involving no proposed flow |
| if (proposedFlow == null) { |
| if (controller.getGroup(controller.getRootGroupId()).isEmpty()) { |
| return; // no sync to perform |
| } else { |
| throw new UninheritableFlowException("Proposed configuration is empty, but the controller contains a data flow."); |
| } |
| } |
| |
| // determine if the controller has been initialized |
| final boolean initialized = controller.isInitialized(); |
| logger.debug("Synching FlowController with proposed flow: Controller is Initialized = {}", initialized); |
| |
| // serialize controller state to bytes |
| final byte[] existingFlow; |
| final boolean existingFlowEmpty; |
| try { |
| if (initialized) { |
| existingFlow = toBytes(controller); |
| existingFlowEmpty = controller.getGroup(controller.getRootGroupId()).isEmpty() && controller.getAllReportingTasks().isEmpty() && controller.getAllControllerServices().isEmpty(); |
| } else { |
| existingFlow = readFlowFromDisk(); |
| if (existingFlow == null || existingFlow.length == 0) { |
| existingFlowEmpty = true; |
| } else { |
| final Document document = parseFlowBytes(existingFlow); |
| final Element rootElement = document.getDocumentElement(); |
| final FlowEncodingVersion encodingVersion = FlowEncodingVersion.parse(rootElement); |
| |
| logger.trace("Setting controller thread counts"); |
| final Integer maxThreadCount = getInteger(rootElement, "maxThreadCount"); |
| if (maxThreadCount == null) { |
| controller.setMaxTimerDrivenThreadCount(getInt(rootElement, "maxTimerDrivenThreadCount")); |
| controller.setMaxEventDrivenThreadCount(getInt(rootElement, "maxEventDrivenThreadCount")); |
| } else { |
| controller.setMaxTimerDrivenThreadCount(maxThreadCount * 2 / 3); |
| controller.setMaxEventDrivenThreadCount(maxThreadCount / 3); |
| } |
| |
| final Element reportingTasksElement = DomUtils.getChild(rootElement, "reportingTasks"); |
| final List<Element> taskElements; |
| if (reportingTasksElement == null) { |
| taskElements = Collections.emptyList(); |
| } else { |
| taskElements = DomUtils.getChildElementsByTagName(reportingTasksElement, "reportingTask"); |
| } |
| |
| final Element controllerServicesElement = DomUtils.getChild(rootElement, "controllerServices"); |
| final List<Element> unrootedControllerServiceElements; |
| if (controllerServicesElement == null) { |
| unrootedControllerServiceElements = Collections.emptyList(); |
| } else { |
| unrootedControllerServiceElements = DomUtils.getChildElementsByTagName(controllerServicesElement, "controllerService"); |
| } |
| |
| logger.trace("Parsing process group from DOM"); |
| final Element rootGroupElement = (Element) rootElement.getElementsByTagName("rootGroup").item(0); |
| final ProcessGroupDTO rootGroupDto = FlowFromDOMFactory.getProcessGroup(null, rootGroupElement, encryptor, encodingVersion); |
| existingFlowEmpty = taskElements.isEmpty() && unrootedControllerServiceElements.isEmpty() && isEmpty(rootGroupDto); |
| logger.debug("Existing Flow Empty = {}", existingFlowEmpty); |
| } |
| } |
| } catch (final IOException e) { |
| throw new FlowSerializationException(e); |
| } |
| |
| logger.trace("Exporting snippets from controller"); |
| final byte[] existingSnippets = controller.getSnippetManager().export(); |
| |
| logger.trace("Getting Authorizer fingerprint from controller"); |
| |
| final byte[] existingAuthFingerprint; |
| final AbstractPolicyBasedAuthorizer policyBasedAuthorizer; |
| final Authorizer authorizer = controller.getAuthorizer(); |
| |
| if (authorizer instanceof AbstractPolicyBasedAuthorizer) { |
| policyBasedAuthorizer = (AbstractPolicyBasedAuthorizer) authorizer; |
| existingAuthFingerprint = policyBasedAuthorizer.getFingerprint().getBytes(StandardCharsets.UTF_8); |
| } else { |
| existingAuthFingerprint = null; |
| policyBasedAuthorizer = null; |
| } |
| |
| final DataFlow existingDataFlow = new StandardDataFlow(existingFlow, existingSnippets, existingAuthFingerprint); |
| |
| // check that the proposed flow is inheritable by the controller |
| try { |
| if (!existingFlowEmpty) { |
| logger.trace("Checking flow inheritability"); |
| final String problemInheritingFlow = checkFlowInheritability(existingDataFlow, proposedFlow, controller); |
| if (problemInheritingFlow != null) { |
| throw new UninheritableFlowException("Proposed configuration is not inheritable by the flow controller because of flow differences: " + problemInheritingFlow); |
| } |
| } |
| } catch (final FingerprintException fe) { |
| throw new FlowSerializationException("Failed to generate flow fingerprints", fe); |
| } |
| |
| logger.trace("Checking authorizer inheritability"); |
| |
| final AuthorizerInheritability authInheritability = checkAuthorizerInheritability(existingDataFlow, proposedFlow); |
| if (!authInheritability.isInheritable() && authInheritability.getReason() != null) { |
| throw new UninheritableFlowException("Proposed Authorizer is not inheritable by the flow controller because of Authorizer differences: " + authInheritability.getReason()); |
| } |
| |
| // create document by parsing proposed flow bytes |
| logger.trace("Parsing proposed flow bytes as DOM document"); |
| final Document configuration = parseFlowBytes(proposedFlow.getFlow()); |
| |
| // attempt to sync controller with proposed flow |
| try { |
| if (configuration != null) { |
| synchronized (configuration) { |
| // get the root element |
| final Element rootElement = (Element) configuration.getElementsByTagName("flowController").item(0); |
| final FlowEncodingVersion encodingVersion = FlowEncodingVersion.parse(rootElement); |
| |
| // set controller config |
| logger.trace("Updating flow config"); |
| final Integer maxThreadCount = getInteger(rootElement, "maxThreadCount"); |
| if (maxThreadCount == null) { |
| controller.setMaxTimerDrivenThreadCount(getInt(rootElement, "maxTimerDrivenThreadCount")); |
| controller.setMaxEventDrivenThreadCount(getInt(rootElement, "maxEventDrivenThreadCount")); |
| } else { |
| controller.setMaxTimerDrivenThreadCount(maxThreadCount * 2 / 3); |
| controller.setMaxEventDrivenThreadCount(maxThreadCount / 3); |
| } |
| |
| // get the root group XML element |
| final Element rootGroupElement = (Element) rootElement.getElementsByTagName("rootGroup").item(0); |
| |
| // if this controller isn't initialized or its empty, add the root group, otherwise update |
| final ProcessGroup rootGroup; |
| if (!initialized || existingFlowEmpty) { |
| logger.trace("Adding root process group"); |
| rootGroup = addProcessGroup(controller, /* parent group */ null, rootGroupElement, encryptor, encodingVersion); |
| } else { |
| logger.trace("Updating root process group"); |
| rootGroup = updateProcessGroup(controller, /* parent group */ null, rootGroupElement, encryptor, encodingVersion); |
| } |
| |
| // If there are any Templates that do not exist in the Proposed Flow that do exist in the 'existing flow', we need |
| // to ensure that we also add those to the appropriate Process Groups, so that we don't lose them. |
| final Document existingFlowConfiguration = parseFlowBytes(existingFlow); |
| if (existingFlowConfiguration != null) { |
| final Element existingRootElement = (Element) existingFlowConfiguration.getElementsByTagName("flowController").item(0); |
| if (existingRootElement != null) { |
| final Element existingRootGroupElement = (Element) existingRootElement.getElementsByTagName("rootGroup").item(0); |
| if (existingRootElement != null) { |
| final FlowEncodingVersion existingEncodingVersion = FlowEncodingVersion.parse(existingFlowConfiguration.getDocumentElement()); |
| addLocalTemplates(existingRootGroupElement, rootGroup, existingEncodingVersion); |
| } |
| } |
| } |
| |
| final Element controllerServicesElement = DomUtils.getChild(rootElement, "controllerServices"); |
| if (controllerServicesElement != null) { |
| final List<Element> serviceElements = DomUtils.getChildElementsByTagName(controllerServicesElement, "controllerService"); |
| |
| if (!initialized || existingFlowEmpty) { |
| // If the encoding version is null, we are loading a flow from NiFi 0.x, where Controller |
| // Services could not be scoped by Process Group. As a result, we want to move the Process Groups |
| // to the root Group. Otherwise, we want to use a null group, which indicates a Controller-level |
| // Controller Service. |
| final ProcessGroup group = (encodingVersion == null) ? rootGroup : null; |
| ControllerServiceLoader.loadControllerServices(serviceElements, controller, group, encryptor, controller.getBulletinRepository(), autoResumeState); |
| } else { |
| for (final Element serviceElement : serviceElements) { |
| updateControllerService(controller, serviceElement, encryptor); |
| } |
| } |
| } |
| |
| scaleRootGroup(rootGroup, encodingVersion); |
| |
| final Element reportingTasksElement = DomUtils.getChild(rootElement, "reportingTasks"); |
| if (reportingTasksElement != null) { |
| final List<Element> taskElements = DomUtils.getChildElementsByTagName(reportingTasksElement, "reportingTask"); |
| for (final Element taskElement : taskElements) { |
| if (!initialized || existingFlowEmpty) { |
| addReportingTask(controller, taskElement, encryptor); |
| } else { |
| updateReportingTask(controller, taskElement, encryptor); |
| } |
| } |
| } |
| } |
| } |
| |
| // clear the snippets that are currently in memory |
| logger.trace("Clearing existing snippets"); |
| final SnippetManager snippetManager = controller.getSnippetManager(); |
| snippetManager.clear(); |
| |
| // if proposed flow has any snippets, load them |
| logger.trace("Loading proposed snippets"); |
| final byte[] proposedSnippets = proposedFlow.getSnippets(); |
| if (proposedSnippets != null && proposedSnippets.length > 0) { |
| for (final StandardSnippet snippet : SnippetManager.parseBytes(proposedSnippets)) { |
| snippetManager.addSnippet(snippet); |
| } |
| } |
| |
| // if auths are inheritable and we have a policy based authorizer, then inherit |
| if (authInheritability.isInheritable() && policyBasedAuthorizer != null) { |
| logger.trace("Inheriting authorizations"); |
| final String proposedAuthFingerprint = new String(proposedFlow.getAuthorizerFingerprint(), StandardCharsets.UTF_8); |
| policyBasedAuthorizer.inheritFingerprint(proposedAuthFingerprint); |
| } |
| |
| logger.debug("Finished syncing flows"); |
| } catch (final Exception ex) { |
| throw new FlowSynchronizationException(ex); |
| } |
| } |
| |
| private void addLocalTemplates(final Element processGroupElement, final ProcessGroup processGroup, final FlowEncodingVersion encodingVersion) { |
| // Replace the templates with those from the proposed flow |
| final List<Element> templateNodeList = getChildrenByTagName(processGroupElement, "template"); |
| if (templateNodeList != null) { |
| for (final Element templateElement : templateNodeList) { |
| final TemplateDTO templateDto = TemplateUtils.parseDto(templateElement); |
| final Template template = new Template(templateDto); |
| |
| // If the Process Group does not have the template, add it. |
| if (processGroup.getTemplate(template.getIdentifier()) == null) { |
| processGroup.addTemplate(template); |
| } |
| } |
| } |
| |
| final List<Element> childGroupElements = getChildrenByTagName(processGroupElement, "processGroup"); |
| for (final Element childGroupElement : childGroupElements) { |
| final String childGroupId = getString(childGroupElement, "id"); |
| final ProcessGroup childGroup = processGroup.getProcessGroup(childGroupId); |
| addLocalTemplates(childGroupElement, childGroup, encodingVersion); |
| } |
| } |
| |
| void scaleRootGroup(final ProcessGroup rootGroup, final FlowEncodingVersion encodingVersion) { |
| if (encodingVersion == null || encodingVersion.getMajorVersion() < 1) { |
| // Calculate new Positions if the encoding version of the flow is older than 1.0. |
| PositionScaler.scale(rootGroup, 1.5, 1.34); |
| } |
| } |
| |
| private static boolean isEmpty(final ProcessGroupDTO dto) { |
| if (dto == null) { |
| return true; |
| } |
| |
| final FlowSnippetDTO contents = dto.getContents(); |
| if (contents == null) { |
| return true; |
| } |
| |
| return CollectionUtils.isEmpty(contents.getProcessors()) |
| && CollectionUtils.isEmpty(contents.getConnections()) |
| && CollectionUtils.isEmpty(contents.getFunnels()) |
| && CollectionUtils.isEmpty(contents.getLabels()) |
| && CollectionUtils.isEmpty(contents.getOutputPorts()) |
| && CollectionUtils.isEmpty(contents.getProcessGroups()) |
| && CollectionUtils.isEmpty(contents.getProcessors()) |
| && CollectionUtils.isEmpty(contents.getRemoteProcessGroups()); |
| } |
| |
| private static Document parseFlowBytes(final byte[] flow) throws FlowSerializationException { |
| // create document by parsing proposed flow bytes |
| try { |
| // create validating document builder |
| final SchemaFactory schemaFactory = SchemaFactory.newInstance(XMLConstants.W3C_XML_SCHEMA_NS_URI); |
| final Schema schema = schemaFactory.newSchema(FLOW_XSD_RESOURCE); |
| final DocumentBuilderFactory docFactory = DocumentBuilderFactory.newInstance(); |
| docFactory.setNamespaceAware(true); |
| docFactory.setSchema(schema); |
| |
| final DocumentBuilder docBuilder = docFactory.newDocumentBuilder(); |
| |
| // parse flow |
| return (flow == null || flow.length == 0) ? null : docBuilder.parse(new ByteArrayInputStream(flow)); |
| } catch (final SAXException | ParserConfigurationException | IOException ex) { |
| throw new FlowSerializationException(ex); |
| } |
| } |
| |
| private byte[] readFlowFromDisk() throws IOException { |
| final Path flowPath = NiFiProperties.getInstance().getFlowConfigurationFile().toPath(); |
| if (!Files.exists(flowPath) || Files.size(flowPath) == 0) { |
| return new byte[0]; |
| } |
| |
| final ByteArrayOutputStream baos = new ByteArrayOutputStream(); |
| |
| try (final InputStream in = Files.newInputStream(flowPath, StandardOpenOption.READ); |
| final InputStream gzipIn = new GZIPInputStream(in)) { |
| FileUtils.copy(gzipIn, baos); |
| } |
| |
| return baos.toByteArray(); |
| } |
| |
| private void updateControllerService(final FlowController controller, final Element controllerServiceElement, final StringEncryptor encryptor) { |
| final ControllerServiceDTO dto = FlowFromDOMFactory.getControllerService(controllerServiceElement, encryptor); |
| |
| final ControllerServiceState dtoState = ControllerServiceState.valueOf(dto.getState()); |
| final boolean dtoEnabled = (dtoState == ControllerServiceState.ENABLED || dtoState == ControllerServiceState.ENABLING); |
| |
| final ControllerServiceNode serviceNode = controller.getControllerServiceNode(dto.getId()); |
| final ControllerServiceState serviceState = serviceNode.getState(); |
| final boolean serviceEnabled = (serviceState == ControllerServiceState.ENABLED || serviceState == ControllerServiceState.ENABLING); |
| |
| if (dtoEnabled && !serviceEnabled) { |
| controller.enableControllerService(controller.getControllerServiceNode(dto.getId())); |
| } else if (!dtoEnabled && serviceEnabled) { |
| controller.disableControllerService(controller.getControllerServiceNode(dto.getId())); |
| } |
| } |
| |
| private void addReportingTask(final FlowController controller, final Element reportingTaskElement, final StringEncryptor encryptor) throws ReportingTaskInstantiationException { |
| final ReportingTaskDTO dto = FlowFromDOMFactory.getReportingTask(reportingTaskElement, encryptor); |
| |
| final ReportingTaskNode reportingTask = controller.createReportingTask(dto.getType(), dto.getId(), false); |
| reportingTask.setName(dto.getName()); |
| reportingTask.setComments(dto.getComments()); |
| reportingTask.setScheduldingPeriod(dto.getSchedulingPeriod()); |
| reportingTask.setSchedulingStrategy(SchedulingStrategy.valueOf(dto.getSchedulingStrategy())); |
| |
| reportingTask.setAnnotationData(dto.getAnnotationData()); |
| |
| for (final Map.Entry<String, String> entry : dto.getProperties().entrySet()) { |
| if (entry.getValue() == null) { |
| reportingTask.removeProperty(entry.getKey()); |
| } else { |
| reportingTask.setProperty(entry.getKey(), entry.getValue()); |
| } |
| } |
| |
| final ComponentLog componentLog = new SimpleProcessLogger(dto.getId(), reportingTask.getReportingTask()); |
| final ReportingInitializationContext config = new StandardReportingInitializationContext(dto.getId(), dto.getName(), |
| SchedulingStrategy.valueOf(dto.getSchedulingStrategy()), dto.getSchedulingPeriod(), componentLog, controller); |
| |
| try { |
| reportingTask.getReportingTask().initialize(config); |
| } catch (final InitializationException ie) { |
| throw new ReportingTaskInstantiationException("Failed to initialize reporting task of type " + dto.getType(), ie); |
| } |
| |
| if (autoResumeState) { |
| if (ScheduledState.RUNNING.name().equals(dto.getState())) { |
| try { |
| controller.startReportingTask(reportingTask); |
| } catch (final Exception e) { |
| logger.error("Failed to start {} due to {}", reportingTask, e); |
| if (logger.isDebugEnabled()) { |
| logger.error("", e); |
| } |
| controller.getBulletinRepository().addBulletin(BulletinFactory.createBulletin( |
| "Reporting Tasks", Severity.ERROR.name(), "Failed to start " + reportingTask + " due to " + e)); |
| } |
| } else if (ScheduledState.DISABLED.name().equals(dto.getState())) { |
| try { |
| controller.disableReportingTask(reportingTask); |
| } catch (final Exception e) { |
| logger.error("Failed to mark {} as disabled due to {}", reportingTask, e); |
| if (logger.isDebugEnabled()) { |
| logger.error("", e); |
| } |
| controller.getBulletinRepository().addBulletin(BulletinFactory.createBulletin( |
| "Reporting Tasks", Severity.ERROR.name(), "Failed to mark " + reportingTask + " as disabled due to " + e)); |
| } |
| } |
| } |
| } |
| |
| private void updateReportingTask(final FlowController controller, final Element reportingTaskElement, final StringEncryptor encryptor) { |
| final ReportingTaskDTO dto = FlowFromDOMFactory.getReportingTask(reportingTaskElement, encryptor); |
| final ReportingTaskNode taskNode = controller.getReportingTaskNode(dto.getId()); |
| |
| if (!taskNode.getScheduledState().name().equals(dto.getState())) { |
| try { |
| switch (ScheduledState.valueOf(dto.getState())) { |
| case DISABLED: |
| if (taskNode.isRunning()) { |
| controller.stopReportingTask(taskNode); |
| } |
| controller.disableReportingTask(taskNode); |
| break; |
| case RUNNING: |
| if (taskNode.getScheduledState() == ScheduledState.DISABLED) { |
| controller.enableReportingTask(taskNode); |
| } |
| controller.startReportingTask(taskNode); |
| break; |
| case STOPPED: |
| if (taskNode.getScheduledState() == ScheduledState.DISABLED) { |
| controller.enableReportingTask(taskNode); |
| } else if (taskNode.getScheduledState() == ScheduledState.RUNNING) { |
| controller.stopReportingTask(taskNode); |
| } |
| break; |
| } |
| } catch (final IllegalStateException ise) { |
| logger.error("Failed to change Scheduled State of {} from {} to {} due to {}", taskNode, taskNode.getScheduledState().name(), dto.getState(), ise.toString()); |
| logger.error("", ise); |
| |
| // create bulletin for the Processor Node |
| controller.getBulletinRepository().addBulletin(BulletinFactory.createBulletin("Node Reconnection", Severity.ERROR.name(), |
| "Failed to change Scheduled State of " + taskNode + " from " + taskNode.getScheduledState().name() + " to " + dto.getState() + " due to " + ise.toString())); |
| |
| // create bulletin at Controller level. |
| controller.getBulletinRepository().addBulletin(BulletinFactory.createBulletin("Node Reconnection", Severity.ERROR.name(), |
| "Failed to change Scheduled State of " + taskNode + " from " + taskNode.getScheduledState().name() + " to " + dto.getState() + " due to " + ise.toString())); |
| } |
| } |
| } |
| |
| private ProcessGroup updateProcessGroup(final FlowController controller, final ProcessGroup parentGroup, final Element processGroupElement, |
| final StringEncryptor encryptor, final FlowEncodingVersion encodingVersion) throws ProcessorInstantiationException { |
| |
| // get the parent group ID |
| final String parentId = (parentGroup == null) ? null : parentGroup.getIdentifier(); |
| |
| // get the process group |
| final ProcessGroupDTO processGroupDto = FlowFromDOMFactory.getProcessGroup(parentId, processGroupElement, encryptor, encodingVersion); |
| |
| // update the process group |
| if (parentId == null) { |
| |
| /* |
| * Labels are not included in the "inherit flow" algorithm, so we cannot |
| * blindly update them because they may not exist in the current flow. |
| * Therefore, we first remove all labels, and then let the updating |
| * process add labels defined in the new flow. |
| */ |
| final ProcessGroup root = controller.getGroup(controller.getRootGroupId()); |
| for (final Label label : root.findAllLabels()) { |
| label.getProcessGroup().removeLabel(label); |
| } |
| } |
| |
| // update the process group |
| controller.updateProcessGroup(processGroupDto); |
| |
| // get the real process group and ID |
| final ProcessGroup processGroup = controller.getGroup(processGroupDto.getId()); |
| |
| // Update Controller Services |
| final List<Element> serviceNodeList = getChildrenByTagName(processGroupElement, "controllerService"); |
| for (final Element serviceNodeElement : serviceNodeList) { |
| updateControllerService(controller, serviceNodeElement, encryptor); |
| } |
| |
| // processors & ports cannot be updated - they must be the same. Except for the scheduled state. |
| final List<Element> processorNodeList = getChildrenByTagName(processGroupElement, "processor"); |
| for (final Element processorElement : processorNodeList) { |
| final ProcessorDTO dto = FlowFromDOMFactory.getProcessor(processorElement, encryptor); |
| final ProcessorNode procNode = processGroup.getProcessor(dto.getId()); |
| |
| if (!procNode.getScheduledState().name().equals(dto.getState())) { |
| try { |
| switch (ScheduledState.valueOf(dto.getState())) { |
| case DISABLED: |
| // switch processor do disabled. This means we have to stop it (if it's already stopped, this method does nothing), |
| // and then we have to disable it. |
| procNode.getProcessGroup().stopProcessor(procNode); |
| procNode.getProcessGroup().disableProcessor(procNode); |
| break; |
| case RUNNING: |
| // we want to run now. Make sure processor is not disabled and then start it. |
| procNode.getProcessGroup().enableProcessor(procNode); |
| procNode.getProcessGroup().startProcessor(procNode); |
| break; |
| case STOPPED: |
| if (procNode.getScheduledState() == ScheduledState.DISABLED) { |
| procNode.getProcessGroup().enableProcessor(procNode); |
| } else if (procNode.getScheduledState() == ScheduledState.RUNNING) { |
| procNode.getProcessGroup().stopProcessor(procNode); |
| } |
| break; |
| } |
| } catch (final IllegalStateException ise) { |
| logger.error("Failed to change Scheduled State of {} from {} to {} due to {}", procNode, procNode.getScheduledState().name(), dto.getState(), ise.toString()); |
| logger.error("", ise); |
| |
| // create bulletin for the Processor Node |
| controller.getBulletinRepository().addBulletin(BulletinFactory.createBulletin(procNode, "Node Reconnection", Severity.ERROR.name(), |
| "Failed to change Scheduled State of " + procNode + " from " + procNode.getScheduledState().name() + " to " + dto.getState() + " due to " + ise.toString())); |
| |
| // create bulletin at Controller level. |
| controller.getBulletinRepository().addBulletin(BulletinFactory.createBulletin("Node Reconnection", Severity.ERROR.name(), |
| "Failed to change Scheduled State of " + procNode + " from " + procNode.getScheduledState().name() + " to " + dto.getState() + " due to " + ise.toString())); |
| } |
| } |
| } |
| |
| final List<Element> inputPortList = getChildrenByTagName(processGroupElement, "inputPort"); |
| for (final Element portElement : inputPortList) { |
| final PortDTO dto = FlowFromDOMFactory.getPort(portElement); |
| final Port port = processGroup.getInputPort(dto.getId()); |
| |
| if (!port.getScheduledState().name().equals(dto.getState())) { |
| switch (ScheduledState.valueOf(dto.getState())) { |
| case DISABLED: |
| // switch processor do disabled. This means we have to stop it (if it's already stopped, this method does nothing), |
| // and then we have to disable it. |
| port.getProcessGroup().stopInputPort(port); |
| port.getProcessGroup().disableInputPort(port); |
| break; |
| case RUNNING: |
| // we want to run now. Make sure processor is not disabled and then start it. |
| port.getProcessGroup().enableInputPort(port); |
| port.getProcessGroup().startInputPort(port); |
| break; |
| case STOPPED: |
| if (port.getScheduledState() == ScheduledState.DISABLED) { |
| port.getProcessGroup().enableInputPort(port); |
| } else if (port.getScheduledState() == ScheduledState.RUNNING) { |
| port.getProcessGroup().stopInputPort(port); |
| } |
| break; |
| } |
| } |
| } |
| |
| final List<Element> outputPortList = getChildrenByTagName(processGroupElement, "outputPort"); |
| for (final Element portElement : outputPortList) { |
| final PortDTO dto = FlowFromDOMFactory.getPort(portElement); |
| final Port port = processGroup.getOutputPort(dto.getId()); |
| |
| if (!port.getScheduledState().name().equals(dto.getState())) { |
| switch (ScheduledState.valueOf(dto.getState())) { |
| case DISABLED: |
| // switch processor do disabled. This means we have to stop it (if it's already stopped, this method does nothing), |
| // and then we have to disable it. |
| port.getProcessGroup().stopOutputPort(port); |
| port.getProcessGroup().disableOutputPort(port); |
| break; |
| case RUNNING: |
| // we want to run now. Make sure processor is not disabled and then start it. |
| port.getProcessGroup().enableOutputPort(port); |
| port.getProcessGroup().startOutputPort(port); |
| break; |
| case STOPPED: |
| if (port.getScheduledState() == ScheduledState.DISABLED) { |
| port.getProcessGroup().enableOutputPort(port); |
| } else if (port.getScheduledState() == ScheduledState.RUNNING) { |
| port.getProcessGroup().stopOutputPort(port); |
| } |
| break; |
| } |
| } |
| } |
| |
| // add labels |
| final List<Element> labelNodeList = getChildrenByTagName(processGroupElement, "label"); |
| for (final Element labelElement : labelNodeList) { |
| final LabelDTO labelDTO = FlowFromDOMFactory.getLabel(labelElement); |
| final Label label = controller.createLabel(labelDTO.getId(), labelDTO.getLabel()); |
| label.setStyle(labelDTO.getStyle()); |
| label.setPosition(new Position(labelDTO.getPosition().getX(), labelDTO.getPosition().getY())); |
| if (labelDTO.getWidth() != null && labelDTO.getHeight() != null) { |
| label.setSize(new Size(labelDTO.getWidth(), labelDTO.getHeight())); |
| } |
| |
| processGroup.addLabel(label); |
| } |
| |
| // update nested process groups (recursively) |
| final List<Element> nestedProcessGroupNodeList = getChildrenByTagName(processGroupElement, "processGroup"); |
| for (final Element nestedProcessGroupElement : nestedProcessGroupNodeList) { |
| updateProcessGroup(controller, processGroup, nestedProcessGroupElement, encryptor, encodingVersion); |
| } |
| |
| // update connections |
| final List<Element> connectionNodeList = getChildrenByTagName(processGroupElement, "connection"); |
| for (final Element connectionElement : connectionNodeList) { |
| final ConnectionDTO dto = FlowFromDOMFactory.getConnection(connectionElement); |
| |
| final Connection connection = processGroup.getConnection(dto.getId()); |
| connection.setName(dto.getName()); |
| connection.setProcessGroup(processGroup); |
| |
| if (dto.getLabelIndex() != null) { |
| connection.setLabelIndex(dto.getLabelIndex()); |
| } |
| if (dto.getzIndex() != null) { |
| connection.setZIndex(dto.getzIndex()); |
| } |
| |
| final List<Position> bendPoints = new ArrayList<>(); |
| for (final PositionDTO bend : dto.getBends()) { |
| bendPoints.add(new Position(bend.getX(), bend.getY())); |
| } |
| connection.setBendPoints(bendPoints); |
| |
| List<FlowFilePrioritizer> newPrioritizers = null; |
| final List<String> prioritizers = dto.getPrioritizers(); |
| if (prioritizers != null) { |
| final List<String> newPrioritizersClasses = new ArrayList<>(prioritizers); |
| newPrioritizers = new ArrayList<>(); |
| for (final String className : newPrioritizersClasses) { |
| try { |
| newPrioritizers.add(controller.createPrioritizer(className)); |
| } catch (final ClassNotFoundException | InstantiationException | IllegalAccessException e) { |
| throw new IllegalArgumentException("Unable to set prioritizer " + className + ": " + e); |
| } |
| } |
| } |
| |
| if (newPrioritizers != null) { |
| connection.getFlowFileQueue().setPriorities(newPrioritizers); |
| } |
| |
| if (dto.getBackPressureObjectThreshold() != null) { |
| connection.getFlowFileQueue().setBackPressureObjectThreshold(dto.getBackPressureObjectThreshold()); |
| } |
| |
| if (dto.getBackPressureDataSizeThreshold() != null && !dto.getBackPressureDataSizeThreshold().trim().isEmpty()) { |
| connection.getFlowFileQueue().setBackPressureDataSizeThreshold(dto.getBackPressureDataSizeThreshold()); |
| } |
| |
| if (dto.getFlowFileExpiration() != null) { |
| connection.getFlowFileQueue().setFlowFileExpiration(dto.getFlowFileExpiration()); |
| } |
| } |
| |
| // Replace the templates with those from the proposed flow |
| final List<Element> templateNodeList = getChildrenByTagName(processGroupElement, "template"); |
| for (final Element templateElement : templateNodeList) { |
| final TemplateDTO templateDto = TemplateUtils.parseDto(templateElement); |
| final Template template = new Template(templateDto); |
| |
| // If the Process Group already has the template, remove it and add it again. We do this |
| // to ensure that all of the nodes have the same view of the template. Templates are immutable, |
| // so any two nodes that have a template with the same ID should have the exact same template. |
| // This just makes sure that they do. |
| if (processGroup.getTemplate(template.getIdentifier()) != null) { |
| processGroup.removeTemplate(template); |
| } |
| processGroup.addTemplate(template); |
| } |
| |
| return processGroup; |
| } |
| |
| private Position toPosition(final PositionDTO dto) { |
| return new Position(dto.getX(), dto.getY()); |
| } |
| |
| private void updateProcessor(final ProcessorNode procNode, final ProcessorDTO processorDTO, final ProcessGroup processGroup, final FlowController controller) |
| throws ProcessorInstantiationException { |
| final ProcessorConfigDTO config = processorDTO.getConfig(); |
| procNode.setPosition(toPosition(processorDTO.getPosition())); |
| procNode.setName(processorDTO.getName()); |
| procNode.setStyle(processorDTO.getStyle()); |
| procNode.setProcessGroup(processGroup); |
| procNode.setComments(config.getComments()); |
| procNode.setLossTolerant(config.isLossTolerant()); |
| procNode.setPenalizationPeriod(config.getPenaltyDuration()); |
| procNode.setYieldPeriod(config.getYieldDuration()); |
| procNode.setBulletinLevel(LogLevel.valueOf(config.getBulletinLevel())); |
| |
| if (config.getSchedulingStrategy() != null) { |
| procNode.setSchedulingStrategy(SchedulingStrategy.valueOf(config.getSchedulingStrategy())); |
| } |
| |
| // must set scheduling strategy before these two |
| procNode.setMaxConcurrentTasks(config.getConcurrentlySchedulableTaskCount()); |
| procNode.setScheduldingPeriod(config.getSchedulingPeriod()); |
| if (config.getRunDurationMillis() != null) { |
| procNode.setRunDuration(config.getRunDurationMillis(), TimeUnit.MILLISECONDS); |
| } |
| |
| procNode.setAnnotationData(config.getAnnotationData()); |
| |
| if (config.getAutoTerminatedRelationships() != null) { |
| final Set<Relationship> relationships = new HashSet<>(); |
| for (final String rel : config.getAutoTerminatedRelationships()) { |
| relationships.add(procNode.getRelationship(rel)); |
| } |
| procNode.setAutoTerminatedRelationships(relationships); |
| } |
| |
| for (final Map.Entry<String, String> entry : config.getProperties().entrySet()) { |
| if (entry.getValue() == null) { |
| procNode.removeProperty(entry.getKey()); |
| } else { |
| procNode.setProperty(entry.getKey(), entry.getValue()); |
| } |
| } |
| |
| final ScheduledState scheduledState = ScheduledState.valueOf(processorDTO.getState()); |
| if (ScheduledState.RUNNING.equals(scheduledState)) { |
| controller.startProcessor(processGroup.getIdentifier(), procNode.getIdentifier()); |
| } else if (ScheduledState.DISABLED.equals(scheduledState)) { |
| processGroup.disableProcessor(procNode); |
| } |
| } |
| |
| private ProcessGroup addProcessGroup(final FlowController controller, final ProcessGroup parentGroup, final Element processGroupElement, |
| final StringEncryptor encryptor, final FlowEncodingVersion encodingVersion) throws ProcessorInstantiationException { |
| // get the parent group ID |
| final String parentId = (parentGroup == null) ? null : parentGroup.getIdentifier(); |
| |
| // add the process group |
| final ProcessGroupDTO processGroupDTO = FlowFromDOMFactory.getProcessGroup(parentId, processGroupElement, encryptor, encodingVersion); |
| final ProcessGroup processGroup = controller.createProcessGroup(processGroupDTO.getId()); |
| processGroup.setComments(processGroupDTO.getComments()); |
| processGroup.setPosition(toPosition(processGroupDTO.getPosition())); |
| processGroup.setName(processGroupDTO.getName()); |
| processGroup.setParent(parentGroup); |
| if (parentGroup == null) { |
| controller.setRootGroup(processGroup); |
| } else { |
| parentGroup.addProcessGroup(processGroup); |
| } |
| |
| // Add Controller Services |
| final List<Element> serviceNodeList = getChildrenByTagName(processGroupElement, "controllerService"); |
| if (!serviceNodeList.isEmpty()) { |
| ControllerServiceLoader.loadControllerServices(serviceNodeList, controller, processGroup, encryptor, controller.getBulletinRepository(), autoResumeState); |
| } |
| |
| // add processors |
| final List<Element> processorNodeList = getChildrenByTagName(processGroupElement, "processor"); |
| for (final Element processorElement : processorNodeList) { |
| final ProcessorDTO processorDTO = FlowFromDOMFactory.getProcessor(processorElement, encryptor); |
| final ProcessorNode procNode = controller.createProcessor(processorDTO.getType(), processorDTO.getId(), false); |
| processGroup.addProcessor(procNode); |
| updateProcessor(procNode, processorDTO, processGroup, controller); |
| } |
| |
| // add input ports |
| final List<Element> inputPortNodeList = getChildrenByTagName(processGroupElement, "inputPort"); |
| for (final Element inputPortElement : inputPortNodeList) { |
| final PortDTO portDTO = FlowFromDOMFactory.getPort(inputPortElement); |
| |
| final Port port; |
| if (processGroup.isRootGroup()) { |
| port = controller.createRemoteInputPort(portDTO.getId(), portDTO.getName()); |
| } else { |
| port = controller.createLocalInputPort(portDTO.getId(), portDTO.getName()); |
| } |
| |
| port.setPosition(toPosition(portDTO.getPosition())); |
| port.setComments(portDTO.getComments()); |
| port.setProcessGroup(processGroup); |
| |
| final Set<String> userControls = portDTO.getUserAccessControl(); |
| if (userControls != null && !userControls.isEmpty()) { |
| if (!(port instanceof RootGroupPort)) { |
| throw new IllegalStateException("Attempting to add User Access Controls to " + port.getIdentifier() + ", but it is not a RootGroupPort"); |
| } |
| ((RootGroupPort) port).setUserAccessControl(userControls); |
| } |
| final Set<String> groupControls = portDTO.getGroupAccessControl(); |
| if (groupControls != null && !groupControls.isEmpty()) { |
| if (!(port instanceof RootGroupPort)) { |
| throw new IllegalStateException("Attempting to add Group Access Controls to " + port.getIdentifier() + ", but it is not a RootGroupPort"); |
| } |
| ((RootGroupPort) port).setGroupAccessControl(groupControls); |
| } |
| |
| processGroup.addInputPort(port); |
| if (portDTO.getConcurrentlySchedulableTaskCount() != null) { |
| port.setMaxConcurrentTasks(portDTO.getConcurrentlySchedulableTaskCount()); |
| } |
| |
| final ScheduledState scheduledState = ScheduledState.valueOf(portDTO.getState()); |
| if (ScheduledState.RUNNING.equals(scheduledState)) { |
| controller.startConnectable(port); |
| } else if (ScheduledState.DISABLED.equals(scheduledState)) { |
| processGroup.disableInputPort(port); |
| } |
| } |
| |
| // add output ports |
| final List<Element> outputPortNodeList = getChildrenByTagName(processGroupElement, "outputPort"); |
| for (final Element outputPortElement : outputPortNodeList) { |
| final PortDTO portDTO = FlowFromDOMFactory.getPort(outputPortElement); |
| |
| final Port port; |
| if (processGroup.isRootGroup()) { |
| port = controller.createRemoteOutputPort(portDTO.getId(), portDTO.getName()); |
| } else { |
| port = controller.createLocalOutputPort(portDTO.getId(), portDTO.getName()); |
| } |
| port.setPosition(toPosition(portDTO.getPosition())); |
| port.setComments(portDTO.getComments()); |
| port.setProcessGroup(processGroup); |
| |
| final Set<String> userControls = portDTO.getUserAccessControl(); |
| if (userControls != null && !userControls.isEmpty()) { |
| if (!(port instanceof RootGroupPort)) { |
| throw new IllegalStateException("Attempting to add User Access Controls to " + port.getIdentifier() + ", but it is not a RootGroupPort"); |
| } |
| ((RootGroupPort) port).setUserAccessControl(userControls); |
| } |
| final Set<String> groupControls = portDTO.getGroupAccessControl(); |
| if (groupControls != null && !groupControls.isEmpty()) { |
| if (!(port instanceof RootGroupPort)) { |
| throw new IllegalStateException("Attempting to add Group Access Controls to " + port.getIdentifier() + ", but it is not a RootGroupPort"); |
| } |
| ((RootGroupPort) port).setGroupAccessControl(groupControls); |
| } |
| |
| processGroup.addOutputPort(port); |
| if (portDTO.getConcurrentlySchedulableTaskCount() != null) { |
| port.setMaxConcurrentTasks(portDTO.getConcurrentlySchedulableTaskCount()); |
| } |
| |
| final ScheduledState scheduledState = ScheduledState.valueOf(portDTO.getState()); |
| if (ScheduledState.RUNNING.equals(scheduledState)) { |
| controller.startConnectable(port); |
| } else if (ScheduledState.DISABLED.equals(scheduledState)) { |
| processGroup.disableOutputPort(port); |
| } |
| } |
| |
| // add funnels |
| final List<Element> funnelNodeList = getChildrenByTagName(processGroupElement, "funnel"); |
| for (final Element funnelElement : funnelNodeList) { |
| final FunnelDTO funnelDTO = FlowFromDOMFactory.getFunnel(funnelElement); |
| final Funnel funnel = controller.createFunnel(funnelDTO.getId()); |
| funnel.setPosition(toPosition(funnelDTO.getPosition())); |
| |
| // Since this is called during startup, we want to add the funnel without enabling it |
| // and then tell the controller to enable it. This way, if the controller is not fully |
| // initialized, the starting of the funnel is delayed until the controller is ready. |
| processGroup.addFunnel(funnel, false); |
| controller.startConnectable(funnel); |
| } |
| |
| // add labels |
| final List<Element> labelNodeList = getChildrenByTagName(processGroupElement, "label"); |
| for (final Element labelElement : labelNodeList) { |
| final LabelDTO labelDTO = FlowFromDOMFactory.getLabel(labelElement); |
| final Label label = controller.createLabel(labelDTO.getId(), labelDTO.getLabel()); |
| label.setStyle(labelDTO.getStyle()); |
| |
| label.setPosition(toPosition(labelDTO.getPosition())); |
| label.setSize(new Size(labelDTO.getWidth(), labelDTO.getHeight())); |
| processGroup.addLabel(label); |
| } |
| |
| // add nested process groups (recursively) |
| final List<Element> nestedProcessGroupNodeList = getChildrenByTagName(processGroupElement, "processGroup"); |
| for (final Element nestedProcessGroupElement : nestedProcessGroupNodeList) { |
| addProcessGroup(controller, processGroup, nestedProcessGroupElement, encryptor, encodingVersion); |
| } |
| |
| // add remote process group |
| final List<Element> remoteProcessGroupNodeList = getChildrenByTagName(processGroupElement, "remoteProcessGroup"); |
| for (final Element remoteProcessGroupElement : remoteProcessGroupNodeList) { |
| final RemoteProcessGroupDTO remoteGroupDto = FlowFromDOMFactory.getRemoteProcessGroup(remoteProcessGroupElement, encryptor); |
| final RemoteProcessGroup remoteGroup = controller.createRemoteProcessGroup(remoteGroupDto.getId(), remoteGroupDto.getTargetUri()); |
| remoteGroup.setComments(remoteGroupDto.getComments()); |
| remoteGroup.setPosition(toPosition(remoteGroupDto.getPosition())); |
| final String name = remoteGroupDto.getName(); |
| if (name != null && !name.trim().isEmpty()) { |
| remoteGroup.setName(name); |
| } |
| remoteGroup.setProcessGroup(processGroup); |
| remoteGroup.setCommunicationsTimeout(remoteGroupDto.getCommunicationsTimeout()); |
| |
| if (remoteGroupDto.getYieldDuration() != null) { |
| remoteGroup.setYieldDuration(remoteGroupDto.getYieldDuration()); |
| } |
| |
| final String transportProtocol = remoteGroupDto.getTransportProtocol(); |
| if (transportProtocol != null && !transportProtocol.trim().isEmpty()) { |
| remoteGroup.setTransportProtocol(SiteToSiteTransportProtocol.valueOf(transportProtocol.toUpperCase())); |
| } |
| |
| if (remoteGroupDto.getProxyHost() != null) { |
| remoteGroup.setProxyHost(remoteGroupDto.getProxyHost()); |
| } |
| |
| if (remoteGroupDto.getProxyPort() != null) { |
| remoteGroup.setProxyPort(remoteGroupDto.getProxyPort()); |
| } |
| |
| if (remoteGroupDto.getProxyUser() != null) { |
| remoteGroup.setProxyUser(remoteGroupDto.getProxyUser()); |
| } |
| |
| if (remoteGroupDto.getProxyPassword() != null) { |
| remoteGroup.setProxyPassword(remoteGroupDto.getProxyPassword()); |
| } |
| |
| final Set<RemoteProcessGroupPortDescriptor> inputPorts = new HashSet<>(); |
| for (final Element portElement : getChildrenByTagName(remoteProcessGroupElement, "inputPort")) { |
| inputPorts.add(FlowFromDOMFactory.getRemoteProcessGroupPort(portElement)); |
| } |
| remoteGroup.setInputPorts(inputPorts); |
| |
| final Set<RemoteProcessGroupPortDescriptor> outputPorts = new HashSet<>(); |
| for (final Element portElement : getChildrenByTagName(remoteProcessGroupElement, "outputPort")) { |
| outputPorts.add(FlowFromDOMFactory.getRemoteProcessGroupPort(portElement)); |
| } |
| remoteGroup.setOutputPorts(outputPorts); |
| processGroup.addRemoteProcessGroup(remoteGroup); |
| |
| for (final RemoteProcessGroupPortDescriptor remoteGroupPortDTO : outputPorts) { |
| final RemoteGroupPort port = remoteGroup.getOutputPort(remoteGroupPortDTO.getId()); |
| if (Boolean.TRUE.equals(remoteGroupPortDTO.isTransmitting())) { |
| controller.startTransmitting(port); |
| } |
| } |
| for (final RemoteProcessGroupPortDescriptor remoteGroupPortDTO : inputPorts) { |
| final RemoteGroupPort port = remoteGroup.getInputPort(remoteGroupPortDTO.getId()); |
| if (Boolean.TRUE.equals(remoteGroupPortDTO.isTransmitting())) { |
| controller.startTransmitting(port); |
| } |
| } |
| } |
| |
| // add connections |
| final List<Element> connectionNodeList = getChildrenByTagName(processGroupElement, "connection"); |
| for (final Element connectionElement : connectionNodeList) { |
| final ConnectionDTO dto = FlowFromDOMFactory.getConnection(connectionElement); |
| |
| final Connectable source; |
| final ConnectableDTO sourceDto = dto.getSource(); |
| if (ConnectableType.REMOTE_OUTPUT_PORT.name().equals(sourceDto.getType())) { |
| final RemoteProcessGroup remoteGroup = processGroup.getRemoteProcessGroup(sourceDto.getGroupId()); |
| source = remoteGroup.getOutputPort(sourceDto.getId()); |
| } else { |
| final ProcessGroup sourceGroup = controller.getGroup(sourceDto.getGroupId()); |
| if (sourceGroup == null) { |
| throw new RuntimeException("Found Invalid ProcessGroup ID for Source: " + dto.getSource().getGroupId()); |
| } |
| |
| source = sourceGroup.getConnectable(sourceDto.getId()); |
| } |
| if (source == null) { |
| throw new RuntimeException("Found Invalid Connectable ID for Source: " + dto.getSource().getId()); |
| } |
| |
| final Connectable destination; |
| final ConnectableDTO destinationDto = dto.getDestination(); |
| if (ConnectableType.REMOTE_INPUT_PORT.name().equals(destinationDto.getType())) { |
| final RemoteProcessGroup remoteGroup = processGroup.getRemoteProcessGroup(destinationDto.getGroupId()); |
| destination = remoteGroup.getInputPort(destinationDto.getId()); |
| } else { |
| final ProcessGroup destinationGroup = controller.getGroup(destinationDto.getGroupId()); |
| if (destinationGroup == null) { |
| throw new RuntimeException("Found Invalid ProcessGroup ID for Destination: " + dto.getDestination().getGroupId()); |
| } |
| |
| destination = destinationGroup.getConnectable(destinationDto.getId()); |
| } |
| if (destination == null) { |
| throw new RuntimeException("Found Invalid Connectable ID for Destination: " + dto.getDestination().getId()); |
| } |
| |
| final Connection connection = controller.createConnection(dto.getId(), dto.getName(), source, destination, dto.getSelectedRelationships()); |
| connection.setProcessGroup(processGroup); |
| |
| final List<Position> bendPoints = new ArrayList<>(); |
| for (final PositionDTO bend : dto.getBends()) { |
| bendPoints.add(new Position(bend.getX(), bend.getY())); |
| } |
| connection.setBendPoints(bendPoints); |
| |
| final Long zIndex = dto.getzIndex(); |
| if (zIndex != null) { |
| connection.setZIndex(zIndex); |
| } |
| |
| if (dto.getLabelIndex() != null) { |
| connection.setLabelIndex(dto.getLabelIndex()); |
| } |
| |
| List<FlowFilePrioritizer> newPrioritizers = null; |
| final List<String> prioritizers = dto.getPrioritizers(); |
| if (prioritizers != null) { |
| final List<String> newPrioritizersClasses = new ArrayList<>(prioritizers); |
| newPrioritizers = new ArrayList<>(); |
| for (final String className : newPrioritizersClasses) { |
| try { |
| newPrioritizers.add(controller.createPrioritizer(className)); |
| } catch (final ClassNotFoundException | InstantiationException | IllegalAccessException e) { |
| throw new IllegalArgumentException("Unable to set prioritizer " + className + ": " + e); |
| } |
| } |
| } |
| if (newPrioritizers != null) { |
| connection.getFlowFileQueue().setPriorities(newPrioritizers); |
| } |
| |
| if (dto.getBackPressureObjectThreshold() != null) { |
| connection.getFlowFileQueue().setBackPressureObjectThreshold(dto.getBackPressureObjectThreshold()); |
| } |
| if (dto.getBackPressureDataSizeThreshold() != null) { |
| connection.getFlowFileQueue().setBackPressureDataSizeThreshold(dto.getBackPressureDataSizeThreshold()); |
| } |
| if (dto.getFlowFileExpiration() != null) { |
| connection.getFlowFileQueue().setFlowFileExpiration(dto.getFlowFileExpiration()); |
| } |
| |
| processGroup.addConnection(connection); |
| } |
| |
| final List<Element> templateNodeList = getChildrenByTagName(processGroupElement, "template"); |
| for (final Element templateNode : templateNodeList) { |
| final TemplateDTO templateDTO = TemplateUtils.parseDto(templateNode); |
| final Template template = new Template(templateDTO); |
| processGroup.addTemplate(template); |
| } |
| |
| return processGroup; |
| } |
| |
| /** |
| * If both authorizers are external authorizers, or if the both are internal authorizers with equal fingerprints, |
| * then an uniheritable result with no reason is returned to indicate nothing to do. |
| * |
| * If both are internal authorizers and the current authorizer is empty, then an inheritable result is returned. |
| * |
| * All other cases return uninheritable with a reason which indicates to throw an exception. |
| * |
| * @param existingFlow the existing DataFlow |
| * @param proposedFlow the proposed DataFlow |
| * @return the AuthorizerInheritability result |
| */ |
| public AuthorizerInheritability checkAuthorizerInheritability(final DataFlow existingFlow, final DataFlow proposedFlow) { |
| final byte[] existing = existingFlow.getAuthorizerFingerprint(); |
| final byte[] proposed = proposedFlow.getAuthorizerFingerprint(); |
| |
| // both are using external authorizers so nothing to inherit, but we don't want to throw an exception |
| if (existing == null && proposed == null) { |
| return AuthorizerInheritability.uninheritable(null); |
| } |
| |
| // current is external, but proposed is internal |
| if (existing == null && proposed != null) { |
| return AuthorizerInheritability.uninheritable( |
| "Current Authorizer is an external Authorizer, but proposed Authorizer is an internal Authorizer"); |
| } |
| |
| // current is internal, but proposed is external |
| if (existing != null && proposed == null) { |
| return AuthorizerInheritability.uninheritable( |
| "Current Authorizer is an internal Authorizer, but proposed Authorizer is an external Authorizer"); |
| } |
| |
| // both are internal, but not the same |
| if (!Arrays.equals(existing, proposed)) { |
| final byte[] emptyAuthBytes = AbstractPolicyBasedAuthorizer.EMPTY_FINGERPRINT.getBytes(StandardCharsets.UTF_8); |
| |
| // if current is empty then we can take all the proposed authorizations |
| // otherwise they are both internal authorizers and don't match so we can't proceed |
| if (Arrays.equals(emptyAuthBytes, existing)) { |
| return AuthorizerInheritability.inheritable(); |
| } else { |
| return AuthorizerInheritability.uninheritable( |
| "Proposed Authorizations do not match current Authorizations"); |
| } |
| } |
| |
| // both are internal and equal |
| return AuthorizerInheritability.uninheritable(null); |
| } |
| |
| /** |
| * Returns true if the given controller can inherit the proposed flow without orphaning flow files. |
| * |
| * @param existingFlow flow |
| * @param controller the running controller |
| * @param proposedFlow the flow to inherit |
| * |
| * @return null if the controller can inherit the specified flow, an explanation of why it cannot be inherited otherwise |
| * |
| * @throws FingerprintException if flow fingerprints could not be generated |
| */ |
| public String checkFlowInheritability(final DataFlow existingFlow, final DataFlow proposedFlow, final FlowController controller) throws FingerprintException { |
| if (existingFlow == null) { |
| return null; // no existing flow, so equivalent to proposed flow |
| } |
| |
| return checkFlowInheritability(existingFlow.getFlow(), proposedFlow.getFlow(), controller); |
| } |
| |
| private String checkFlowInheritability(final byte[] existingFlow, final byte[] proposedFlow, final FlowController controller) { |
| if (existingFlow == null) { |
| return null; // no existing flow, so equivalent to proposed flow |
| } |
| |
| // check if the Flow is inheritable |
| final FingerprintFactory fingerprintFactory = new FingerprintFactory(encryptor); |
| final String existingFlowFingerprintBeforeHash = fingerprintFactory.createFingerprint(existingFlow, controller); |
| if (existingFlowFingerprintBeforeHash.trim().isEmpty()) { |
| return null; // no existing flow, so equivalent to proposed flow |
| } |
| |
| if (proposedFlow == null || proposedFlow.length == 0) { |
| return "Proposed Flow was empty but Current Flow is not"; // existing flow is not empty and proposed flow is empty (we could orphan flowfiles) |
| } |
| |
| final String proposedFlowFingerprintBeforeHash = fingerprintFactory.createFingerprint(proposedFlow, controller); |
| if (proposedFlowFingerprintBeforeHash.trim().isEmpty()) { |
| return "Proposed Flow was empty but Current Flow is not"; // existing flow is not empty and proposed flow is empty (we could orphan flowfiles) |
| } |
| |
| final boolean inheritable = existingFlowFingerprintBeforeHash.equals(proposedFlowFingerprintBeforeHash); |
| if (!inheritable) { |
| return findFirstDiscrepancy(existingFlowFingerprintBeforeHash, proposedFlowFingerprintBeforeHash, "Flows"); |
| } |
| |
| return null; |
| } |
| |
| private String findFirstDiscrepancy(final String existing, final String proposed, final String comparisonDescription) { |
| final int shortestFileLength = Math.min(existing.length(), proposed.length()); |
| for (int i = 0; i < shortestFileLength; i++) { |
| if (existing.charAt(i) != proposed.charAt(i)) { |
| final String formattedExistingDelta = formatFlowDiscrepancy(existing, i, 100); |
| final String formattedProposedDelta = formatFlowDiscrepancy(proposed, i, 100); |
| return String.format("Found difference in %s:\nLocal Fingerprint: %s\nCluster Fingerprint: %s", comparisonDescription, formattedExistingDelta, formattedProposedDelta); |
| } |
| } |
| |
| // existing must startWith proposed or proposed must startWith existing |
| if (existing.length() > proposed.length()) { |
| final String formattedExistingDelta = existing.substring(proposed.length(), Math.min(existing.length(), proposed.length() + 200)); |
| return String.format("Found difference in %s:\nLocal Fingerprint contains additional configuration from Cluster Fingerprint: %s", comparisonDescription, formattedExistingDelta); |
| } else if (proposed.length() > existing.length()) { |
| final String formattedProposedDelta = proposed.substring(existing.length(), Math.min(proposed.length(), existing.length() + 200)); |
| return String.format("Found difference in %s:\nCluster Fingerprint contains additional configuration from Local Fingerprint: %s", comparisonDescription, formattedProposedDelta); |
| } |
| |
| return "Unable to find any discrepancies between fingerprints. Please contact the NiFi support team"; |
| } |
| |
| private byte[] toBytes(final FlowController flowController) throws FlowSerializationException { |
| final ByteArrayOutputStream result = new ByteArrayOutputStream(); |
| final StandardFlowSerializer flowSerializer = new StandardFlowSerializer(encryptor); |
| flowController.serialize(flowSerializer, result); |
| return result.toByteArray(); |
| } |
| |
| private static String getString(final Element element, final String childElementName) { |
| final List<Element> nodeList = getChildrenByTagName(element, childElementName); |
| if (nodeList == null || nodeList.isEmpty()) { |
| return ""; |
| } |
| final Element childElement = nodeList.get(0); |
| return childElement.getTextContent(); |
| } |
| |
| private static int getInt(final Element element, final String childElementName) { |
| return Integer.parseInt(getString(element, childElementName)); |
| } |
| |
| private static Integer getInteger(final Element element, final String childElementName) { |
| final String value = getString(element, childElementName); |
| return (value == null || value.trim().equals("") ? null : Integer.parseInt(value)); |
| } |
| |
| private static List<Element> getChildrenByTagName(final Element element, final String tagName) { |
| final List<Element> matches = new ArrayList<>(); |
| final NodeList nodeList = element.getChildNodes(); |
| for (int i = 0; i < nodeList.getLength(); i++) { |
| final Node node = nodeList.item(i); |
| if (!(node instanceof Element)) { |
| continue; |
| } |
| |
| final Element child = (Element) nodeList.item(i); |
| if (child.getNodeName().equals(tagName)) { |
| matches.add(child); |
| } |
| } |
| |
| return matches; |
| } |
| |
| private String formatFlowDiscrepancy(final String flowFingerprint, final int deltaIndex, final int deltaPad) { |
| return flowFingerprint.substring(Math.max(0, deltaIndex - deltaPad), Math.min(flowFingerprint.length(), deltaIndex + deltaPad)); |
| } |
| |
| /** |
| * Holder for the result of determining if a proposed Authorizer is inheritable. |
| */ |
| private static final class AuthorizerInheritability { |
| |
| private final boolean inheritable; |
| private final String reason; |
| |
| public AuthorizerInheritability(boolean inheritable, String reason) { |
| this.inheritable = inheritable; |
| this.reason = reason; |
| } |
| |
| public boolean isInheritable() { |
| return inheritable; |
| } |
| |
| public String getReason() { |
| return reason; |
| } |
| |
| public static AuthorizerInheritability uninheritable(String reason) { |
| return new AuthorizerInheritability(false, reason); |
| } |
| |
| public static AuthorizerInheritability inheritable() { |
| return new AuthorizerInheritability(true, null); |
| } |
| |
| } |
| |
| } |