blob: c06460d596d7e3984439df01fdf632ac0055af83 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.camel.component.flink;
import org.apache.camel.Consumer;
import org.apache.camel.Processor;
import org.apache.camel.Producer;
import org.apache.camel.spi.Metadata;
import org.apache.camel.spi.UriEndpoint;
import org.apache.camel.spi.UriParam;
import org.apache.camel.spi.UriPath;
import org.apache.camel.support.DefaultEndpoint;
import org.apache.flink.api.java.DataSet;
import org.apache.flink.streaming.api.datastream.DataStream;
/**
* The flink component can be used to send DataSet jobs to Apache Flink cluster.
*/
@UriEndpoint(firstVersion = "2.18.0", scheme = "flink", title = "Apache Flink", syntax = "flink:endpointType", producerOnly = true, label = "hadoop")
public class FlinkEndpoint extends DefaultEndpoint {
@UriPath
@Metadata(required = true)
private EndpointType endpointType;
@UriParam
private DataSet dataSet;
@UriParam
private DataSetCallback dataSetCallback;
@UriParam
private DataStream dataStream;
@UriParam
private DataStreamCallback dataStreamCallback;
@UriParam(defaultValue = "true")
private boolean collect = true;
public FlinkEndpoint(String endpointUri, FlinkComponent component, EndpointType endpointType) {
super(endpointUri, component);
this.endpointType = endpointType;
}
@Override
protected void doStart() throws Exception {
super.doStart();
if (dataSet == null) {
dataSet = getComponent().getDataSet();
}
if (dataSetCallback == null) {
dataSetCallback = getComponent().getDataSetCallback();
}
}
@Override
public Producer createProducer() throws Exception {
if (endpointType == EndpointType.dataset) {
return new DataSetFlinkProducer(this);
} else if (endpointType == EndpointType.datastream) {
return new DataStreamFlinkProducer(this);
} else {
return null;
}
}
@Override
public Consumer createConsumer(Processor processor) throws Exception {
throw new UnsupportedOperationException("Flink Component supports producer endpoints only.");
}
@Override
public FlinkComponent getComponent() {
return (FlinkComponent) super.getComponent();
}
/**
* Type of the endpoint (dataset, datastream).
*/
public void setEndpointType(EndpointType endpointType) {
this.endpointType = endpointType;
}
public DataSet getDataSet() {
return dataSet;
}
public DataStream getDataStream() {
return dataStream;
}
/**
* DataSet to compute against.
*/
public void setDataSet(DataSet ds) {
this.dataSet = ds;
}
/**
* DataStream to compute against.
*/
public void setDataStream(DataStream ds) {
this.dataStream = ds;
}
public DataSetCallback getDataSetCallback() {
return dataSetCallback;
}
public DataStreamCallback getDataStreamCallback() {
return dataStreamCallback;
}
/**
* Function performing action against a DataSet.
*/
public void setDataSetCallback(DataSetCallback dataSetCallback) {
this.dataSetCallback = dataSetCallback;
}
/**
* Function performing action against a DataStream.
*/
public void setDataStreamCallback(DataStreamCallback dataStreamCallback) {
this.dataStreamCallback = dataStreamCallback;
}
public boolean isCollect() {
return collect;
}
/**
* Indicates if results should be collected or counted.
*/
public void setCollect(boolean collect) {
this.collect = collect;
}
}