blob: ead2c44fe0884f82e50b3b483dc08756ddc9e3f0 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.flink.connector.jdbc;
import org.apache.flink.api.common.restartstrategy.RestartStrategies;
import org.apache.flink.configuration.Configuration;
import org.apache.flink.connector.jdbc.testutils.JdbcITCaseBase;
import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
import org.apache.flink.util.function.FunctionWithException;
import org.junit.jupiter.api.Test;
import java.io.Serializable;
import java.sql.Connection;
import java.sql.DriverManager;
import java.sql.ResultSet;
import java.sql.SQLException;
import java.sql.Statement;
import java.sql.Types;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.List;
import java.util.concurrent.atomic.AtomicInteger;
import static org.apache.flink.configuration.PipelineOptions.OBJECT_REUSE;
import static org.apache.flink.connector.jdbc.JdbcConnectionOptions.JdbcConnectionOptionsBuilder;
import static org.apache.flink.connector.jdbc.JdbcTestFixture.INPUT_TABLE;
import static org.apache.flink.connector.jdbc.JdbcTestFixture.INSERT_TEMPLATE;
import static org.apache.flink.connector.jdbc.JdbcTestFixture.TEST_DATA;
import static org.apache.flink.connector.jdbc.JdbcTestFixture.TestEntry;
import static org.apache.flink.util.Preconditions.checkNotNull;
import static org.assertj.core.api.Assertions.assertThat;
/** Smoke tests for the {@link JdbcSink} and the underlying classes. */
public class JdbcITCase extends JdbcTestBase implements JdbcITCaseBase {
public static final JdbcStatementBuilder<TestEntry> TEST_ENTRY_JDBC_STATEMENT_BUILDER =
(ps, t) -> {
ps.setInt(1, t.id);
ps.setString(2, t.title);
ps.setString(3, t.author);
if (t.price == null) {
ps.setNull(4, Types.DOUBLE);
} else {
ps.setDouble(4, t.price);
}
ps.setInt(5, t.qty);
};
@Test
void testInsert() throws Exception {
StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
env.setRestartStrategy(new RestartStrategies.NoRestartStrategyConfiguration());
env.setParallelism(1);
env.fromElements(TEST_DATA)
.addSink(
JdbcSink.sink(
String.format(INSERT_TEMPLATE, INPUT_TABLE),
TEST_ENTRY_JDBC_STATEMENT_BUILDER,
new JdbcConnectionOptionsBuilder()
.withUrl(getMetadata().getJdbcUrl())
.withDriverName(getMetadata().getDriverClass())
.build()));
env.execute();
assertThat(selectBooks()).isEqualTo(Arrays.asList(TEST_DATA));
}
@Test
void testObjectReuse() throws Exception {
Configuration configuration = new Configuration();
configuration.set(OBJECT_REUSE, true);
StreamExecutionEnvironment env =
StreamExecutionEnvironment.getExecutionEnvironment(configuration);
env.setRestartStrategy(new RestartStrategies.NoRestartStrategyConfiguration());
env.setParallelism(1);
AtomicInteger counter = new AtomicInteger(0);
String[] words = {"a", "and", "b", "were", "sitting in the buffer"};
StringHolder reused = new StringHolder();
env.fromElements(words)
.map(
word -> {
reused.setContent(word);
return reused;
})
.addSink(
JdbcSink.sink(
JdbcTestFixture.INSERT_INTO_WORDS_TEMPLATE,
(ps, e) -> {
ps.setInt(1, counter.getAndIncrement());
ps.setString(2, e.content);
},
new JdbcConnectionOptionsBuilder()
.withUrl(getMetadata().getJdbcUrl())
.withDriverName(getMetadata().getDriverClass())
.build()));
env.execute();
assertThat(selectWords()).isEqualTo(Arrays.asList(words));
}
private List<String> selectWords() throws SQLException {
ArrayList<String> strings = new ArrayList<>();
try (Connection connection = DriverManager.getConnection(getMetadata().getJdbcUrl())) {
try (Statement st = connection.createStatement()) {
try (ResultSet rs = st.executeQuery("select word from words")) {
while (rs.next()) {
strings.add(rs.getString(1));
}
}
}
}
return strings;
}
private static class StringHolder implements Serializable {
private static final long serialVersionUID = 1L;
private String content;
public String getContent() {
return checkNotNull(content);
}
public void setContent(String payload) {
this.content = checkNotNull(payload);
}
}
private List<TestEntry> selectBooks() throws SQLException {
List<TestEntry> result = new ArrayList<>();
try (Connection connection = DriverManager.getConnection(getMetadata().getJdbcUrl())) {
connection.setTransactionIsolation(Connection.TRANSACTION_READ_COMMITTED);
connection.setReadOnly(true);
try (Statement st = connection.createStatement()) {
try (ResultSet rs =
st.executeQuery(
"select id, title, author, price, qty from " + INPUT_TABLE)) {
while (rs.next()) {
result.add(
new TestEntry(
getNullable(rs, r -> r.getInt(1)),
getNullable(rs, r -> r.getString(2)),
getNullable(rs, r -> r.getString(3)),
getNullable(rs, r -> r.getDouble(4)),
getNullable(rs, r -> r.getInt(5))));
}
}
}
}
return result;
}
private static <T> T getNullable(
ResultSet rs, FunctionWithException<ResultSet, T, SQLException> get)
throws SQLException {
T value = get.apply(rs);
return rs.wasNull() ? null : value;
}
}