| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package org.apache.flink.fs.s3hadoop; |
| |
| import org.apache.flink.configuration.Configuration; |
| import org.apache.flink.configuration.CoreOptions; |
| import org.apache.flink.core.fs.FSDataInputStream; |
| import org.apache.flink.core.fs.FileSystem; |
| import org.apache.flink.core.fs.Path; |
| import org.apache.flink.core.fs.RecoverableFsDataOutputStream; |
| import org.apache.flink.core.fs.RecoverableWriter; |
| import org.apache.flink.core.io.SimpleVersionedSerializer; |
| import org.apache.flink.fs.s3.common.FlinkS3FileSystem; |
| import org.apache.flink.fs.s3.common.writer.S3Recoverable; |
| import org.apache.flink.testutils.s3.S3TestCredentials; |
| import org.apache.flink.util.MathUtils; |
| import org.apache.flink.util.StringUtils; |
| import org.apache.flink.util.TestLogger; |
| |
| import org.junit.After; |
| import org.junit.AfterClass; |
| import org.junit.Assert; |
| import org.junit.Before; |
| import org.junit.BeforeClass; |
| import org.junit.ClassRule; |
| import org.junit.Test; |
| import org.junit.rules.TemporaryFolder; |
| |
| import java.io.BufferedReader; |
| import java.io.FileNotFoundException; |
| import java.io.IOException; |
| import java.io.InputStreamReader; |
| import java.nio.charset.StandardCharsets; |
| import java.nio.file.Files; |
| import java.nio.file.Paths; |
| import java.util.HashMap; |
| import java.util.Map; |
| import java.util.Random; |
| import java.util.UUID; |
| import java.util.stream.Stream; |
| |
| import static org.apache.flink.fs.s3.common.AbstractS3FileSystemFactory.MAX_CONCURRENT_UPLOADS; |
| import static org.apache.flink.fs.s3.common.AbstractS3FileSystemFactory.PART_UPLOAD_MIN_SIZE; |
| |
| /** |
| * Tests for the {@link org.apache.flink.fs.s3.common.writer.S3RecoverableWriter S3RecoverableWriter}. |
| */ |
| public class HadoopS3RecoverableWriterITCase extends TestLogger { |
| |
| // ----------------------- S3 general configuration ----------------------- |
| |
| private static final long PART_UPLOAD_MIN_SIZE_VALUE = 7L << 20; |
| private static final int MAX_CONCURRENT_UPLOADS_VALUE = 2; |
| |
| // ----------------------- Test Specific configuration ----------------------- |
| |
| private static final Random RND = new Random(); |
| |
| private static Path basePath; |
| |
| private static FlinkS3FileSystem fileSystem; |
| |
| // this is set for every test @Before |
| private Path basePathForTest; |
| |
| // ----------------------- Test Data to be used ----------------------- |
| |
| private static final String testData1 = "THIS IS A TEST 1."; |
| private static final String testData2 = "THIS IS A TEST 2."; |
| private static final String testData3 = "THIS IS A TEST 3."; |
| |
| private static final String bigDataChunk = createBigDataChunk(testData1, PART_UPLOAD_MIN_SIZE_VALUE); |
| |
| // ----------------------- Test Lifecycle ----------------------- |
| |
| private static boolean skipped = true; |
| |
| @ClassRule |
| public static final TemporaryFolder TEMP_FOLDER = new TemporaryFolder(); |
| |
| @BeforeClass |
| public static void checkCredentialsAndSetup() throws IOException { |
| // check whether credentials exist |
| S3TestCredentials.assumeCredentialsAvailable(); |
| |
| basePath = new Path(S3TestCredentials.getTestBucketUri() + "tests-" + UUID.randomUUID()); |
| |
| // initialize configuration with valid credentials |
| final Configuration conf = new Configuration(); |
| conf.setString("s3.access.key", S3TestCredentials.getS3AccessKey()); |
| conf.setString("s3.secret.key", S3TestCredentials.getS3SecretKey()); |
| |
| conf.setLong(PART_UPLOAD_MIN_SIZE, PART_UPLOAD_MIN_SIZE_VALUE); |
| conf.setInteger(MAX_CONCURRENT_UPLOADS, MAX_CONCURRENT_UPLOADS_VALUE); |
| |
| final String defaultTmpDir = TEMP_FOLDER.getRoot().getAbsolutePath() + "s3_tmp_dir"; |
| conf.setString(CoreOptions.TMP_DIRS, defaultTmpDir); |
| |
| FileSystem.initialize(conf); |
| |
| skipped = false; |
| } |
| |
| @AfterClass |
| public static void cleanUp() throws Exception { |
| if (!skipped) { |
| getFileSystem().delete(basePath, true); |
| } |
| FileSystem.initialize(new Configuration()); |
| } |
| |
| @Before |
| public void prepare() throws Exception { |
| basePathForTest = new Path( |
| basePath, |
| StringUtils.getRandomString(RND, 16, 16, 'a', 'z')); |
| |
| cleanupLocalDir(); |
| } |
| |
| private void cleanupLocalDir() throws Exception { |
| final String defaultTmpDir = getFileSystem().getLocalTmpDir(); |
| final java.nio.file.Path defaultTmpPath = Paths.get(defaultTmpDir); |
| |
| if (Files.exists(defaultTmpPath)) { |
| try (Stream<java.nio.file.Path> files = Files.list(defaultTmpPath)) { |
| files.forEach(p -> { |
| try { |
| Files.delete(p); |
| } catch (IOException e) { |
| e.printStackTrace(); |
| } |
| }); |
| } |
| } else { |
| Files.createDirectory(defaultTmpPath); |
| } |
| } |
| |
| @After |
| public void cleanupAndCheckTmpCleanup() throws Exception { |
| final String defaultTmpDir = getFileSystem().getLocalTmpDir(); |
| final java.nio.file.Path localTmpDir = Paths.get(defaultTmpDir); |
| |
| // delete local tmp dir. |
| Assert.assertTrue(Files.exists(localTmpDir)); |
| try (Stream<java.nio.file.Path> files = Files.list(localTmpDir)) { |
| Assert.assertEquals(0L, files.count()); |
| } |
| Files.delete(localTmpDir); |
| |
| // delete also S3 dir. |
| getFileSystem().delete(basePathForTest, true); |
| } |
| |
| private static FlinkS3FileSystem getFileSystem() throws Exception { |
| if (fileSystem == null) { |
| fileSystem = (FlinkS3FileSystem) FileSystem.get(basePath.toUri()); |
| } |
| return fileSystem; |
| } |
| |
| // ----------------------- Test Normal Execution ----------------------- |
| |
| @Test |
| public void testCloseWithNoData() throws Exception { |
| final RecoverableWriter writer = getRecoverableWriter(); |
| final Path path = new Path(basePathForTest, "part-0"); |
| |
| final RecoverableFsDataOutputStream stream = writer.open(path); |
| |
| stream.closeForCommit().commit(); |
| } |
| |
| @Test |
| public void testCommitAfterNormalClose() throws Exception { |
| final RecoverableWriter writer = getRecoverableWriter(); |
| final Path path = new Path(basePathForTest, "part-0"); |
| |
| final RecoverableFsDataOutputStream stream = writer.open(path); |
| stream.write(bytesOf(testData1)); |
| stream.closeForCommit().commit(); |
| |
| Assert.assertEquals(testData1, getContentsOfFile(path)); |
| } |
| |
| @Test |
| public void testCommitAfterPersist() throws Exception { |
| final RecoverableWriter writer = getRecoverableWriter(); |
| final Path path = new Path(basePathForTest, "part-0"); |
| |
| final RecoverableFsDataOutputStream stream = writer.open(path); |
| stream.write(bytesOf(testData1)); |
| stream.persist(); |
| |
| stream.write(bytesOf(testData2)); |
| stream.closeForCommit().commit(); |
| |
| Assert.assertEquals(testData1 + testData2, getContentsOfFile(path)); |
| } |
| |
| @Test(expected = FileNotFoundException.class) |
| public void testCleanupRecoverableState() throws Exception { |
| final RecoverableWriter writer = getRecoverableWriter(); |
| final Path path = new Path(basePathForTest, "part-0"); |
| |
| final RecoverableFsDataOutputStream stream = writer.open(path); |
| stream.write(bytesOf(testData1)); |
| S3Recoverable recoverable = (S3Recoverable) stream.persist(); |
| |
| stream.closeForCommit().commit(); |
| |
| // still the data is there as we have not deleted them from the tmp object |
| final String content = getContentsOfFile(new Path('/' + recoverable.incompleteObjectName())); |
| Assert.assertEquals(testData1, content); |
| |
| boolean successfullyDeletedState = writer.cleanupRecoverableState(recoverable); |
| Assert.assertTrue(successfullyDeletedState); |
| |
| // this should throw the exception as we deleted the file. |
| getContentsOfFile(new Path('/' + recoverable.incompleteObjectName())); |
| } |
| |
| @Test |
| public void testCallingDeleteObjectTwiceDoesNotThroughException() throws Exception { |
| final RecoverableWriter writer = getRecoverableWriter(); |
| final Path path = new Path(basePathForTest, "part-0"); |
| |
| final RecoverableFsDataOutputStream stream = writer.open(path); |
| stream.write(bytesOf(testData1)); |
| S3Recoverable recoverable = (S3Recoverable) stream.persist(); |
| |
| stream.closeForCommit().commit(); |
| |
| // still the data is there as we have not deleted them from the tmp object |
| final String content = getContentsOfFile(new Path('/' + recoverable.incompleteObjectName())); |
| Assert.assertEquals(testData1, content); |
| |
| boolean successfullyDeletedState = writer.cleanupRecoverableState(recoverable); |
| Assert.assertTrue(successfullyDeletedState); |
| |
| boolean unsuccessfulDeletion = writer.cleanupRecoverableState(recoverable); |
| Assert.assertFalse(unsuccessfulDeletion); |
| } |
| |
| // ----------------------- Test Recovery ----------------------- |
| |
| @Test |
| public void testCommitAfterRecovery() throws Exception { |
| final Path path = new Path(basePathForTest, "part-0"); |
| |
| final RecoverableWriter initWriter = getRecoverableWriter(); |
| |
| final RecoverableFsDataOutputStream stream = initWriter.open(path); |
| stream.write(bytesOf(testData1)); |
| |
| stream.persist(); |
| stream.persist(); |
| |
| // and write some more data |
| stream.write(bytesOf(testData2)); |
| |
| final RecoverableWriter.CommitRecoverable recoverable = stream.closeForCommit().getRecoverable(); |
| |
| final byte[] serializedRecoverable = initWriter.getCommitRecoverableSerializer().serialize(recoverable); |
| |
| // get a new serializer from a new writer to make sure that no pre-initialized state leaks in. |
| final RecoverableWriter newWriter = getRecoverableWriter(); |
| |
| final SimpleVersionedSerializer<RecoverableWriter.CommitRecoverable> deserializer = newWriter.getCommitRecoverableSerializer(); |
| final RecoverableWriter.CommitRecoverable recoveredRecoverable = deserializer.deserialize(deserializer.getVersion(), serializedRecoverable); |
| |
| final RecoverableFsDataOutputStream.Committer committer = newWriter.recoverForCommit(recoveredRecoverable); |
| committer.commitAfterRecovery(); |
| |
| Assert.assertEquals(testData1 + testData2, getContentsOfFile(path)); |
| } |
| |
| private static final String INIT_EMPTY_PERSIST = "EMPTY"; |
| private static final String INTERM_WITH_STATE_PERSIST = "INTERM-STATE"; |
| private static final String INTERM_WITH_NO_ADDITIONAL_STATE_PERSIST = "INTERM-IMEDIATE"; |
| private static final String FINAL_WITH_EXTRA_STATE = "FINAL"; |
| |
| @Test |
| public void testRecoverWithEmptyState() throws Exception { |
| testResumeAfterMultiplePersistWithSmallData( |
| INIT_EMPTY_PERSIST, |
| testData3); |
| } |
| |
| @Test |
| public void testRecoverWithState() throws Exception { |
| testResumeAfterMultiplePersistWithSmallData( |
| INTERM_WITH_STATE_PERSIST, |
| testData1 + testData3); |
| } |
| |
| @Test |
| public void testRecoverFromIntermWithoutAdditionalState() throws Exception { |
| testResumeAfterMultiplePersistWithSmallData( |
| INTERM_WITH_NO_ADDITIONAL_STATE_PERSIST, |
| testData1 + testData3); |
| } |
| |
| @Test |
| public void testRecoverAfterMultiplePersistsState() throws Exception { |
| testResumeAfterMultiplePersistWithSmallData( |
| FINAL_WITH_EXTRA_STATE, |
| testData1 + testData2 + testData3); |
| } |
| |
| @Test |
| public void testRecoverWithStateWithMultiPart() throws Exception { |
| testResumeAfterMultiplePersistWithMultiPartUploads( |
| INTERM_WITH_STATE_PERSIST, |
| bigDataChunk + bigDataChunk); |
| } |
| |
| @Test |
| public void testRecoverFromIntermWithoutAdditionalStateWithMultiPart() throws Exception { |
| testResumeAfterMultiplePersistWithMultiPartUploads( |
| INTERM_WITH_NO_ADDITIONAL_STATE_PERSIST, |
| bigDataChunk + bigDataChunk); |
| } |
| |
| @Test |
| public void testRecoverAfterMultiplePersistsStateWithMultiPart() throws Exception { |
| testResumeAfterMultiplePersistWithMultiPartUploads( |
| FINAL_WITH_EXTRA_STATE, |
| bigDataChunk + bigDataChunk + bigDataChunk); |
| } |
| |
| private void testResumeAfterMultiplePersistWithSmallData( |
| final String persistName, |
| final String expectedFinalContents) throws Exception { |
| testResumeAfterMultiplePersist( |
| persistName, |
| expectedFinalContents, |
| testData1, |
| testData2, |
| testData3 |
| ); |
| } |
| |
| private void testResumeAfterMultiplePersistWithMultiPartUploads( |
| final String persistName, |
| final String expectedFinalContents) throws Exception { |
| testResumeAfterMultiplePersist( |
| persistName, |
| expectedFinalContents, |
| bigDataChunk, |
| bigDataChunk, |
| bigDataChunk |
| ); |
| } |
| |
| private void testResumeAfterMultiplePersist( |
| final String persistName, |
| final String expectedFinalContents, |
| final String firstItemToWrite, |
| final String secondItemToWrite, |
| final String thirdItemToWrite) throws Exception { |
| |
| final Path path = new Path(basePathForTest, "part-0"); |
| final RecoverableWriter initWriter = getRecoverableWriter(); |
| |
| final Map<String, RecoverableWriter.ResumeRecoverable> recoverables = new HashMap<>(4); |
| try (final RecoverableFsDataOutputStream stream = initWriter.open(path)) { |
| recoverables.put(INIT_EMPTY_PERSIST, stream.persist()); |
| |
| stream.write(bytesOf(firstItemToWrite)); |
| |
| recoverables.put(INTERM_WITH_STATE_PERSIST, stream.persist()); |
| recoverables.put(INTERM_WITH_NO_ADDITIONAL_STATE_PERSIST, stream.persist()); |
| |
| // and write some more data |
| stream.write(bytesOf(secondItemToWrite)); |
| |
| recoverables.put(FINAL_WITH_EXTRA_STATE, stream.persist()); |
| } |
| |
| final SimpleVersionedSerializer<RecoverableWriter.ResumeRecoverable> serializer = initWriter.getResumeRecoverableSerializer(); |
| final byte[] serializedRecoverable = serializer.serialize(recoverables.get(persistName)); |
| |
| // get a new serializer from a new writer to make sure that no pre-initialized state leaks in. |
| final RecoverableWriter newWriter = getRecoverableWriter(); |
| final SimpleVersionedSerializer<RecoverableWriter.ResumeRecoverable> deserializer = newWriter.getResumeRecoverableSerializer(); |
| final RecoverableWriter.ResumeRecoverable recoveredRecoverable = |
| deserializer.deserialize(serializer.getVersion(), serializedRecoverable); |
| |
| final RecoverableFsDataOutputStream recoveredStream = newWriter.recover(recoveredRecoverable); |
| recoveredStream.write(bytesOf(thirdItemToWrite)); |
| recoveredStream.closeForCommit().commit(); |
| |
| Assert.assertEquals(expectedFinalContents, getContentsOfFile(path)); |
| } |
| |
| // -------------------------- Test Utilities -------------------------- |
| |
| private String getContentsOfFile(Path path) throws Exception { |
| final StringBuilder builder = new StringBuilder(); |
| try ( |
| FSDataInputStream inStream = getFileSystem().open(path); |
| BufferedReader reader = new BufferedReader(new InputStreamReader(inStream)) |
| ) { |
| String line; |
| while ((line = reader.readLine()) != null) { |
| builder.append(line); |
| } |
| } |
| return builder.toString(); |
| } |
| |
| // ----------------------- Test utilities ----------------------- |
| |
| private static String createBigDataChunk(String pattern, long size) { |
| final StringBuilder stringBuilder = new StringBuilder(); |
| |
| int sampleLength = bytesOf(pattern).length; |
| int repeats = MathUtils.checkedDownCast(size) / sampleLength + 100; |
| |
| for (int i = 0; i < repeats; i++) { |
| stringBuilder.append(pattern); |
| } |
| return stringBuilder.toString(); |
| } |
| |
| private static byte[] bytesOf(String str) { |
| return str.getBytes(StandardCharsets.UTF_8); |
| } |
| |
| private RecoverableWriter getRecoverableWriter() throws Exception { |
| return getFileSystem().createRecoverableWriter(); |
| } |
| } |