| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * <p> |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * <p> |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.crunch.kafka.offset.hdfs; |
| |
| import org.apache.crunch.kafka.offset.OffsetReader; |
| import org.apache.crunch.kafka.offset.OffsetWriter; |
| import org.apache.hadoop.conf.Configuration; |
| import org.apache.hadoop.fs.FileSystem; |
| import org.apache.hadoop.fs.Path; |
| import org.apache.kafka.common.TopicPartition; |
| import org.junit.After; |
| import org.junit.Before; |
| import org.junit.Rule; |
| import org.junit.Test; |
| import org.junit.rules.TemporaryFolder; |
| import org.junit.rules.TestName; |
| |
| import java.io.IOException; |
| import java.util.Collections; |
| import java.util.HashMap; |
| import java.util.LinkedList; |
| import java.util.List; |
| import java.util.Map; |
| |
| import static org.hamcrest.CoreMatchers.nullValue; |
| import static org.hamcrest.core.Is.is; |
| import static org.junit.Assert.assertThat; |
| |
| public class HDFSOffsetReaderTest { |
| |
| @Rule |
| public TemporaryFolder tempFolder = new TemporaryFolder(); |
| |
| @Rule |
| public TestName testName = new TestName(); |
| |
| private Path basePath; |
| private FileSystem fileSystem; |
| private OffsetWriter writer; |
| private OffsetReader reader; |
| |
| |
| @Before |
| public void setup() throws IOException { |
| Configuration config = new Configuration(); |
| config.set(FileSystem.DEFAULT_FS, tempFolder.newFolder().getAbsolutePath()); |
| |
| fileSystem = FileSystem.newInstance(config); |
| basePath = new Path(tempFolder.newFolder().toString(), testName.getMethodName()); |
| |
| writer = new HDFSOffsetWriter(config, basePath); |
| |
| reader = new HDFSOffsetReader(config, basePath); |
| } |
| |
| @After |
| public void cleanup() throws IOException { |
| writer.close(); |
| reader.close(); |
| fileSystem.close(); |
| } |
| |
| @Test(expected = IllegalArgumentException.class) |
| public void constructNullConfig() { |
| new HDFSOffsetReader(null, new Path("/")); |
| } |
| |
| @Test(expected = IllegalArgumentException.class) |
| public void constructNullPath() { |
| new HDFSOffsetReader(new Configuration(), null); |
| } |
| |
| @Test |
| public void getStoredOffsetPersistenceTimesNoValues() throws IOException { |
| List<Long> storedOffsetPersistenceTimes = reader.getStoredOffsetPersistenceTimes(); |
| assertThat(storedOffsetPersistenceTimes, is(Collections.<Long>emptyList())); |
| } |
| |
| @Test |
| public void getStoredOffsetPersistenceTimesMultipleValues() throws IOException { |
| long current = 1464992662000L; |
| List<Long> persistedTimes = new LinkedList<>(); |
| for (int i = 0; i < 10; i++) { |
| persistedTimes.add(current + (i * 18000)); |
| } |
| |
| for (Long t : persistedTimes) { |
| try { |
| writer.write(t, Collections.<TopicPartition, Long>emptyMap()); |
| } catch (IOException e) { |
| e.printStackTrace(); |
| } |
| } |
| |
| List<Long> storedTimes = reader.getStoredOffsetPersistenceTimes(); |
| |
| assertThat(storedTimes, is(persistedTimes)); |
| } |
| |
| @Test |
| public void readOffsetNoMatchForTime() throws IOException { |
| Map<TopicPartition, Long> offsets = reader.readOffsets(12345L); |
| assertThat(offsets, is(nullValue())); |
| } |
| |
| @Test |
| public void readOffsetLatestNone() throws IOException { |
| assertThat(reader.readLatestOffsets(), is(Collections.<TopicPartition, Long>emptyMap())); |
| } |
| |
| @Test |
| public void readOffsetLatest() throws IOException { |
| long current = 1464992662000L; |
| List<Long> persistedTimes = new LinkedList<>(); |
| for (int i = 0; i < 10; i++) { |
| persistedTimes.add(current + (i * 18000)); |
| } |
| |
| for (Long t : persistedTimes) { |
| try { |
| writer.write(t, Collections.<TopicPartition, Long>emptyMap()); |
| } catch (IOException e) { |
| e.printStackTrace(); |
| } |
| } |
| |
| long expectedTime = persistedTimes.get(persistedTimes.size() - 1); |
| |
| Map<TopicPartition, Long> offsets = new HashMap<>(); |
| for (int i = 0; i < 9; i++) { |
| for (int j = 0; j < 5; j++) { |
| offsets.put(new TopicPartition("topic" + i, j), (long) j); |
| } |
| } |
| |
| writer.write(expectedTime, offsets); |
| |
| Map<TopicPartition, Long> retrievedOffsets = reader.readLatestOffsets(); |
| |
| assertThat(retrievedOffsets, is(offsets)); |
| } |
| |
| |
| @Test |
| public void readOffsetForTime() throws IOException { |
| long current = 1464992662000L; |
| List<Long> persistedTimes = new LinkedList<>(); |
| for (int i = 0; i < 10; i++) { |
| persistedTimes.add(current + (i * 18000)); |
| } |
| for (Long t : persistedTimes) { |
| try { |
| writer.write(t, Collections.<TopicPartition, Long>emptyMap()); |
| } catch (IOException e) { |
| e.printStackTrace(); |
| } |
| } |
| |
| long expectedTime = persistedTimes.get(2); |
| |
| Map<TopicPartition, Long> offsets = new HashMap<>(); |
| for (int i = 0; i < 9; i++) { |
| for (int j = 0; j < 5; j++) { |
| offsets.put(new TopicPartition("topic" + i, j), (long) j); |
| } |
| } |
| |
| writer.write(expectedTime, offsets); |
| |
| Map<TopicPartition, Long> retrievedOffsets = reader.readOffsets(expectedTime); |
| |
| assertThat(retrievedOffsets, is(offsets)); |
| } |
| |
| |
| @Test |
| public void skipReadingDirectory() throws IOException { |
| long current = 1464992662000L; |
| List<Long> persistedTimes = new LinkedList<>(); |
| for (int i = 0; i < 10; i++) { |
| persistedTimes.add(current + (i * 18000)); |
| } |
| |
| for (Long t : persistedTimes) { |
| try { |
| writer.write(t, Collections.<TopicPartition, Long>emptyMap()); |
| } catch (IOException e) { |
| e.printStackTrace(); |
| } |
| } |
| fileSystem.mkdirs(new Path(basePath, "imadirectory")); |
| |
| List<Long> storedTimes = reader.getStoredOffsetPersistenceTimes(); |
| |
| assertThat(storedTimes, is(persistedTimes)); |
| } |
| |
| @Test |
| public void skipInvalidFile() throws IOException { |
| long current = 1464992662000L; |
| List<Long> persistedTimes = new LinkedList<>(); |
| for (int i = 0; i < 10; i++) { |
| persistedTimes.add(current + (i * 18000)); |
| } |
| |
| for (Long t : persistedTimes) { |
| try { |
| writer.write(t, Collections.<TopicPartition, Long>emptyMap()); |
| } catch (IOException e) { |
| e.printStackTrace(); |
| } |
| } |
| |
| fileSystem.createNewFile(new Path(basePath, "imabadfile.json")); |
| fileSystem.createNewFile(new Path(basePath, "imabadfile.txt")); |
| |
| List<Long> storedTimes = reader.getStoredOffsetPersistenceTimes(); |
| |
| assertThat(storedTimes, is(persistedTimes)); |
| } |
| } |