blob: 241cad340f41ff0932ffaf21f7b8d33578bea450 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.uniffle.storage.handler.impl;
import java.nio.ByteBuffer;
import java.util.LinkedList;
import java.util.List;
import java.util.Random;
import java.util.Set;
import com.google.common.collect.Lists;
import com.google.common.collect.Sets;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.junit.jupiter.api.Test;
import org.apache.uniffle.common.BufferSegment;
import org.apache.uniffle.common.ShuffleDataResult;
import org.apache.uniffle.common.ShufflePartitionedBlock;
import org.apache.uniffle.common.util.BlockIdSet;
import org.apache.uniffle.storage.HadoopTestBase;
import org.apache.uniffle.storage.common.FileBasedShuffleSegment;
import static org.junit.jupiter.api.Assertions.assertTrue;
public class HadoopHandlerTest extends HadoopTestBase {
@Test
public void initTest() throws Exception {
String basePath = HDFS_URI + "test_base";
new HadoopShuffleWriteHandler("appId", 0, 0, 0, basePath, "test", conf);
Path path = new Path(basePath);
assertTrue(fs.isDirectory(path));
}
@Test
public void writeTest() throws Exception {
String basePath = HDFS_URI + "writeTest";
HadoopShuffleWriteHandler writeHandler =
new HadoopShuffleWriteHandler("appId", 1, 1, 1, basePath, "test", conf);
List<ShufflePartitionedBlock> blocks = new LinkedList<>();
List<Long> expectedBlockId = new LinkedList<>();
List<byte[]> expectedData = new LinkedList<>();
List<FileBasedShuffleSegment> expectedIndex = new LinkedList<>();
int pos = 0;
for (int i = 1; i < 13; ++i) {
byte[] buf = new byte[i * 8];
new Random().nextBytes(buf);
expectedData.add(buf);
blocks.add(new ShufflePartitionedBlock(i * 8, i * 8, i, i, 0, buf));
expectedBlockId.add(Long.valueOf(i));
expectedIndex.add(new FileBasedShuffleSegment(i, pos, i * 8, i * 8, i, 0));
pos += i * 8;
}
writeHandler.write(blocks);
compareDataAndIndex("appId", 1, 1, basePath, expectedData, expectedBlockId);
// append the exist data and index files
List<ShufflePartitionedBlock> blocksAppend = new LinkedList<>();
for (int i = 13; i < 23; ++i) {
byte[] buf = new byte[i * 8];
new Random().nextBytes(buf);
expectedData.add(buf);
expectedBlockId.add(Long.valueOf(i));
blocksAppend.add(new ShufflePartitionedBlock(i * 8, i * 8, i, i, i, buf));
expectedIndex.add(new FileBasedShuffleSegment(i, pos, i * 8, i * 8, i, i));
pos += i * 8;
}
writeHandler = new HadoopShuffleWriteHandler("appId", 1, 1, 1, basePath, "test", conf);
writeHandler.write(blocksAppend);
compareDataAndIndex("appId", 1, 1, basePath, expectedData, expectedBlockId);
}
private void compareDataAndIndex(
String appId,
int shuffleId,
int partitionId,
String basePath,
List<byte[]> expectedData,
List<Long> expectedBlockId)
throws IllegalStateException {
BlockIdSet expectBlockIds = BlockIdSet.empty();
BlockIdSet processBlockIds = BlockIdSet.empty();
for (long blockId : expectedBlockId) {
expectBlockIds.add(blockId);
}
// read directly and compare
HadoopClientReadHandler readHandler =
new HadoopClientReadHandler(
appId,
shuffleId,
partitionId,
100,
1,
10,
10000,
expectBlockIds,
processBlockIds,
basePath,
new Configuration());
try {
List<ByteBuffer> actual = readData(readHandler, Sets.newHashSet(expectedBlockId));
compareBytes(expectedData, actual);
} finally {
readHandler.close();
}
}
private List<ByteBuffer> readData(HadoopClientReadHandler handler, Set<Long> blockIds)
throws IllegalStateException {
ShuffleDataResult sdr = handler.readShuffleData();
List<BufferSegment> bufferSegments = sdr.getBufferSegments();
List<ByteBuffer> result = Lists.newArrayList();
for (BufferSegment bs : bufferSegments) {
byte[] data = new byte[bs.getLength()];
System.arraycopy(sdr.getData(), bs.getOffset(), data, 0, bs.getLength());
result.add(ByteBuffer.wrap(data));
}
return result;
}
}