blob: 60ceb1de050e8541b842a73f31ce7bc69c9eaef4 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.apache.accumulo.test.replication;
import static org.easymock.EasyMock.createMock;
import static org.easymock.EasyMock.expectLastCall;
import static org.easymock.EasyMock.replay;
import static org.easymock.EasyMock.verify;
import java.util.HashSet;
import java.util.Set;
import java.util.UUID;
import org.apache.accumulo.core.client.Accumulo;
import org.apache.accumulo.core.client.AccumuloClient;
import org.apache.accumulo.core.client.BatchWriter;
import org.apache.accumulo.core.data.Mutation;
import org.apache.accumulo.core.data.TableId;
import org.apache.accumulo.core.protobuf.ProtobufUtil;
import org.apache.accumulo.core.replication.ReplicationSchema.OrderSection;
import org.apache.accumulo.core.replication.ReplicationSchema.WorkSection;
import org.apache.accumulo.core.replication.ReplicationTable;
import org.apache.accumulo.core.replication.ReplicationTarget;
import org.apache.accumulo.core.security.TablePermission;
import org.apache.accumulo.fate.zookeeper.ZooCache;
import org.apache.accumulo.manager.replication.UnorderedWorkAssigner;
import org.apache.accumulo.server.replication.DistributedWorkQueueWorkAssignerHelper;
import org.apache.accumulo.server.replication.StatusUtil;
import org.apache.accumulo.server.replication.proto.Replication.Status;
import org.apache.accumulo.server.zookeeper.DistributedWorkQueue;
import org.apache.accumulo.test.functional.ConfigurableMacBase;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.Text;
import org.junit.Before;
import org.junit.Ignore;
import org.junit.Test;
@Ignore("Replication ITs are not stable and not currently maintained")
public class UnorderedWorkAssignerIT extends ConfigurableMacBase {
private AccumuloClient client;
private MockUnorderedWorkAssigner assigner;
private static class MockUnorderedWorkAssigner extends UnorderedWorkAssigner {
public MockUnorderedWorkAssigner(AccumuloClient client) {
super(null, client);
}
@Override
protected void setQueuedWork(Set<String> queuedWork) {
super.setQueuedWork(queuedWork);
}
@Override
protected void setWorkQueue(DistributedWorkQueue workQueue) {
super.setWorkQueue(workQueue);
}
@Override
protected boolean queueWork(Path path, ReplicationTarget target) {
return super.queueWork(path, target);
}
@Override
protected void initializeQueuedWork() {
super.initializeQueuedWork();
}
@Override
protected Set<String> getQueuedWork() {
return super.getQueuedWork();
}
@Override
protected void setClient(AccumuloClient client) {
super.setClient(client);
}
@Override
protected void setMaxQueueSize(int maxQueueSize) {
super.setMaxQueueSize(maxQueueSize);
}
@Override
protected void createWork() {
super.createWork();
}
@Override
protected void setZooCache(ZooCache zooCache) {
super.setZooCache(zooCache);
}
@Override
protected void cleanupFinishedWork() {
super.cleanupFinishedWork();
}
}
@Before
public void init() throws Exception {
client = Accumulo.newClient().from(getClientProperties()).build();
assigner = new MockUnorderedWorkAssigner(client);
ReplicationTable.setOnline(client);
client.securityOperations().grantTablePermission(client.whoami(), ReplicationTable.NAME,
TablePermission.WRITE);
client.securityOperations().grantTablePermission(client.whoami(), ReplicationTable.NAME,
TablePermission.READ);
}
@Test
public void createWorkForFilesNeedingIt() throws Exception {
ReplicationTarget target1 = new ReplicationTarget("cluster1", "table1", TableId.of("1")),
target2 = new ReplicationTarget("cluster1", "table2", TableId.of("2"));
Text serializedTarget1 = target1.toText(), serializedTarget2 = target2.toText();
String keyTarget1 =
target1.getPeerName() + DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR
+ target1.getRemoteIdentifier() + DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR
+ target1.getSourceTableId(),
keyTarget2 = target2.getPeerName() + DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR
+ target2.getRemoteIdentifier() + DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR
+ target2.getSourceTableId();
Status.Builder builder = Status.newBuilder().setBegin(0).setEnd(0).setInfiniteEnd(true)
.setClosed(false).setCreatedTime(5L);
Status status1 = builder.build();
builder.setCreatedTime(10L);
Status status2 = builder.build();
// Create two mutations, both of which need replication work done
BatchWriter bw = ReplicationTable.getBatchWriter(client);
String filename1 = UUID.randomUUID().toString(), filename2 = UUID.randomUUID().toString();
String file1 = "/accumulo/wal/tserver+port/" + filename1,
file2 = "/accumulo/wal/tserver+port/" + filename2;
Mutation m = new Mutation(file1);
WorkSection.add(m, serializedTarget1, ProtobufUtil.toValue(status1));
bw.addMutation(m);
m = OrderSection.createMutation(file1, status1.getCreatedTime());
OrderSection.add(m, target1.getSourceTableId(), ProtobufUtil.toValue(status1));
bw.addMutation(m);
m = new Mutation(file2);
WorkSection.add(m, serializedTarget2, ProtobufUtil.toValue(status2));
bw.addMutation(m);
m = OrderSection.createMutation(file2, status2.getCreatedTime());
OrderSection.add(m, target2.getSourceTableId(), ProtobufUtil.toValue(status2));
bw.addMutation(m);
bw.close();
DistributedWorkQueue workQueue = createMock(DistributedWorkQueue.class);
HashSet<String> queuedWork = new HashSet<>();
assigner.setQueuedWork(queuedWork);
assigner.setWorkQueue(workQueue);
assigner.setMaxQueueSize(Integer.MAX_VALUE);
// Make sure we expect the invocations in the order they were created
String key = filename1 + "|" + keyTarget1;
workQueue.addWork(key, file1);
expectLastCall().once();
key = filename2 + "|" + keyTarget2;
workQueue.addWork(key, file2);
expectLastCall().once();
replay(workQueue);
assigner.createWork();
verify(workQueue);
}
@Test
public void doNotCreateWorkForFilesNotNeedingIt() throws Exception {
ReplicationTarget target1 = new ReplicationTarget("cluster1", "table1", TableId.of("1")),
target2 = new ReplicationTarget("cluster1", "table2", TableId.of("2"));
Text serializedTarget1 = target1.toText(), serializedTarget2 = target2.toText();
// Create two mutations, both of which need replication work done
BatchWriter bw = ReplicationTable.getBatchWriter(client);
String filename1 = UUID.randomUUID().toString(), filename2 = UUID.randomUUID().toString();
String file1 = "/accumulo/wal/tserver+port/" + filename1,
file2 = "/accumulo/wal/tserver+port/" + filename2;
Mutation m = new Mutation(file1);
WorkSection.add(m, serializedTarget1, StatusUtil.fileCreatedValue(5));
bw.addMutation(m);
m = new Mutation(file2);
WorkSection.add(m, serializedTarget2, StatusUtil.fileCreatedValue(10));
bw.addMutation(m);
bw.close();
DistributedWorkQueue workQueue = createMock(DistributedWorkQueue.class);
HashSet<String> queuedWork = new HashSet<>();
assigner.setQueuedWork(queuedWork);
assigner.setMaxQueueSize(Integer.MAX_VALUE);
replay(workQueue);
assigner.createWork();
verify(workQueue);
}
@Test
public void workNotReAdded() throws Exception {
Set<String> queuedWork = new HashSet<>();
assigner.setQueuedWork(queuedWork);
ReplicationTarget target = new ReplicationTarget("cluster1", "table1", TableId.of("1"));
String serializedTarget = target.getPeerName()
+ DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR + target.getRemoteIdentifier()
+ DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR + target.getSourceTableId();
queuedWork.add("wal1|" + serializedTarget);
// Create two mutations, both of which need replication work done
BatchWriter bw = ReplicationTable.getBatchWriter(client);
String file1 = "/accumulo/wal/tserver+port/wal1";
Mutation m = new Mutation(file1);
WorkSection.add(m, target.toText(), StatusUtil.openWithUnknownLengthValue());
bw.addMutation(m);
bw.close();
DistributedWorkQueue workQueue = createMock(DistributedWorkQueue.class);
assigner.setWorkQueue(workQueue);
assigner.setMaxQueueSize(Integer.MAX_VALUE);
replay(workQueue);
assigner.createWork();
verify(workQueue);
}
}