| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| */ |
| package org.apache.accumulo.manager.replication; |
| |
| import static org.apache.accumulo.fate.util.UtilWaitThread.sleepUninterruptibly; |
| |
| import java.util.HashSet; |
| import java.util.Iterator; |
| import java.util.Set; |
| import java.util.concurrent.TimeUnit; |
| |
| import org.apache.accumulo.core.client.AccumuloClient; |
| import org.apache.accumulo.core.conf.AccumuloConfiguration; |
| import org.apache.accumulo.core.replication.ReplicationConstants; |
| import org.apache.accumulo.core.replication.ReplicationTarget; |
| import org.apache.accumulo.fate.zookeeper.ZooUtil; |
| import org.apache.accumulo.server.replication.DistributedWorkQueueWorkAssignerHelper; |
| import org.apache.hadoop.fs.Path; |
| import org.apache.zookeeper.KeeperException; |
| import org.slf4j.Logger; |
| import org.slf4j.LoggerFactory; |
| |
| /** |
| * Read work records from the replication table, create work entries for other nodes to complete. |
| * <p> |
| * Uses the DistributedWorkQueue to make the work available for any tserver. This approach does not |
| * consider the locality of the tabletserver performing the work in relation to the data being |
| * replicated (local HDFS blocks). |
| * <p> |
| * The implementation allows for multiple tservers to concurrently replicate data to peer(s), |
| * however it is possible that data for a table is replayed on the peer in a different order than |
| * the manager. The {@link SequentialWorkAssigner} should be used if this must be guaranteed at the |
| * cost of replication throughput. |
| */ |
| public class UnorderedWorkAssigner extends DistributedWorkQueueWorkAssigner { |
| private static final Logger log = LoggerFactory.getLogger(UnorderedWorkAssigner.class); |
| private static final String NAME = "Unordered Work Assigner"; |
| |
| private Set<String> queuedWork; |
| |
| public UnorderedWorkAssigner() {} |
| |
| public UnorderedWorkAssigner(AccumuloConfiguration conf, AccumuloClient client) { |
| configure(conf, client); |
| } |
| |
| @Override |
| public String getName() { |
| return NAME; |
| } |
| |
| protected Set<String> getQueuedWork() { |
| return queuedWork; |
| } |
| |
| protected void setQueuedWork(Set<String> queuedWork) { |
| this.queuedWork = queuedWork; |
| } |
| |
| /** |
| * Initialize the queuedWork set with the work already sent out |
| */ |
| @Override |
| protected void initializeQueuedWork() { |
| if (queuedWork != null) { |
| return; |
| } |
| |
| queuedWork = new HashSet<>(); |
| while (true) { |
| try { |
| queuedWork.addAll(workQueue.getWorkQueued()); |
| return; |
| } catch (KeeperException e) { |
| if (KeeperException.Code.NONODE.equals(e.code())) { |
| log.warn("Could not find ZK root for replication work queue, will retry", e); |
| sleepUninterruptibly(500, TimeUnit.MILLISECONDS); |
| continue; |
| } |
| |
| log.error("Error reading existing queued replication work from ZooKeeper", e); |
| throw new RuntimeException("Error reading existing queued replication work from ZooKeeper", |
| e); |
| } catch (InterruptedException e) { |
| log.error("Error reading existing queued replication work from ZooKeeper", e); |
| throw new RuntimeException("Error reading existing queued replication work from ZooKeeper", |
| e); |
| } |
| } |
| } |
| |
| /** |
| * Distribute the work for the given path with filename |
| * |
| * @param path |
| * Path to the file being replicated |
| * @param target |
| * Target for the file to be replicated to |
| */ |
| @Override |
| protected boolean queueWork(Path path, ReplicationTarget target) { |
| String queueKey = DistributedWorkQueueWorkAssignerHelper.getQueueKey(path.getName(), target); |
| if (queuedWork.contains(queueKey)) { |
| log.debug("{} is already queued to be replicated to {}, not re-queueing", path, target); |
| return false; |
| } |
| |
| try { |
| log.debug("Queued work for {} and {}", queueKey, path); |
| workQueue.addWork(queueKey, path.toString()); |
| queuedWork.add(queueKey); |
| } catch (KeeperException | InterruptedException e) { |
| log.warn("Could not queue work for {}", path, e); |
| return false; |
| } |
| |
| return true; |
| } |
| |
| /** |
| * Iterate over the queued work to remove entries that have been completed. |
| */ |
| @Override |
| protected void cleanupFinishedWork() { |
| final Iterator<String> work = queuedWork.iterator(); |
| final String instanceId = client.instanceOperations().getInstanceID(); |
| while (work.hasNext()) { |
| String filename = work.next(); |
| // Null equates to the work was finished |
| if (zooCache |
| .get(ZooUtil.getRoot(instanceId) + ReplicationConstants.ZOO_WORK_QUEUE + "/" + filename) |
| == null) { |
| work.remove(); |
| } |
| } |
| } |
| |
| @Override |
| protected boolean shouldQueueWork(ReplicationTarget target) { |
| // We don't care about ordering, just replicate it all |
| return true; |
| } |
| |
| @Override |
| protected int getQueueSize() { |
| return this.queuedWork.size(); |
| } |
| |
| @Override |
| protected Set<String> getQueuedWork(ReplicationTarget target) { |
| String desiredQueueKeySuffix = |
| DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR + target.getPeerName() |
| + DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR + target.getRemoteIdentifier() |
| + DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR + target.getSourceTableId(); |
| Set<String> queuedWorkForTarget = new HashSet<>(); |
| for (String queuedWork : this.queuedWork) { |
| if (queuedWork.endsWith(desiredQueueKeySuffix)) { |
| queuedWorkForTarget.add(queuedWork); |
| } |
| } |
| |
| return queuedWorkForTarget; |
| } |
| |
| @Override |
| protected void removeQueuedWork(ReplicationTarget target, String queueKey) { |
| this.queuedWork.remove(queueKey); |
| } |
| } |