| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| */ |
| package org.apache.jackrabbit.oak.plugins.document; |
| |
| import com.google.common.collect.Iterables; |
| |
| import org.apache.jackrabbit.oak.plugins.document.mongo.MongoDocumentStore; |
| import org.apache.jackrabbit.oak.plugins.document.mongo.MongoMissingLastRevSeeker; |
| import org.apache.jackrabbit.oak.plugins.document.rdb.RDBDocumentStore; |
| import org.apache.jackrabbit.oak.plugins.document.rdb.RDBMissingLastRevSeeker; |
| import org.apache.jackrabbit.oak.plugins.document.util.Utils; |
| import org.apache.jackrabbit.oak.spi.commit.CommitInfo; |
| import org.apache.jackrabbit.oak.spi.commit.EmptyHook; |
| import org.apache.jackrabbit.oak.spi.state.NodeBuilder; |
| import org.apache.jackrabbit.oak.stats.Clock; |
| import org.junit.After; |
| import org.junit.Before; |
| import org.junit.Test; |
| |
| import static org.apache.jackrabbit.oak.plugins.document.ClusterNodeInfo.DEFAULT_LEASE_DURATION_MILLIS; |
| import static org.apache.jackrabbit.oak.plugins.document.NodeDocument.NUM_REVS_THRESHOLD; |
| import static org.apache.jackrabbit.oak.plugins.document.RecoveryHandler.NOOP; |
| import static org.junit.Assert.assertEquals; |
| import static org.junit.Assert.assertFalse; |
| import static org.junit.Assert.assertNotNull; |
| import static org.junit.Assert.assertNull; |
| import static org.junit.Assert.assertTrue; |
| |
| public class MissingLastRevSeekerTest extends AbstractDocumentStoreTest { |
| |
| private Clock clock; |
| private DocumentStore store; |
| private MissingLastRevSeeker seeker; |
| |
| public MissingLastRevSeekerTest(DocumentStoreFixture dsf) { |
| super(dsf); |
| } |
| |
| @Before |
| public void before() throws Exception { |
| clock = new Clock.Virtual(); |
| clock.waitUntil(System.currentTimeMillis()); |
| Revision.setClock(clock); |
| ClusterNodeInfo.setClock(clock); |
| store = ds; |
| if (dsf == DocumentStoreFixture.MONGO) { |
| seeker = new MongoMissingLastRevSeeker((MongoDocumentStore) store, clock); |
| } else if (store instanceof RDBDocumentStore) { |
| seeker = new RDBMissingLastRevSeeker((RDBDocumentStore) store, clock); |
| } else { |
| seeker = new MissingLastRevSeeker(store, clock); |
| } |
| removeMeClusterNodes.add("1"); |
| removeMeClusterNodes.add("2"); |
| } |
| |
| @After |
| public void after() { |
| ClusterNodeInfo.resetClockToDefault(); |
| Revision.resetClockToDefault(); |
| } |
| |
| private void markDocumentsForCleanup() { |
| for (NodeDocument doc : Utils.getAllDocuments(ds)) { |
| removeMe.add(doc.getId()); |
| } |
| } |
| |
| @Test |
| public void acquireRecoveryLockOnActiveClusterNode() { |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 1); |
| |
| assertFalse(seeker.acquireRecoveryLock(1, 2)); |
| } |
| |
| @Test |
| public void acquireRecoveryLockOnInactiveClusterNode() { |
| ClusterNodeInfo nodeInfo1 = ClusterNodeInfo.getInstance(store, NOOP, null, null, 1); |
| nodeInfo1.dispose(); |
| |
| assertFalse(seeker.acquireRecoveryLock(1, 2)); |
| } |
| |
| @Test |
| public void acquireRecoveryLockOnExpiredLease() throws Exception { |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 1); |
| // expire the lease |
| clock.waitUntil(clock.getTime() + DEFAULT_LEASE_DURATION_MILLIS + 1); |
| |
| assertTrue(seeker.acquireRecoveryLock(1, 2)); |
| } |
| |
| @Test |
| public void acquireRecoveryLockOnAlreadyLocked() throws Exception { |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 1); |
| // expire the lease |
| clock.waitUntil(clock.getTime() + DEFAULT_LEASE_DURATION_MILLIS + 1); |
| |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 2); |
| |
| assertTrue(seeker.acquireRecoveryLock(1, 2)); |
| assertFalse(seeker.acquireRecoveryLock(1, 3)); |
| } |
| |
| @Test |
| public void acquireRecoveryLockAgain() throws Exception { |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 1); |
| // expire the lease |
| clock.waitUntil(clock.getTime() + DEFAULT_LEASE_DURATION_MILLIS + 1); |
| |
| assertTrue(seeker.acquireRecoveryLock(1, 2)); |
| assertTrue(seeker.acquireRecoveryLock(1, 2)); |
| } |
| |
| @Test |
| public void releaseRecoveryLockSuccessTrue() throws Exception { |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 1); |
| // expire the lease |
| clock.waitUntil(clock.getTime() + DEFAULT_LEASE_DURATION_MILLIS + 1); |
| |
| assertTrue(seeker.acquireRecoveryLock(1, 2)); |
| assertTrue(getClusterNodeInfo(1).isBeingRecovered()); |
| assertTrue(getClusterNodeInfo(1).isActive()); |
| seeker.releaseRecoveryLock(1, true); |
| assertFalse(getClusterNodeInfo(1).isBeingRecovered()); |
| assertFalse(getClusterNodeInfo(1).isActive()); |
| // recovery not needed anymore |
| assertFalse(getClusterNodeInfo(1).isRecoveryNeeded(clock.getTime())); |
| assertFalse(seeker.acquireRecoveryLock(1, 2)); |
| } |
| |
| @Test |
| public void releaseRecoveryLockSuccessFalse() throws Exception { |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 1); |
| // expire the lease |
| clock.waitUntil(clock.getTime() + DEFAULT_LEASE_DURATION_MILLIS + 1); |
| |
| assertTrue(seeker.acquireRecoveryLock(1, 2)); |
| assertTrue(getClusterNodeInfo(1).isBeingRecovered()); |
| assertTrue(getClusterNodeInfo(1).isActive()); |
| seeker.releaseRecoveryLock(1, false); |
| assertFalse(getClusterNodeInfo(1).isBeingRecovered()); |
| assertTrue(getClusterNodeInfo(1).isActive()); |
| // recovery still needed |
| assertTrue(getClusterNodeInfo(1).isRecoveryNeeded(clock.getTime())); |
| assertTrue(seeker.acquireRecoveryLock(1, 2)); |
| } |
| |
| @Test |
| public void isRecoveryNeeded() throws Exception { |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 1); |
| // expire the lease |
| clock.waitUntil(clock.getTime() + DEFAULT_LEASE_DURATION_MILLIS + 1); |
| |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 2); |
| |
| assertTrue(seeker.isRecoveryNeeded()); |
| assertTrue(getClusterNodeInfo(1).isRecoveryNeeded(clock.getTime())); |
| assertFalse(getClusterNodeInfo(2).isRecoveryNeeded(clock.getTime())); |
| |
| assertTrue(seeker.acquireRecoveryLock(1, 2)); |
| seeker.releaseRecoveryLock(1, true); |
| |
| assertFalse(seeker.isRecoveryNeeded()); |
| assertFalse(getClusterNodeInfo(1).isRecoveryNeeded(clock.getTime())); |
| assertFalse(getClusterNodeInfo(2).isRecoveryNeeded(clock.getTime())); |
| } |
| |
| @Test |
| public void isRecoveryNeededWithRecoveryLock() throws Exception { |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 1); |
| // expire the lease |
| clock.waitUntil(clock.getTime() + DEFAULT_LEASE_DURATION_MILLIS + 1); |
| |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 2); |
| |
| assertTrue(seeker.acquireRecoveryLock(1, 2)); |
| |
| assertTrue(seeker.isRecoveryNeeded()); |
| assertTrue(getClusterNodeInfo(1).isRecoveryNeeded(clock.getTime())); |
| |
| seeker.releaseRecoveryLock(1, true); |
| |
| assertFalse(seeker.isRecoveryNeeded()); |
| assertFalse(getClusterNodeInfo(1).isRecoveryNeeded(clock.getTime())); |
| } |
| |
| |
| @Test |
| public void getAllClusterNodes() { |
| assertEquals(0, Iterables.size(seeker.getAllClusters())); |
| |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 1); |
| |
| assertEquals(1, Iterables.size(seeker.getAllClusters())); |
| |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 2); |
| |
| assertEquals(2, Iterables.size(seeker.getAllClusters())); |
| } |
| |
| @Test |
| public void getClusterNodeInfo() { |
| assertNull(getClusterNodeInfo(1)); |
| |
| ClusterNodeInfo.getInstance(store, NOOP, null, null, 1); |
| |
| assertNotNull(getClusterNodeInfo(1)); |
| } |
| |
| private ClusterNodeInfoDocument getClusterNodeInfo(int clusterId) { |
| return seeker.getClusterNodeInfo(clusterId); |
| } |
| |
| @Test |
| public void getNonSplitDocs() throws Exception { |
| String nodeName = this.getClass().getName() + "-foo"; |
| DocumentNodeStore dns = getBuilder().clock(clock).setAsyncDelay(0).setDocumentStore(new DocumentStoreWrapper(store) { |
| @Override |
| public void dispose() { |
| // do not close underlying store, otherwise cleanup |
| // cannot remove documents after the test |
| } |
| }).getNodeStore(); |
| NodeBuilder b1 = dns.getRoot().builder(); |
| b1.child(nodeName); |
| dns.merge(b1, EmptyHook.INSTANCE, CommitInfo.EMPTY); |
| //Modify and commit changes on this node 100 times to create a split document |
| for (int i = 0; i < NUM_REVS_THRESHOLD; i++) { |
| b1 = dns.getRoot().builder(); |
| b1.child(nodeName).setProperty("prop",i); |
| dns.merge(b1, EmptyHook.INSTANCE, CommitInfo.EMPTY); |
| } |
| dns.runBackgroundOperations(); |
| //seeker should return only non split documents |
| int docs = Iterables.size(seeker.getCandidates(0)); |
| assertEquals(2, docs); |
| markDocumentsForCleanup(); |
| dns.dispose(); |
| } |
| } |