blob: 78038b60a82fd37a89e23eaa1825b9cf4c9f45b7 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.ignite.internal.processors.cache.transactions;
import java.util.HashSet;
import java.util.Iterator;
import java.util.List;
import java.util.Map;
import java.util.Queue;
import java.util.Set;
import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.ConcurrentLinkedQueue;
import java.util.concurrent.CountDownLatch;
import java.util.concurrent.TimeUnit;
import java.util.function.Supplier;
import java.util.stream.IntStream;
import javax.cache.Cache;
import org.apache.ignite.Ignite;
import org.apache.ignite.IgniteCache;
import org.apache.ignite.internal.IgniteEx;
import org.apache.ignite.internal.IgniteInterruptedCheckedException;
import org.apache.ignite.internal.TestRecordingCommunicationSpi;
import org.apache.ignite.internal.util.future.GridFutureAdapter;
import org.apache.ignite.internal.util.typedef.T2;
import org.apache.ignite.internal.util.typedef.internal.U;
import org.apache.ignite.lang.IgniteClosure;
import org.apache.ignite.lang.IgniteUuid;
import org.junit.Test;
/**
*
*/
public class TxPartitionCounterStateTwoPrimaryTwoBackupsTest extends TxPartitionCounterStateAbstractTest {
/** Tx sizes. */
private static final int[] SIZES = new int[] {5, 7, 3};
/** */
private static final int TOTAL = IntStream.of(SIZES).sum() + PRELOAD_KEYS_CNT;
/** */
private static final int PARTITION_ID = 0;
/** TODO test should pass with any combination of partition, add more variants ? */
private static final int PARTITION_ID_2 = PARTITION_ID + 1;
/** */
private static final int BACKUPS = 1;
/** */
private static final int NODES_CNT = 3;
/** */
@Test
public void testFailoverOnPrepare2Partitions() throws Exception {
doTestFailoverOnPrepare2Partitions(false);
}
/** */
@Test
public void testFailoverOnPrepare2PartitionsSkipCheckpoint() throws Exception {
doTestFailoverOnPrepare2Partitions(true);
}
/**
* Test scenario: 2pc, 2 partitions, 3 transactions.
* <p>
* 1. txs prepared in order 0, 1, 2
* <p>
* 2. prevent tx[0] and tx[1] prepare for part2 on primary.
* <p>
* 3. tx[2] committed for both partition.
* <p>
* 4. wait until tx[2] fully committed for both partitions and tx[0] and tx[1] fully prepared
* <p>
* 5. stop primary node for part2.
* <p>
* 6. tx[0], tx[1] rolled back due to prepare fail for part2.
* <p>
*
* Pass condition: counters for rolled back txs are incremented on primary and backup nodes.
*
* @param skipCheckpoint Skip checkpoint.
*/
private void doTestFailoverOnPrepare2Partitions(boolean skipCheckpoint) throws Exception {
final int finishedTxIdx = 2;
CountDownLatch commitLatch = new CountDownLatch(1); // Wait until tx is finished before stopping primary.
Map<Integer, T2<Ignite, List<Ignite>>> txTop = runOnPartition(PARTITION_ID, new Supplier<Integer>() {
@Override public Integer get() {
return PARTITION_ID_2;
}
}, BACKUPS, NODES_CNT,
new IgniteClosure<Map<Integer, T2<Ignite, List<Ignite>>>, TxCallback>() {
@Override public TxCallback apply(Map<Integer, T2<Ignite, List<Ignite>>> txTop) {
return new TxCallbackAdapter() {
/** */
private Queue<Integer> prepOrder = new ConcurrentLinkedQueue<Integer>();
{
prepOrder.add(0);
prepOrder.add(1);
prepOrder.add(2);
}
/** */
private Map<IgniteUuid, GridFutureAdapter<?>> prepFuts = new ConcurrentHashMap<>();
/** {@inheritDoc} */
@Override public boolean beforePrimaryPrepare(IgniteEx primary, IgniteUuid nearXidVer,
GridFutureAdapter<?> proceedFut) {
if (txTop.get(PARTITION_ID).get1() == primary) { // Order prepare for part1
runAsync(() -> {
prepFuts.put(nearXidVer, proceedFut);
// Order prepares.
if (prepFuts.size() == SIZES.length) { // Wait until all prep requests queued and force prepare order.
prepFuts.remove(version(prepOrder.poll())).onDone();
}
});
return true;
}
return order(nearXidVer) != finishedTxIdx; // Delay txs 0 and 1 preparation for part2, allow tx 2 to finish.
}
/** {@inheritDoc} */
@Override public boolean afterPrimaryPrepare(IgniteEx primary, IgniteInternalTx tx, IgniteUuid nearXidVer,
GridFutureAdapter<?> proceedFut) {
if (txTop.get(PARTITION_ID).get1() == primary) {
runAsync(() -> {
log.info("TX: Prepared part1: " + order(nearXidVer));
if (prepOrder.isEmpty()) {
log.info("TX: All prepared part1");
try {
assertTrue(U.await(commitLatch, 30, TimeUnit.SECONDS));
}
catch (IgniteInterruptedCheckedException e) {
fail();
}
// fail primary for second partition and trigger rollback for prepared transactions on
// when txs for part1 are fully prepared and tx with idx=finishedTxIdx was committed.
stopGrid(skipCheckpoint, txTop.get(PARTITION_ID_2).get1().name());
TestRecordingCommunicationSpi.stopBlockAll();
return;
}
prepFuts.remove(version(prepOrder.poll())).onDone();
});
}
return order(nearXidVer) != finishedTxIdx; // Delay final preparation for tx[0] and tx[1]
}
@Override public boolean afterPrimaryFinish(IgniteEx primary, IgniteUuid nearXidVer,
GridFutureAdapter<?> proceedFut) {
commitLatch.countDown();
return super.afterPrimaryFinish(primary, nearXidVer, proceedFut);
}
};
}
},
SIZES);
waitForTopology(NODES_CNT);
awaitPartitionMapExchange();
IgniteCache<Integer, Integer> cache = grid(CLIENT_GRID_NAME).cache(DEFAULT_CACHE_NAME);
Iterator<Cache.Entry<Integer, Integer>> iter = cache.iterator();
Set<Integer> keys = new HashSet<>();
while (iter.hasNext()) {
Cache.Entry<Integer, Integer> next = iter.next();
keys.add(next.getKey());
}
Set<Integer> expKeys = new HashSet<>();
List<Integer> part1Keys = partitionKeys(cache, PARTITION_ID, TOTAL, 0);
expKeys.addAll(part1Keys.subList(0, PRELOAD_KEYS_CNT)); // Preload keys.
expKeys.addAll(part1Keys.subList(PRELOAD_KEYS_CNT + SIZES[0] + SIZES[1], TOTAL)); // Preload keys.
List<Integer> part2Keys = partitionKeys(cache, PARTITION_ID_2, SIZES.length, 0);
expKeys.add(part2Keys.get(finishedTxIdx));
assertEqualsCollections(expKeys, keys);
// Expect only one committed tx.
assertEquals(
PRELOAD_KEYS_CNT + SIZES[finishedTxIdx] + 1 /** Key for part2. */,
grid(CLIENT_GRID_NAME).cache(DEFAULT_CACHE_NAME).size()
);
// Expect consistent partitions.
assertPartitionsSame(idleVerify(grid(CLIENT_GRID_NAME), DEFAULT_CACHE_NAME));
assertCountersSame(PARTITION_ID, true, DEFAULT_CACHE_NAME);
}
}