| /* |
| * Licensed to the Apache Software Foundation (ASF) under one or more |
| * contributor license agreements. See the NOTICE file distributed with |
| * this work for additional information regarding copyright ownership. |
| * The ASF licenses this file to You under the Apache License, Version 2.0 |
| * (the "License"); you may not use this file except in compliance with |
| * the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package org.apache.ignite.internal.processors.cache; |
| |
| import java.util.Arrays; |
| import java.util.LinkedHashMap; |
| import java.util.Map; |
| import java.util.Set; |
| import java.util.concurrent.CountDownLatch; |
| import java.util.concurrent.atomic.AtomicBoolean; |
| import java.util.concurrent.atomic.AtomicInteger; |
| import org.apache.ignite.Ignite; |
| import org.apache.ignite.IgniteCache; |
| import org.apache.ignite.IgniteCheckedException; |
| import org.apache.ignite.IgniteDataStreamer; |
| import org.apache.ignite.cache.CacheAtomicityMode; |
| import org.apache.ignite.cache.CacheWriteSynchronizationMode; |
| import org.apache.ignite.cache.affinity.rendezvous.RendezvousAffinityFunction; |
| import org.apache.ignite.cluster.ClusterState; |
| import org.apache.ignite.configuration.CacheConfiguration; |
| import org.apache.ignite.configuration.DataStorageConfiguration; |
| import org.apache.ignite.configuration.IgniteConfiguration; |
| import org.apache.ignite.internal.IgniteEx; |
| import org.apache.ignite.internal.IgniteInternalFuture; |
| import org.apache.ignite.internal.processors.cache.distributed.dht.IgniteClusterReadOnlyException; |
| import org.apache.ignite.internal.util.GridConcurrentHashSet; |
| import org.apache.ignite.internal.util.typedef.G; |
| import org.apache.ignite.internal.util.typedef.internal.CU; |
| import org.apache.ignite.spi.IgniteSpiException; |
| import org.apache.ignite.testframework.GridTestUtils; |
| import org.apache.ignite.testframework.MvccFeatureChecker; |
| import org.junit.Assert; |
| import org.junit.Assume; |
| import org.junit.Test; |
| |
| import static org.apache.ignite.cluster.ClusterState.ACTIVE; |
| import static org.apache.ignite.cluster.ClusterState.ACTIVE_READ_ONLY; |
| import static org.apache.ignite.cluster.ClusterState.INACTIVE; |
| import static org.apache.ignite.testframework.GridTestUtils.assertActive; |
| import static org.apache.ignite.testframework.GridTestUtils.assertThrowsAnyCause; |
| import static org.apache.ignite.testframework.GridTestUtils.assertThrowsWithCause; |
| |
| /** |
| * |
| */ |
| public class IgniteClusterActivateDeactivateTestWithPersistence extends IgniteClusterActivateDeactivateTest { |
| /** {@inheritDoc} */ |
| @Override protected boolean persistenceEnabled() { |
| return true; |
| } |
| |
| /** {@inheritDoc} */ |
| @Override protected void beforeTest() throws Exception { |
| cleanPersistenceDir(); |
| |
| super.beforeTest(); |
| } |
| |
| /** {@inheritDoc} */ |
| @Override protected void afterTest() throws Exception { |
| super.afterTest(); |
| |
| cleanPersistenceDir(); |
| } |
| |
| /** {@inheritDoc} */ |
| @Override protected IgniteConfiguration getConfiguration(String igniteInstanceName) throws Exception { |
| return super.getConfiguration(igniteInstanceName).setAutoActivationEnabled(false); |
| } |
| |
| /** |
| * @throws Exception If failed. |
| */ |
| @Test |
| public void testActivateCachesRestore_SingleNode() throws Exception { |
| activateCachesRestore(1, false, ACTIVE); |
| } |
| |
| /** |
| * @throws Exception If failed. |
| */ |
| @Test |
| public void testActivateReadOnlyCachesRestore_SingleNode() throws Exception { |
| activateCachesRestore(1, false, ACTIVE_READ_ONLY); |
| } |
| |
| /** |
| * @throws Exception If failed. |
| */ |
| @Test |
| public void testActivateCachesRestore_SingleNode_WithNewCaches() throws Exception { |
| activateCachesRestore(1, true, ACTIVE); |
| } |
| |
| /** |
| * @throws Exception If failed. |
| */ |
| @Test |
| public void testActivateReadOnlyCachesRestore_SingleNode_WithNewCaches() throws Exception { |
| activateCachesRestore(1, true, ACTIVE_READ_ONLY); |
| } |
| |
| /** |
| * @throws Exception If failed. |
| */ |
| @Test |
| public void testActivateCachesRestore_5_Servers() throws Exception { |
| activateCachesRestore(5, false, ACTIVE); |
| } |
| |
| /** |
| * @throws Exception If failed. |
| */ |
| @Test |
| public void testActivateReadOnlyCachesRestore_5_Servers() throws Exception { |
| activateCachesRestore(5, false, ACTIVE_READ_ONLY); |
| } |
| |
| /** |
| * @throws Exception If failed. |
| */ |
| @Test |
| public void testActivateCachesRestore_5_Servers_WithNewCaches() throws Exception { |
| activateCachesRestore(5, true, ACTIVE); |
| } |
| |
| /** |
| * @throws Exception If failed. |
| */ |
| @Test |
| public void testActivateReadOnlyCachesRestore_5_Servers_WithNewCaches() throws Exception { |
| activateCachesRestore(5, true, ACTIVE_READ_ONLY); |
| } |
| |
| /** |
| * Test deactivation on cluster that is not yet activated. |
| * |
| * @throws Exception If failed. |
| */ |
| @Test |
| public void testDeactivateInactiveCluster() throws Exception { |
| checkDeactivateInactiveCluster(ACTIVE); |
| } |
| |
| /** |
| * Test deactivation on cluster that is not yet activated. |
| * |
| * @throws Exception If failed. |
| */ |
| @Test |
| public void testDeactivateInactiveClusterReadOnly() throws Exception { |
| checkDeactivateInactiveCluster(ACTIVE_READ_ONLY); |
| } |
| |
| /** */ |
| private void checkDeactivateInactiveCluster(ClusterState activationMode) throws Exception { |
| assertActive(activationMode); |
| |
| ccfgs = new CacheConfiguration[] { |
| new CacheConfiguration<>("test_cache_1") |
| .setGroupName("test_cache") |
| .setAtomicityMode(CacheAtomicityMode.TRANSACTIONAL), |
| new CacheConfiguration<>("test_cache_2") |
| .setGroupName("test_cache") |
| .setAtomicityMode(CacheAtomicityMode.TRANSACTIONAL) |
| }; |
| |
| Ignite ignite = startGrids(3); |
| |
| ignite.cluster().state(activationMode); |
| |
| if (activationMode == ACTIVE) { |
| ignite.cache("test_cache_1") |
| .put("key1", "val1"); |
| ignite.cache("test_cache_2") |
| .put("key1", "val1"); |
| } |
| |
| ignite.cluster().state(INACTIVE); |
| |
| assertEquals(INACTIVE, ignite.cluster().state()); |
| |
| stopAllGrids(); |
| |
| ignite = startGrids(2); |
| |
| assertEquals(INACTIVE, ignite.cluster().state()); |
| |
| ignite.cluster().state(INACTIVE); |
| |
| assertEquals(INACTIVE, ignite.cluster().state()); |
| } |
| |
| /** */ |
| private Map<Integer, Integer> startGridsAndLoadData(int srvs, ClusterState activationMode) throws Exception { |
| assertActive(activationMode); |
| |
| Ignite srv = startGrids(srvs); |
| |
| srv.cluster().state(ACTIVE); |
| |
| srv.createCaches(Arrays.asList(cacheConfigurations1())); |
| |
| srv.cluster().state(activationMode); |
| |
| Map<Integer, Integer> cacheData = new LinkedHashMap<>(); |
| |
| for (CacheConfiguration ccfg : cacheConfigurations1()) { |
| for (int i = 1; i <= 100; i++) { |
| int key = -i; |
| int val = i; |
| |
| if (activationMode == ACTIVE) { |
| srv.cache(ccfg.getName()).put(key, val); |
| |
| cacheData.put(key, val); |
| } |
| else { |
| assertThrowsWithCause(() -> srv.cache(ccfg.getName()).put(key, val), IgniteClusterReadOnlyException.class); |
| |
| cacheData.put(key, null); |
| } |
| } |
| } |
| |
| return cacheData; |
| } |
| |
| /** |
| * @param srvs Number of server nodes. |
| * @param withNewCaches If {@code true} then after restart has new caches in configuration. |
| * @param activationMode Cluster activation mode. |
| * @throws Exception If failed. |
| */ |
| private void activateCachesRestore(int srvs, boolean withNewCaches, ClusterState activationMode) throws Exception { |
| assertActive(activationMode); |
| |
| Map<Integer, Integer> cacheData = startGridsAndLoadData(srvs, activationMode); |
| |
| stopAllGrids(); |
| |
| for (int i = 0; i < srvs; i++) { |
| if (withNewCaches) |
| ccfgs = cacheConfigurations2(); |
| |
| startGrid(i); |
| } |
| |
| Ignite srv = ignite(0); |
| |
| checkNoCaches(srvs); |
| |
| srv.cluster().state(activationMode); |
| |
| final int CACHES = withNewCaches ? 4 : 2; |
| |
| for (int i = 0; i < srvs; i++) |
| checkCachesOnNode(i, CACHES); |
| |
| DataStorageConfiguration dsCfg = srv.configuration().getDataStorageConfiguration(); |
| |
| checkCachesData(cacheData, dsCfg); |
| |
| checkCaches(srvs, CACHES); |
| |
| int nodes = srvs; |
| |
| startGrid(nodes++, false); |
| |
| for (int i = 0; i < nodes; i++) |
| checkCachesOnNode(i, CACHES); |
| |
| checkCaches(nodes, CACHES); |
| |
| startGrid(nodes++, true); |
| |
| checkCachesOnNode(nodes - 1, CACHES, false); |
| |
| checkCaches(nodes, CACHES); |
| |
| for (int i = 0; i < nodes; i++) |
| checkCachesOnNode(i, CACHES); |
| |
| checkCachesData(cacheData, dsCfg); |
| } |
| |
| |
| /** {@inheritDoc} */ |
| @Override protected void doFinalChecks(int startNodes, int nodesCnt) { |
| for (int i = 0; i < startNodes; i++) { |
| int j = i; |
| |
| assertThrowsAnyCause(log, () -> startGrid(j), IgniteSpiException.class, "not compatible"); |
| } |
| } |
| |
| /** |
| * @see <a href="https://issues.apache.org/jira/browse/IGNITE-7330">IGNITE-7330</a> for more information about context of the test |
| */ |
| @Test |
| public void testClientJoinsWhenActivationIsInProgress() throws Exception { |
| checkClientJoinsWhenActivationIsInProgress(ACTIVE); |
| } |
| |
| /** */ |
| @Test |
| public void testClientJoinsWhenActivationReanOnlyIsInProgress() throws Exception { |
| checkClientJoinsWhenActivationIsInProgress(ACTIVE_READ_ONLY); |
| } |
| |
| /** */ |
| private void checkClientJoinsWhenActivationIsInProgress(ClusterState state) throws Exception { |
| assertActive(state); |
| |
| startGridsAndLoadData(5, state); |
| |
| stopAllGrids(); |
| |
| Ignite srv = startGrids(5); |
| |
| final CountDownLatch clientStartLatch = new CountDownLatch(1); |
| |
| IgniteInternalFuture clStartFut = GridTestUtils.runAsync( |
| () -> { |
| try { |
| clientStartLatch.await(); |
| |
| Thread.sleep(10); |
| |
| Ignite cl = startClientGrid("client0"); |
| |
| IgniteCache<Object, Object> atomicCache = cl.cache(CACHE_NAME_PREFIX + '0'); |
| IgniteCache<Object, Object> txCache = cl.cache(CACHE_NAME_PREFIX + '1'); |
| |
| assertEquals(state == ACTIVE ? 100 : 0, atomicCache.size()); |
| assertEquals(state == ACTIVE ? 100 : 0, txCache.size()); |
| } |
| catch (Exception e) { |
| log.error("Error occurred", e); |
| |
| fail("Error occurred in client thread. Msg: " + e.getMessage()); |
| } |
| }, |
| "client-starter-thread" |
| ); |
| |
| clientStartLatch.countDown(); |
| srv.cluster().state(state); |
| |
| clStartFut.get(); |
| } |
| |
| /** |
| * Checks that persistent caches are present with actual data and volatile caches are missing. |
| * |
| * @param cacheData Cache data. |
| * @param dsCfg DataStorageConfiguration. |
| */ |
| private void checkCachesData(Map<Integer, Integer> cacheData, DataStorageConfiguration dsCfg) { |
| for (CacheConfiguration ccfg : cacheConfigurations1()) { |
| if (CU.isPersistentCache(ccfg, dsCfg)) |
| checkCacheData(cacheData, ccfg.getName()); |
| else { |
| for (Ignite node : G.allGrids()) |
| assertTrue(node.cache(ccfg.getName()) == null || node.cache(ccfg.getName()).size() == 0); |
| } |
| } |
| } |
| |
| /** |
| * @throws Exception If failed. |
| */ |
| @Test |
| public void testActivateCacheRestoreConfigurationConflict() throws Exception { |
| checkActivateCacheRestoreConfigurationConflict(ACTIVE); |
| } |
| |
| /** |
| * @throws Exception If failed. |
| */ |
| @Test |
| public void testActivateReadOnlyCacheRestoreConfigurationConflict() throws Exception { |
| checkActivateCacheRestoreConfigurationConflict(ACTIVE_READ_ONLY); |
| } |
| |
| /** */ |
| private void checkActivateCacheRestoreConfigurationConflict(ClusterState state) throws Exception { |
| assertActive(state); |
| |
| final int SRVS = 3; |
| |
| Ignite srv = startGrids(SRVS); |
| |
| srv.cluster().state(ACTIVE); |
| |
| CacheConfiguration ccfg = new CacheConfiguration(DEFAULT_CACHE_NAME) |
| .setAtomicityMode(CacheAtomicityMode.TRANSACTIONAL); |
| |
| srv.createCache(ccfg); |
| |
| srv.cluster().state(state); |
| |
| stopAllGrids(); |
| |
| ccfg = new CacheConfiguration(DEFAULT_CACHE_NAME + 1) |
| .setAtomicityMode(CacheAtomicityMode.TRANSACTIONAL); |
| |
| ccfg.setGroupName(DEFAULT_CACHE_NAME); |
| |
| ccfgs = new CacheConfiguration[] {ccfg}; |
| |
| assertThrowsAnyCause(log, () -> startGrids(SRVS), IgniteCheckedException.class, "Failed to start configured cache."); |
| } |
| |
| /** |
| * Test that after deactivation during eviction and rebalance and activation again after |
| * all data in cache is consistent. |
| * |
| * @throws Exception If failed. |
| */ |
| @Test |
| public void testDeactivateDuringEvictionAndRebalance() throws Exception { |
| Assume.assumeFalse("https://issues.apache.org/jira/browse/IGNITE-7384", MvccFeatureChecker.forcedMvcc()); |
| |
| IgniteEx srv = startGrids(3); |
| |
| srv.cluster().state(ACTIVE); |
| |
| CacheConfiguration ccfg = new CacheConfiguration(DEFAULT_CACHE_NAME) |
| .setBackups(1) |
| .setWriteSynchronizationMode(CacheWriteSynchronizationMode.FULL_SYNC) |
| .setIndexedTypes(Integer.class, Integer.class) |
| .setAffinity(new RendezvousAffinityFunction(false, 64)) |
| .setAtomicityMode(CacheAtomicityMode.TRANSACTIONAL); |
| |
| IgniteCache cache = srv.createCache(ccfg); |
| |
| // High number of keys triggers long partition eviction. |
| final int keysCount = 100_000; |
| |
| try (IgniteDataStreamer ds = srv.dataStreamer(DEFAULT_CACHE_NAME)) { |
| log.info("Writing initial data..."); |
| |
| ds.allowOverwrite(true); |
| for (int k = 1; k <= keysCount; k++) { |
| ds.addData(k, k); |
| |
| if (k % 50_000 == 0) |
| log.info("Written " + k + " entities."); |
| } |
| |
| log.info("Writing initial data finished."); |
| } |
| |
| AtomicInteger keyCounter = new AtomicInteger(keysCount); |
| AtomicBoolean stop = new AtomicBoolean(false); |
| |
| Set<Integer> addedKeys = new GridConcurrentHashSet<>(); |
| |
| IgniteInternalFuture cacheLoadFuture = GridTestUtils.runMultiThreadedAsync( |
| () -> { |
| while (!stop.get()) { |
| int key = keyCounter.incrementAndGet(); |
| try { |
| cache.put(key, key); |
| |
| addedKeys.add(key); |
| |
| Thread.sleep(10); |
| } |
| catch (Exception ignored) { } |
| } |
| }, |
| 2, |
| "cache-load" |
| ); |
| |
| stopGrid(2); |
| |
| // Wait for some data. |
| Thread.sleep(3000); |
| |
| startGrid(2); |
| |
| log.info("Stop load..."); |
| |
| stop.set(true); |
| |
| cacheLoadFuture.get(); |
| |
| // Deactivate and activate again. |
| srv.cluster().state(INACTIVE); |
| |
| srv.cluster().state(ACTIVE); |
| |
| awaitPartitionMapExchange(); |
| |
| log.info("Checking data..."); |
| |
| for (Ignite ignite : G.allGrids()) { |
| IgniteCache cache1 = ignite.getOrCreateCache(DEFAULT_CACHE_NAME); |
| |
| for (int k = 1; k <= keysCount; k++) { |
| Object val = cache1.get(k); |
| |
| Assert.assertNotNull("node=" + ignite.name() + ", key=" + k, val); |
| |
| Assert.assertTrue("node=" + ignite.name() + ", key=" + k + ", val=" + val, (int) val == k); |
| } |
| |
| for (int k : addedKeys) { |
| Object val = cache1.get(k); |
| |
| Assert.assertNotNull("node=" + ignite.name() + ", key=" + k, val); |
| |
| Assert.assertTrue("node=" + ignite.name() + ", key=" + k + ", val=" + val, (int) val == k); |
| } |
| } |
| } |
| } |