blob: 0f67cb6a22dd8e610aa4892caa45d563b02b4ac3 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one or more contributor license
* agreements. See the NOTICE file distributed with this work for additional information regarding
* copyright ownership. The ASF licenses this file to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance with the License. You may obtain a
* copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software distributed under the License
* is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express
* or implied. See the License for the specific language governing permissions and limitations under
* the License.
*/
package org.apache.geode.cache.lucene.internal;
import static org.apache.geode.distributed.ConfigurationProperties.MCAST_PORT;
import static org.junit.Assert.assertNotNull;
import static org.junit.Assert.assertTrue;
import java.io.IOException;
import java.util.concurrent.TimeUnit;
import org.apache.lucene.analysis.Analyzer;
import org.apache.lucene.analysis.standard.StandardAnalyzer;
import org.junit.After;
import org.junit.Assert;
import org.junit.Before;
import org.junit.experimental.categories.Category;
import org.apache.geode.cache.*;
import org.apache.geode.cache.lucene.LuceneIndex;
import org.apache.geode.cache.lucene.LuceneService;
import org.apache.geode.cache.lucene.LuceneServiceProvider;
import org.apache.geode.cache.lucene.internal.filesystem.FileSystemStats;
import org.apache.geode.cache.lucene.internal.repository.IndexRepository;
import org.apache.geode.cache.lucene.internal.repository.RepositoryManager;
import org.apache.geode.cache.lucene.internal.repository.serializer.HeterogeneousLuceneSerializer;
import org.apache.geode.internal.cache.BucketNotFoundException;
import org.apache.geode.internal.cache.GemFireCacheImpl;
import org.apache.geode.internal.cache.PartitionedRegion;
import org.apache.geode.test.junit.categories.IntegrationTest;
import org.apache.geode.test.junit.categories.LuceneTest;
@Category({IntegrationTest.class, LuceneTest.class})
public class LuceneIndexRecoveryHAIntegrationTest {
String[] indexedFields = new String[] {"txt"};
HeterogeneousLuceneSerializer mapper = new HeterogeneousLuceneSerializer();
Analyzer analyzer = new StandardAnalyzer();
Cache cache;
private LuceneIndexStats indexStats;
private FileSystemStats fileSystemStats;
@Before
public void setup() {
indexedFields = new String[] {"txt"};
mapper = new HeterogeneousLuceneSerializer();
analyzer = new StandardAnalyzer();
LuceneServiceImpl.registerDataSerializables();
cache = new CacheFactory().set(MCAST_PORT, "0").create();
}
@After
public void tearDown() {
if (cache != null) {
cache.close();
}
}
/**
* On rebalance, new repository manager will be created. It will try to read fileAndChunkRegion
* and construct index. This test simulates the same.
*/
// @Test
public void recoverRepoInANewNode()
throws BucketNotFoundException, IOException, InterruptedException {
LuceneServiceImpl service = (LuceneServiceImpl) LuceneServiceProvider.get(cache);
service.createIndexFactory().setFields(indexedFields).create("index1", "/userRegion");
PartitionAttributes<String, String> attrs =
new PartitionAttributesFactory().setTotalNumBuckets(1).create();
RegionFactory<String, String> regionfactory =
cache.createRegionFactory(RegionShortcut.PARTITION);
regionfactory.setPartitionAttributes(attrs);
PartitionedRegion userRegion = (PartitionedRegion) regionfactory.create("userRegion");
LuceneIndexForPartitionedRegion index =
(LuceneIndexForPartitionedRegion) service.getIndex("index1", "/userRegion");
// put an entry to create the bucket
userRegion.put("rebalance", "test");
service.waitUntilFlushed("index1", "userRegion", 30000, TimeUnit.MILLISECONDS);
RepositoryManager manager = new PartitionedRepositoryManager((LuceneIndexImpl) index, mapper);
IndexRepository repo = manager.getRepository(userRegion, 0, null);
assertNotNull(repo);
repo.create("rebalance", "test");
repo.commit();
// close the region to simulate bucket movement. New node will create repo using data persisted
// by old region
// ((PartitionedRegion)index.fileAndChunkRegion).close();
// ((PartitionedRegion)index.chunkRegion).close();
userRegion.close();
userRegion = (PartitionedRegion) regionfactory.create("userRegion");
userRegion.put("rebalance", "test");
manager = new PartitionedRepositoryManager((LuceneIndexImpl) index, mapper);
IndexRepository newRepo = manager.getRepository(userRegion, 0, null);
Assert.assertNotEquals(newRepo, repo);
}
private void verifyIndexFinishFlushing(String indexName, String regionName)
throws InterruptedException {
LuceneService service = LuceneServiceProvider.get(cache);
LuceneIndex index = service.getIndex(indexName, regionName);
boolean flushed = service.waitUntilFlushed(indexName, regionName, 60000, TimeUnit.MILLISECONDS);
assertTrue(flushed);
}
}