blob: 5d72660fe188936e4708dceee41842317d32d9e2 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.cassandra.distributed.test;
import java.util.HashMap;
import java.util.Map;
import java.util.concurrent.atomic.AtomicLong;
import org.apache.cassandra.io.util.File;
import org.junit.Test;
import net.bytebuddy.ByteBuddy;
import net.bytebuddy.dynamic.loading.ClassLoadingStrategy;
import net.bytebuddy.implementation.MethodDelegation;
import org.apache.cassandra.config.DatabaseDescriptor;
import org.apache.cassandra.db.ColumnFamilyStore;
import org.apache.cassandra.db.Keyspace;
import org.apache.cassandra.db.compaction.ActiveCompactions;
import org.apache.cassandra.db.compaction.CompactionStrategyManager;
import org.apache.cassandra.distributed.Cluster;
import org.apache.cassandra.streaming.StreamManager;
import static net.bytebuddy.matcher.ElementMatchers.named;
import static net.bytebuddy.matcher.ElementMatchers.takesArguments;
import static org.apache.cassandra.distributed.api.Feature.GOSSIP;
import static org.apache.cassandra.distributed.api.Feature.NETWORK;
public class StreamsDiskSpaceTest extends TestBaseImpl
{
@Test
public void testAbortStreams() throws Exception
{
try(Cluster cluster = init(Cluster.build(2)
.withConfig(config -> config.set("hinted_handoff_enabled", false)
.with(GOSSIP)
.with(NETWORK))
.withInstanceInitializer((cl, id) -> BB.doInstall(cl, id, StreamManager.class, "getTotalRemainingOngoingBytes"))
.start()))
{
cluster.schemaChange("create table " + KEYSPACE + ".tbl (id int primary key, t int) with compaction={'class': 'SizeTieredCompactionStrategy'}");
for (int i = 0; i < 10000; i++)
cluster.get(1).executeInternal("INSERT INTO " + KEYSPACE + ".tbl (id, t) values (?,?)", i, i);
cluster.get(1).flush(KEYSPACE);
cluster.get(2).runOnInstance(() -> BB.ongoing.set(Long.MAX_VALUE / 2));
cluster.get(1).nodetoolResult("repair", "-full").asserts().failure();
cluster.get(2).nodetoolResult("repair", "-full").asserts().failure();
cluster.get(2).runOnInstance(() -> BB.ongoing.set(0));
cluster.get(1).nodetoolResult("repair", "-full").asserts().success();
}
}
@Test
public void testAbortStreamsWhenOngoingCompactionsLeaveInsufficientSpace() throws Exception
{
try(Cluster cluster = init(Cluster.build(2)
.withConfig(config -> config.set("hinted_handoff_enabled", false)
.with(GOSSIP)
.with(NETWORK))
.withInstanceInitializer((cl, id) -> BB.doInstall(cl, id, ActiveCompactions.class, "estimatedRemainingWriteBytes"))
.start()))
{
cluster.schemaChange("create table " + KEYSPACE + ".tbl (id int primary key, t int) with compaction={'class': 'SizeTieredCompactionStrategy'}");
for (int i = 0; i < 10000; i++)
cluster.get(1).executeInternal("INSERT INTO " + KEYSPACE + ".tbl (id, t) values (?,?)", i, i);
cluster.get(1).flush(KEYSPACE);
cluster.get(2).runOnInstance(() -> {
ColumnFamilyStore cfs = Keyspace.open(KEYSPACE).getColumnFamilyStore("tbl");
BB.datadir = cfs.getDirectories().getLocationForDisk(cfs.getDirectories().getWriteableLocation(0));
});
cluster.get(2).runOnInstance(() -> BB.ongoing.set(Long.MAX_VALUE / 2));
cluster.get(1).nodetoolResult("repair", "-full").asserts().failure();
cluster.get(2).nodetoolResult("repair", "-full").asserts().failure();
cluster.get(2).runOnInstance(() -> BB.ongoing.set(0));
cluster.get(1).nodetoolResult("repair", "-full").asserts().success();
}
}
@Test
public void testAbortStreamsWhenTooManyPendingCompactions() throws Exception
{
try(Cluster cluster = init(Cluster.build(2)
.withConfig(config -> config.set("hinted_handoff_enabled", false)
.set("reject_repair_compaction_threshold", 1024)
.with(GOSSIP)
.with(NETWORK))
.withInstanceInitializer(BB::installCSMGetEstimatedRemainingTasks)
.start()))
{
cluster.schemaChange("create table " + KEYSPACE + ".tbl (id int primary key, t int) with compaction={'class': 'SizeTieredCompactionStrategy'}");
for (int i = 0; i < 10000; i++)
cluster.get(1).executeInternal("INSERT INTO " + KEYSPACE + ".tbl (id, t) values (?,?)", i, i);
cluster.get(1).flush(KEYSPACE);
cluster.get(2).runOnInstance(() -> {
BB.ongoing.set(DatabaseDescriptor.getRepairPendingCompactionRejectThreshold() + 1);
});
cluster.get(1).nodetoolResult("repair", "-full").asserts().failure();
cluster.get(2).nodetoolResult("repair", "-full").asserts().failure();
cluster.get(2).runOnInstance(() -> BB.ongoing.set(0));
cluster.get(1).nodetoolResult("repair", "-full").asserts().success();
}
}
public static class BB
{
public static AtomicLong ongoing = new AtomicLong();
public static File datadir;
private static void doInstall(ClassLoader cl, int id, Class<?> clazz, String method)
{
if (id != 2)
return;
new ByteBuddy().rebase(clazz)
.method(named(method))
.intercept(MethodDelegation.to(BB.class))
.make()
.load(cl, ClassLoadingStrategy.Default.INJECTION);
}
public static long getTotalRemainingOngoingBytes()
{
return ongoing.get();
}
public static Map<File, Long> estimatedRemainingWriteBytes()
{
Map<File, Long> ret = new HashMap<>();
if (datadir != null)
ret.put(datadir, ongoing.get());
return ret;
}
public static int getEstimatedRemainingTasks(int additionalSSTables, long additionalBytes, boolean isIncremental)
{
return (int) ongoing.get();
}
private static void installCSMGetEstimatedRemainingTasks(ClassLoader cl, int nodeNumber)
{
if (nodeNumber == 2)
{
new ByteBuddy().redefine(CompactionStrategyManager.class)
.method(named("getEstimatedRemainingTasks").and(takesArguments(3)))
.intercept(MethodDelegation.to(BB.class))
.make()
.load(cl, ClassLoadingStrategy.Default.INJECTION);
}
}
}
}