| /** |
| * Licensed to the Apache Software Foundation (ASF) under one or more |
| * contributor license agreements. See the NOTICE file distributed with |
| * this work for additional information regarding copyright ownership. |
| * The ASF licenses this file to You under the Apache License, Version 2.0 |
| * (the "License"); you may not use this file except in compliance with |
| * the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.activemq.bugs; |
| |
| import static org.junit.Assert.assertEquals; |
| import static org.junit.Assert.assertNull; |
| import static org.junit.Assert.assertNotNull; |
| import static org.junit.Assert.assertTrue; |
| import static org.junit.Assert.fail; |
| |
| import java.io.File; |
| import java.io.IOException; |
| import java.util.Collection; |
| import java.util.concurrent.TimeUnit; |
| |
| import javax.jms.Connection; |
| import javax.jms.Destination; |
| import javax.jms.JMSException; |
| import javax.jms.Message; |
| import javax.jms.MessageConsumer; |
| import javax.jms.MessageProducer; |
| import javax.jms.Queue; |
| import javax.jms.Session; |
| import javax.jms.Topic; |
| |
| import org.apache.activemq.ActiveMQConnectionFactory; |
| import org.apache.activemq.ActiveMQSession; |
| import org.apache.activemq.broker.BrokerService; |
| import org.apache.activemq.command.ActiveMQQueue; |
| import org.apache.activemq.command.ActiveMQTopic; |
| import org.apache.activemq.leveldb.LevelDBStore; |
| import org.apache.activemq.store.PersistenceAdapter; |
| import org.apache.activemq.store.kahadb.KahaDBPersistenceAdapter; |
| import org.apache.activemq.store.kahadb.disk.journal.DataFile; |
| import org.apache.activemq.store.kahadb.disk.journal.Journal; |
| import org.apache.activemq.util.Wait; |
| import org.junit.After; |
| import org.junit.Test; |
| import org.slf4j.Logger; |
| import org.slf4j.LoggerFactory; |
| |
| public class AMQ2832Test { |
| |
| private static final Logger LOG = LoggerFactory.getLogger(AMQ2832Test.class); |
| |
| BrokerService broker = null; |
| private ActiveMQConnectionFactory cf; |
| private final Destination destination = new ActiveMQQueue("AMQ2832Test"); |
| private String connectionUri; |
| |
| protected void startBroker() throws Exception { |
| doStartBroker(true, false); |
| } |
| |
| protected void restartBroker() throws Exception { |
| if (broker != null) { |
| broker.stop(); |
| broker.waitUntilStopped(); |
| } |
| doStartBroker(false, false); |
| } |
| |
| protected void recoverBroker() throws Exception { |
| if (broker != null) { |
| broker.stop(); |
| broker.waitUntilStopped(); |
| } |
| doStartBroker(false, true); |
| } |
| |
| private void doStartBroker(boolean delete, boolean recover) throws Exception { |
| broker = new BrokerService(); |
| broker.setDeleteAllMessagesOnStartup(delete); |
| broker.setPersistent(true); |
| broker.setUseJmx(true); |
| broker.addConnector("tcp://localhost:0"); |
| |
| configurePersistence(broker, recover); |
| |
| connectionUri = "vm://localhost?create=false"; |
| cf = new ActiveMQConnectionFactory(connectionUri); |
| |
| broker.start(); |
| LOG.info("Starting broker.."); |
| } |
| |
| protected void configurePersistence(BrokerService brokerService, boolean recover) throws Exception { |
| KahaDBPersistenceAdapter adapter = (KahaDBPersistenceAdapter) brokerService.getPersistenceAdapter(); |
| |
| // ensure there are a bunch of data files but multiple entries in each |
| adapter.setJournalMaxFileLength(1024 * 20); |
| |
| // speed up the test case, checkpoint an cleanup early and often |
| adapter.setCheckpointInterval(5000); |
| adapter.setCleanupInterval(5000); |
| adapter.setPreallocationScope(Journal.PreallocationScope.ENTIRE_JOURNAL.name()); |
| |
| if (recover) { |
| adapter.setForceRecoverIndex(true); |
| } |
| } |
| |
| @After |
| public void tearDown() throws Exception { |
| if (broker != null) { |
| broker.stop(); |
| broker.waitUntilStopped(); |
| } |
| } |
| |
| /** |
| * Scenario: |
| * db-1.log has an unacknowledged message, |
| * db-2.log contains acks for the messages from db-1.log, |
| * db-3.log contains acks for the messages from db-2.log |
| * |
| * Expected behavior: since db-1.log is blocked, db-2.log and db-3.log should not be removed during the cleanup. |
| * Current situation on 5.10.0, 5.10.1 is that db-3.log is removed causing all messages from db-2.log, whose acks were in db-3.log, to be replayed. |
| * |
| * @throws Exception |
| */ |
| @Test |
| public void testAckChain() throws Exception { |
| startBroker(); |
| |
| makeAckChain(); |
| |
| broker.stop(); |
| broker.waitUntilStopped(); |
| |
| recoverBroker(); |
| |
| StagedConsumer consumer = new StagedConsumer(); |
| Message message = consumer.receive(1); |
| assertNotNull("One message stays unacked from db-1.log", message); |
| message.acknowledge(); |
| message = consumer.receive(1); |
| assertNull("There should not be any unconsumed messages any more", message); |
| consumer.close(); |
| } |
| |
| private void makeAckChain() throws Exception { |
| StagedConsumer consumer = new StagedConsumer(); |
| // file #1 |
| produceMessagesToConsumeMultipleDataFiles(5); |
| // acknowledge first 2 messages and leave the 3rd one unacknowledged blocking db-1.log |
| consumer.receive(3); |
| |
| // send messages by consuming and acknowledging every message right after sent in order to get KahadbAdd and Remove command to be saved together |
| // this is necessary in order to get KahaAddMessageCommand to be saved in one db file and the corresponding KahaRemoveMessageCommand in the next one |
| produceAndConsumeImmediately(20, consumer); |
| consumer.receive(2).acknowledge(); // consume and ack the last 2 unconsumed |
| |
| // now we have 3 files written and started with #4 |
| consumer.close(); |
| } |
| |
| @Test |
| public void testNoRestartOnMissingAckDataFile() throws Exception { |
| startBroker(); |
| |
| // reuse scenario from previous test |
| makeAckChain(); |
| |
| File dataDir = broker.getPersistenceAdapter().getDirectory(); |
| broker.stop(); |
| broker.waitUntilStopped(); |
| |
| File secondLastDataFile = new File(dataDir, "db-3.log"); |
| LOG.info("Whacking data file with acks: " + secondLastDataFile); |
| secondLastDataFile.delete(); |
| |
| try { |
| doStartBroker(false, false); |
| fail("Expect failure to start with corrupt journal"); |
| } catch (IOException expected) { |
| } |
| } |
| |
| |
| private void produceAndConsumeImmediately(int numOfMsgs, StagedConsumer consumer) throws Exception { |
| for (int i = 0; i < numOfMsgs; i++) { |
| produceMessagesToConsumeMultipleDataFiles(1); |
| consumer.receive(1).acknowledge(); |
| } |
| } |
| |
| @Test |
| public void testAckRemovedMessageReplayedAfterRecovery() throws Exception { |
| |
| startBroker(); |
| |
| StagedConsumer consumer = new StagedConsumer(); |
| int numMessagesAvailable = produceMessagesToConsumeMultipleDataFiles(20); |
| // this will block the reclaiming of one data file |
| Message firstUnacked = consumer.receive(10); |
| LOG.info("first unacked: " + firstUnacked.getJMSMessageID()); |
| Message secondUnacked = consumer.receive(1); |
| LOG.info("second unacked: " + secondUnacked.getJMSMessageID()); |
| numMessagesAvailable -= 11; |
| |
| numMessagesAvailable += produceMessagesToConsumeMultipleDataFiles(10); |
| // ensure ack is another data file |
| LOG.info("Acking firstUnacked: " + firstUnacked.getJMSMessageID()); |
| firstUnacked.acknowledge(); |
| |
| numMessagesAvailable += produceMessagesToConsumeMultipleDataFiles(10); |
| |
| consumer.receive(numMessagesAvailable).acknowledge(); |
| |
| // second unacked should keep first data file available but journal with the first ack |
| // may get whacked |
| consumer.close(); |
| |
| broker.stop(); |
| broker.waitUntilStopped(); |
| |
| recoverBroker(); |
| |
| consumer = new StagedConsumer(); |
| // need to force recovery? |
| |
| Message msg = consumer.receive(1, 5); |
| assertNotNull("One messages left after recovery", msg); |
| msg.acknowledge(); |
| |
| // should be no more messages |
| msg = consumer.receive(1, 5); |
| assertEquals("Only one messages left after recovery: " + msg, null, msg); |
| consumer.close(); |
| } |
| |
| @Test |
| public void testAlternateLossScenario() throws Exception { |
| |
| startBroker(); |
| PersistenceAdapter pa = broker.getPersistenceAdapter(); |
| if (pa instanceof LevelDBStore) { |
| return; |
| } |
| |
| ActiveMQQueue queue = new ActiveMQQueue("MyQueue"); |
| ActiveMQQueue disposable = new ActiveMQQueue("MyDisposableQueue"); |
| ActiveMQTopic topic = new ActiveMQTopic("MyDurableTopic"); |
| |
| // This ensure that data file 1 never goes away. |
| createInactiveDurableSub(topic); |
| assertEquals(1, getNumberOfJournalFiles()); |
| |
| // One Queue Message that will be acked in another data file. |
| produceMessages(queue, 1); |
| assertEquals(1, getNumberOfJournalFiles()); |
| |
| // Add some messages to consume space |
| produceMessages(disposable, 50); |
| |
| int dataFilesCount = getNumberOfJournalFiles(); |
| assertTrue(dataFilesCount > 1); |
| |
| // Create an ack for the single message on this queue |
| drainQueue(queue); |
| |
| // Add some more messages to consume space beyond tha data file with the ack |
| produceMessages(disposable, 50); |
| |
| assertTrue(dataFilesCount < getNumberOfJournalFiles()); |
| dataFilesCount = getNumberOfJournalFiles(); |
| |
| restartBroker(); |
| |
| // Clear out all queue data |
| broker.getAdminView().removeQueue(disposable.getQueueName()); |
| |
| // Once this becomes true our ack could be lost. |
| assertTrue("Less than three journal file expected, was " + getNumberOfJournalFiles(), Wait.waitFor(new Wait.Condition() { |
| @Override |
| public boolean isSatisified() throws Exception { |
| return getNumberOfJournalFiles() <= 4; |
| } |
| }, TimeUnit.MINUTES.toMillis(3))); |
| |
| // Recover and the Message should not be replayed but if the old MessageAck is lost |
| // then it could be. |
| recoverBroker(); |
| |
| assertTrue(drainQueue(queue) == 0); |
| } |
| |
| private int getNumberOfJournalFiles() throws IOException { |
| |
| Collection<DataFile> files = |
| ((KahaDBPersistenceAdapter) broker.getPersistenceAdapter()).getStore().getJournal().getFileMap().values(); |
| LOG.info("Data files: " + files); |
| int reality = 0; |
| for (DataFile file : files) { |
| if (file != null) { |
| reality++; |
| } |
| } |
| |
| return reality; |
| } |
| |
| private void createInactiveDurableSub(Topic topic) throws Exception { |
| Connection connection = cf.createConnection(); |
| connection.setClientID("Inactive"); |
| Session session = connection.createSession(false, Session.AUTO_ACKNOWLEDGE); |
| MessageConsumer consumer = session.createDurableSubscriber(topic, "Inactive"); |
| consumer.close(); |
| connection.close(); |
| produceMessages(topic, 1); |
| } |
| |
| private int drainQueue(Queue queue) throws Exception { |
| Connection connection = cf.createConnection(); |
| connection.setClientID("Inactive"); |
| connection.start(); |
| Session session = connection.createSession(false, Session.AUTO_ACKNOWLEDGE); |
| MessageConsumer consumer = session.createConsumer(queue); |
| int count = 0; |
| while (consumer.receive(5000) != null) { |
| count++; |
| } |
| consumer.close(); |
| connection.close(); |
| return count; |
| } |
| |
| private int produceMessages(Destination destination, int numToSend) throws Exception { |
| int sent = 0; |
| Connection connection = new ActiveMQConnectionFactory( |
| broker.getTransportConnectors().get(0).getConnectUri()).createConnection(); |
| connection.start(); |
| try { |
| Session session = connection.createSession(false, Session.AUTO_ACKNOWLEDGE); |
| MessageProducer producer = session.createProducer(destination); |
| for (int i = 0; i < numToSend; i++) { |
| producer.send(createMessage(session, i)); |
| sent++; |
| } |
| } finally { |
| connection.close(); |
| } |
| |
| return sent; |
| } |
| |
| private int produceMessagesToConsumeMultipleDataFiles(int numToSend) throws Exception { |
| return produceMessages(destination, numToSend); |
| } |
| |
| final String payload = new String(new byte[1024]); |
| |
| private Message createMessage(Session session, int i) throws Exception { |
| return session.createTextMessage(payload + "::" + i); |
| } |
| |
| private class StagedConsumer { |
| Connection connection; |
| MessageConsumer consumer; |
| |
| StagedConsumer() throws Exception { |
| connection = new ActiveMQConnectionFactory("failover://" + |
| broker.getTransportConnectors().get(0).getConnectUri().toString()).createConnection(); |
| connection.start(); |
| consumer = connection.createSession(false, ActiveMQSession.INDIVIDUAL_ACKNOWLEDGE).createConsumer(destination); |
| } |
| |
| public Message receive(int numToReceive) throws Exception { |
| return receive(numToReceive, 2); |
| } |
| |
| public Message receive(int numToReceive, int timeoutInSeconds) throws Exception { |
| Message msg = null; |
| for (; numToReceive > 0; numToReceive--) { |
| |
| do { |
| msg = consumer.receive(1*1000); |
| } while (msg == null && --timeoutInSeconds > 0); |
| |
| if (numToReceive > 1) { |
| msg.acknowledge(); |
| } |
| |
| if (msg != null) { |
| LOG.debug("received: " + msg.getJMSMessageID()); |
| } |
| } |
| // last message, unacked |
| return msg; |
| } |
| |
| void close() throws JMSException { |
| consumer.close(); |
| connection.close(); |
| } |
| } |
| } |