blob: 1c10be515669c26faefcef99c18736f79241bfcd [file] [log] [blame]
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.falcon.messaging;
import java.io.InputStream;
import java.io.OutputStream;
import javax.jms.Connection;
import javax.jms.ConnectionFactory;
import javax.jms.Destination;
import javax.jms.JMSException;
import javax.jms.MapMessage;
import javax.jms.MessageConsumer;
import javax.jms.Session;
import org.apache.activemq.ActiveMQConnectionFactory;
import org.apache.activemq.broker.BrokerService;
import org.apache.activemq.util.ByteArrayInputStream;
import org.apache.falcon.cluster.util.EmbeddedCluster;
import org.apache.falcon.workflow.WorkflowExecutionArgs;
import org.apache.falcon.workflow.WorkflowExecutionContext;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IOUtils;
import org.testng.Assert;
import org.testng.annotations.AfterClass;
import org.testng.annotations.BeforeClass;
import org.testng.annotations.Test;
/**
* Test for feed message producer.
*/
public class FeedProducerTest {
private String[] args;
private static final String BROKER_URL = "vm://localhost?broker.useJmx=false&broker.persistent=true";
private static final String BROKER_IMPL_CLASS = "org.apache.activemq.ActiveMQConnectionFactory";
private static final String TOPIC_NAME = "Falcon.process1.click-logs";
private BrokerService broker;
private Path logFile;
private volatile AssertionError error;
private EmbeddedCluster dfsCluster;
private Configuration conf;
@BeforeClass
public void setup() throws Exception {
this.dfsCluster = EmbeddedCluster.newCluster("testCluster");
conf = dfsCluster.getConf();
logFile = new Path(conf.get("fs.default.name"),
"/falcon/feed/agg-logs/instance-2012-01-01-10-00.csv");
args = new String[] {
"-" + WorkflowExecutionArgs.ENTITY_NAME.getName(), TOPIC_NAME,
"-" + WorkflowExecutionArgs.FEED_NAMES.getName(), "click-logs",
"-" + WorkflowExecutionArgs.FEED_INSTANCE_PATHS.getName(),
"/click-logs/10/05/05/00/20",
"-" + WorkflowExecutionArgs.WORKFLOW_ID.getName(), "workflow-01-00",
"-" + WorkflowExecutionArgs.WORKFLOW_USER.getName(), "falcon",
"-" + WorkflowExecutionArgs.RUN_ID.getName(), "1",
"-" + WorkflowExecutionArgs.NOMINAL_TIME.getName(), "2011-01-01-01-00",
"-" + WorkflowExecutionArgs.TIMESTAMP.getName(), "2012-01-01-01-00",
"-" + WorkflowExecutionArgs.BRKR_URL.getName(), BROKER_URL,
"-" + WorkflowExecutionArgs.BRKR_IMPL_CLASS.getName(), BROKER_IMPL_CLASS,
"-" + WorkflowExecutionArgs.USER_BRKR_URL.getName(), BROKER_URL,
"-" + WorkflowExecutionArgs.USER_BRKR_IMPL_CLASS.getName(), BROKER_IMPL_CLASS,
"-" + WorkflowExecutionArgs.ENTITY_TYPE.getName(), "FEED",
"-" + WorkflowExecutionArgs.OPERATION.getName(), "DELETE",
"-" + WorkflowExecutionArgs.LOG_FILE.getName(), logFile.toString(),
"-" + WorkflowExecutionArgs.LOG_DIR.getName(), "/falcon/feed/agg-logs/",
"-" + WorkflowExecutionArgs.TOPIC_NAME.getName(), TOPIC_NAME,
"-" + WorkflowExecutionArgs.STATUS.getName(), "SUCCEEDED",
"-" + WorkflowExecutionArgs.BRKR_TTL.getName(), "10",
"-" + WorkflowExecutionArgs.CLUSTER_NAME.getName(), "corp",
};
broker = new BrokerService();
broker.addConnector(BROKER_URL);
broker.setDataDirectory("target/activemq");
broker.start();
}
@AfterClass
public void tearDown() throws Exception {
broker.deleteAllMessages();
broker.stop();
this.dfsCluster.shutdown();
}
@Test
public void testLogFile() throws Exception {
FileSystem fs = dfsCluster.getFileSystem();
OutputStream out = fs.create(logFile);
InputStream in = new ByteArrayInputStream(
("instancePaths=/falcon/feed/agg-logs/path1/2010/10/10/20,"
+ "/falcon/feed/agg-logs/path1/2010/10/10/21,"
+ "/falcon/feed/agg-logs/path1/2010/10/10/22,"
+ "/falcon/feed/agg-logs/path1/2010/10/10/23")
.getBytes());
IOUtils.copyBytes(in, out, conf);
testProcessMessageCreator();
}
@Test
public void testEmptyLogFile() throws Exception {
FileSystem fs = dfsCluster.getFileSystem();
OutputStream out = fs.create(logFile);
InputStream in = new ByteArrayInputStream(("instancePaths=").getBytes());
IOUtils.copyBytes(in, out, conf);
WorkflowExecutionContext context = WorkflowExecutionContext.create(
args, WorkflowExecutionContext.Type.POST_PROCESSING);
JMSMessageProducer jmsMessageProducer = JMSMessageProducer.builder(context)
.type(JMSMessageProducer.MessageType.USER).build();
jmsMessageProducer.sendMessage();
}
private void testProcessMessageCreator() throws Exception {
Thread t = new Thread() {
@Override
public void run() {
try {
consumer();
} catch (AssertionError e) {
error = e;
} catch (JMSException ignore) {
error = null;
}
}
};
t.start();
Thread.sleep(100);
WorkflowExecutionContext context = WorkflowExecutionContext.create(
args, WorkflowExecutionContext.Type.POST_PROCESSING);
JMSMessageProducer jmsMessageProducer = JMSMessageProducer.builder(context)
.type(JMSMessageProducer.MessageType.USER).build();
jmsMessageProducer.sendMessage();
t.join();
if (error != null) {
throw error;
}
}
private void consumer() throws JMSException {
ConnectionFactory connectionFactory = new ActiveMQConnectionFactory(BROKER_URL);
Connection connection = connectionFactory.createConnection();
connection.start();
Session session = connection.createSession(false, Session.AUTO_ACKNOWLEDGE);
Destination destination = session.createTopic(TOPIC_NAME);
MessageConsumer consumer = session.createConsumer(destination);
// wait till you get atleast one message
MapMessage m;
for (m = null; m == null;) {
m = (MapMessage) consumer.receive();
}
System.out.println("Consumed: " + m.toString());
assertMessage(m);
Assert.assertEquals(m.getString(WorkflowExecutionArgs.FEED_INSTANCE_PATHS.getName()),
"/falcon/feed/agg-logs/path1/2010/10/10/20");
for (m = null; m == null;) {
m = (MapMessage) consumer.receive();
}
System.out.println("Consumed: " + m.toString());
assertMessage(m);
Assert.assertEquals(m.getString(WorkflowExecutionArgs.FEED_INSTANCE_PATHS.getName()),
"/falcon/feed/agg-logs/path1/2010/10/10/21");
for (m = null; m == null;) {
m = (MapMessage) consumer.receive();
}
System.out.println("Consumed: " + m.toString());
assertMessage(m);
Assert.assertEquals(m.getString(WorkflowExecutionArgs.FEED_INSTANCE_PATHS.getName()),
"/falcon/feed/agg-logs/path1/2010/10/10/22");
for (m = null; m == null;) {
m = (MapMessage) consumer.receive();
}
System.out.println("Consumed: " + m.toString());
assertMessage(m);
Assert.assertEquals(m.getString(WorkflowExecutionArgs.FEED_INSTANCE_PATHS.getName()),
"/falcon/feed/agg-logs/path1/2010/10/10/23");
connection.close();
}
private void assertMessage(MapMessage m) throws JMSException {
Assert.assertEquals(m.getString(WorkflowExecutionArgs.ENTITY_NAME.getName()),
TOPIC_NAME);
Assert.assertEquals(m.getString(WorkflowExecutionArgs.OPERATION.getName()), "DELETE");
Assert.assertEquals(m.getString(WorkflowExecutionArgs.WORKFLOW_ID.getName()),
"workflow-01-00");
Assert.assertEquals(m.getString(WorkflowExecutionArgs.WORKFLOW_USER.getName()),
"falcon");
Assert.assertEquals(m.getString(WorkflowExecutionArgs.RUN_ID.getName()), "1");
Assert.assertEquals(m.getString(WorkflowExecutionArgs.NOMINAL_TIME.getName()),
"2011-01-01-01-00");
Assert.assertEquals(m.getString(WorkflowExecutionArgs.TIMESTAMP.getName()),
"2012-01-01-01-00");
Assert.assertEquals(m.getString(WorkflowExecutionArgs.STATUS.getName()), "SUCCEEDED");
}
}