blob: b045b4fced35cdd96ac8ff4b6721359c48d9cd74 [file] [log] [blame]
<!DOCTYPE html>
<!--[if lt IE 7]> <html class="no-js lt-ie9 lt-ie8 lt-ie7"> <![endif]-->
<!--[if IE 7]> <html class="no-js lt-ie9 lt-ie8"> <![endif]-->
<!--[if IE 8]> <html class="no-js lt-ie9"> <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js"> <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta http-equiv="X-UA-Compatible" content="IE=edge,chrome=1">
<title>Spark Streaming + Flume Integration Guide - Spark 1.6.2 Documentation</title>
<link rel="stylesheet" href="css/bootstrap.min.css">
<style>
body {
padding-top: 60px;
padding-bottom: 40px;
}
</style>
<meta name="viewport" content="width=device-width">
<link rel="stylesheet" href="css/bootstrap-responsive.min.css">
<link rel="stylesheet" href="css/main.css">
<script src="js/vendor/modernizr-2.6.1-respond-1.1.0.min.js"></script>
<link rel="stylesheet" href="css/pygments-default.css">
<!-- Google analytics script -->
<script type="text/javascript">
var _gaq = _gaq || [];
_gaq.push(['_setAccount', 'UA-32518208-2']);
_gaq.push(['_trackPageview']);
(function() {
var ga = document.createElement('script'); ga.type = 'text/javascript'; ga.async = true;
ga.src = ('https:' == document.location.protocol ? 'https://ssl' : 'http://www') + '.google-analytics.com/ga.js';
var s = document.getElementsByTagName('script')[0]; s.parentNode.insertBefore(ga, s);
})();
</script>
</head>
<body>
<!--[if lt IE 7]>
<p class="chromeframe">You are using an outdated browser. <a href="http://browsehappy.com/">Upgrade your browser today</a> or <a href="http://www.google.com/chromeframe/?redirect=true">install Google Chrome Frame</a> to better experience this site.</p>
<![endif]-->
<!-- This code is taken from http://twitter.github.com/bootstrap/examples/hero.html -->
<div class="navbar navbar-fixed-top" id="topbar">
<div class="navbar-inner">
<div class="container">
<div class="brand"><a href="index.html">
<img src="img/spark-logo-hd.png" style="height:50px;"/></a><span class="version">1.6.2</span>
</div>
<ul class="nav">
<!--TODO(andyk): Add class="active" attribute to li some how.-->
<li><a href="index.html">Overview</a></li>
<li class="dropdown">
<a href="#" class="dropdown-toggle" data-toggle="dropdown">Programming Guides<b class="caret"></b></a>
<ul class="dropdown-menu">
<li><a href="quick-start.html">Quick Start</a></li>
<li><a href="programming-guide.html">Spark Programming Guide</a></li>
<li class="divider"></li>
<li><a href="streaming-programming-guide.html">Spark Streaming</a></li>
<li><a href="sql-programming-guide.html">DataFrames, Datasets and SQL</a></li>
<li><a href="mllib-guide.html">MLlib (Machine Learning)</a></li>
<li><a href="graphx-programming-guide.html">GraphX (Graph Processing)</a></li>
<li><a href="bagel-programming-guide.html">Bagel (Pregel on Spark)</a></li>
<li><a href="sparkr.html">SparkR (R on Spark)</a></li>
</ul>
</li>
<li class="dropdown">
<a href="#" class="dropdown-toggle" data-toggle="dropdown">API Docs<b class="caret"></b></a>
<ul class="dropdown-menu">
<li><a href="api/scala/index.html#org.apache.spark.package">Scala</a></li>
<li><a href="api/java/index.html">Java</a></li>
<li><a href="api/python/index.html">Python</a></li>
<li><a href="api/R/index.html">R</a></li>
</ul>
</li>
<li class="dropdown">
<a href="#" class="dropdown-toggle" data-toggle="dropdown">Deploying<b class="caret"></b></a>
<ul class="dropdown-menu">
<li><a href="cluster-overview.html">Overview</a></li>
<li><a href="submitting-applications.html">Submitting Applications</a></li>
<li class="divider"></li>
<li><a href="spark-standalone.html">Spark Standalone</a></li>
<li><a href="running-on-mesos.html">Mesos</a></li>
<li><a href="running-on-yarn.html">YARN</a></li>
<li class="divider"></li>
<li><a href="ec2-scripts.html">Amazon EC2</a></li>
</ul>
</li>
<li class="dropdown">
<a href="api.html" class="dropdown-toggle" data-toggle="dropdown">More<b class="caret"></b></a>
<ul class="dropdown-menu">
<li><a href="configuration.html">Configuration</a></li>
<li><a href="monitoring.html">Monitoring</a></li>
<li><a href="tuning.html">Tuning Guide</a></li>
<li><a href="job-scheduling.html">Job Scheduling</a></li>
<li><a href="security.html">Security</a></li>
<li><a href="hardware-provisioning.html">Hardware Provisioning</a></li>
<li class="divider"></li>
<li><a href="building-spark.html">Building Spark</a></li>
<li><a href="https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark">Contributing to Spark</a></li>
<li><a href="https://cwiki.apache.org/confluence/display/SPARK/Supplemental+Spark+Projects">Supplemental Projects</a></li>
</ul>
</li>
</ul>
<!--<p class="navbar-text pull-right"><span class="version-text">v1.6.2</span></p>-->
</div>
</div>
</div>
<div class="container-wrapper">
<div class="content" id="content">
<h1 class="title">Spark Streaming + Flume Integration Guide</h1>
<p><a href="https://flume.apache.org/">Apache Flume</a> is a distributed, reliable, and available service for efficiently collecting, aggregating, and moving large amounts of log data. Here we explain how to configure Flume and Spark Streaming to receive data from Flume. There are two approaches to this.</p>
<h2 id="approach-1-flume-style-push-based-approach">Approach 1: Flume-style Push-based Approach</h2>
<p>Flume is designed to push data between Flume agents. In this approach, Spark Streaming essentially sets up a receiver that acts an Avro agent for Flume, to which Flume can push the data. Here are the configuration steps.</p>
<h4 id="general-requirements">General Requirements</h4>
<p>Choose a machine in your cluster such that</p>
<ul>
<li>
<p>When your Flume + Spark Streaming application is launched, one of the Spark workers must run on that machine.</p>
</li>
<li>
<p>Flume can be configured to push data to a port on that machine.</p>
</li>
</ul>
<p>Due to the push model, the streaming application needs to be up, with the receiver scheduled and listening on the chosen port, for Flume to be able push data.</p>
<h4 id="configuring-flume">Configuring Flume</h4>
<p>Configure Flume agent to send data to an Avro sink by having the following in the configuration file.</p>
<pre><code>agent.sinks = avroSink
agent.sinks.avroSink.type = avro
agent.sinks.avroSink.channel = memoryChannel
agent.sinks.avroSink.hostname = &lt;chosen machine's hostname&gt;
agent.sinks.avroSink.port = &lt;chosen port on the machine&gt;
</code></pre>
<p>See the <a href="https://flume.apache.org/documentation.html">Flume&#8217;s documentation</a> for more information about
configuring Flume agents.</p>
<h4 id="configuring-spark-streaming-application">Configuring Spark Streaming Application</h4>
<ol>
<li>
<p><strong>Linking:</strong> In your SBT/Maven projrect definition, link your streaming application against the following artifact (see <a href="streaming-programming-guide.html#linking">Linking section</a> in the main programming guide for further information).</p>
<pre><code> groupId = org.apache.spark
artifactId = spark-streaming-flume_2.10
version = 1.6.2
</code></pre>
</li>
<li>
<p><strong>Programming:</strong> In the streaming application code, import <code>FlumeUtils</code> and create input DStream as follows.</p>
<div class="codetabs">
<div data-lang="scala">
<pre><code> import org.apache.spark.streaming.flume._
val flumeStream = FlumeUtils.createStream(streamingContext, [chosen machine's hostname], [chosen port])
</code></pre>
<p>See the <a href="api/scala/index.html#org.apache.spark.streaming.flume.FlumeUtils$">API docs</a>
and the <a href="https://github.com/apache/spark/tree/master/examples/src/main/scala/org/apache/spark/examples/streaming/FlumeEventCount.scala">example</a>.</p>
</div>
<div data-lang="java">
<pre><code> import org.apache.spark.streaming.flume.*;
JavaReceiverInputDStream&lt;SparkFlumeEvent&gt; flumeStream =
FlumeUtils.createStream(streamingContext, [chosen machine's hostname], [chosen port]);
</code></pre>
<p>See the <a href="api/java/index.html?org/apache/spark/streaming/flume/FlumeUtils.html">API docs</a>
and the <a href="https://github.com/apache/spark/tree/master/examples/src/main/java/org/apache/spark/examples/streaming/JavaFlumeEventCount.java">example</a>.</p>
</div>
<div data-lang="python">
<pre><code> from pyspark.streaming.flume import FlumeUtils
flumeStream = FlumeUtils.createStream(streamingContext, [chosen machine's hostname], [chosen port])
</code></pre>
<p>By default, the Python API will decode Flume event body as UTF8 encoded strings. You can specify your custom decoding function to decode the body byte arrays in Flume events to any arbitrary data type.
See the <a href="api/python/pyspark.streaming.html#pyspark.streaming.flume.FlumeUtils">API docs</a>
and the <a href="https://github.com/apache/spark/blob/master/examples/src/main/python/streaming/flume_wordcount.py">example</a>.</p>
</div>
</div>
<p>Note that the hostname should be the same as the one used by the resource manager in the
cluster (Mesos, YARN or Spark Standalone), so that resource allocation can match the names and launch
the receiver in the right machine.</p>
</li>
<li>
<p><strong>Deploying:</strong> As with any Spark applications, <code>spark-submit</code> is used to launch your application. However, the details are slightly different for Scala/Java applications and Python applications.</p>
<p>For Scala and Java applications, if you are using SBT or Maven for project management, then package <code>spark-streaming-flume_2.10</code> and its dependencies into the application JAR. Make sure <code>spark-core_2.10</code> and <code>spark-streaming_2.10</code> are marked as <code>provided</code> dependencies as those are already present in a Spark installation. Then use <code>spark-submit</code> to launch your application (see <a href="streaming-programming-guide.html#deploying-applications">Deploying section</a> in the main programming guide).</p>
<p>For Python applications which lack SBT/Maven project management, <code>spark-streaming-flume_2.10</code> and its dependencies can be directly added to <code>spark-submit</code> using <code>--packages</code> (see <a href="submitting-applications.html">Application Submission Guide</a>). That is,</p>
<pre><code> ./bin/spark-submit --packages org.apache.spark:spark-streaming-flume_2.10:1.6.2 ...
</code></pre>
<p>Alternatively, you can also download the JAR of the Maven artifact <code>spark-streaming-flume-assembly</code> from the
<a href="http://search.maven.org/#search|ga|1|a%3A%22spark-streaming-flume-assembly_2.10%22%20AND%20v%3A%221.6.2%22">Maven repository</a> and add it to <code>spark-submit</code> with <code>--jars</code>.</p>
</li>
</ol>
<h2 id="approach-2-pull-based-approach-using-a-custom-sink">Approach 2: Pull-based Approach using a Custom Sink</h2>
<p>Instead of Flume pushing data directly to Spark Streaming, this approach runs a custom Flume sink that allows the following.</p>
<ul>
<li>Flume pushes data into the sink, and the data stays buffered.</li>
<li>Spark Streaming uses a <a href="streaming-programming-guide.html#receiver-reliability">reliable Flume receiver</a>
and transactions to pull data from the sink. Transactions succeed only after data is received and
replicated by Spark Streaming.</li>
</ul>
<p>This ensures stronger reliability and
<a href="streaming-programming-guide.html#fault-tolerance-semantics">fault-tolerance guarantees</a>
than the previous approach. However, this requires configuring Flume to run a custom sink.
Here are the configuration steps.</p>
<h4 id="general-requirements-1">General Requirements</h4>
<p>Choose a machine that will run the custom sink in a Flume agent. The rest of the Flume pipeline is configured to send data to that agent. Machines in the Spark cluster should have access to the chosen machine running the custom sink.</p>
<h4 id="configuring-flume-1">Configuring Flume</h4>
<p>Configuring Flume on the chosen machine requires the following two steps.</p>
<ol>
<li>
<p><strong>Sink JARs</strong>: Add the following JARs to Flume&#8217;s classpath (see <a href="https://flume.apache.org/documentation.html">Flume&#8217;s documentation</a> to see how) in the machine designated to run the custom sink .</p>
<p>(i) <em>Custom sink JAR</em>: Download the JAR corresponding to the following artifact (or <a href="http://search.maven.org/remotecontent?filepath=org/apache/spark/spark-streaming-flume-sink_2.10/1.6.2/spark-streaming-flume-sink_2.10-1.6.2.jar">direct link</a>).</p>
<pre><code> groupId = org.apache.spark
artifactId = spark-streaming-flume-sink_2.10
version = 1.6.2
</code></pre>
<p>(ii) <em>Scala library JAR</em>: Download the Scala library JAR for Scala 2.10.5. It can be found with the following artifact detail (or, <a href="http://search.maven.org/remotecontent?filepath=org/scala-lang/scala-library/2.10.5/scala-library-2.10.5.jar">direct link</a>).</p>
<pre><code> groupId = org.scala-lang
artifactId = scala-library
version = 2.10.5
</code></pre>
<p>(iii) <em>Commons Lang 3 JAR</em>: Download the Commons Lang 3 JAR. It can be found with the following artifact detail (or, <a href="http://search.maven.org/remotecontent?filepath=org/apache/commons/commons-lang3/3.3.2/commons-lang3-3.3.2.jar">direct link</a>).</p>
<pre><code> groupId = org.apache.commons
artifactId = commons-lang3
version = 3.3.2
</code></pre>
</li>
<li>
<p><strong>Configuration file</strong>: On that machine, configure Flume agent to send data to an Avro sink by having the following in the configuration file.</p>
<pre><code> agent.sinks = spark
agent.sinks.spark.type = org.apache.spark.streaming.flume.sink.SparkSink
agent.sinks.spark.hostname = &lt;hostname of the local machine&gt;
agent.sinks.spark.port = &lt;port to listen on for connection from Spark&gt;
agent.sinks.spark.channel = memoryChannel
</code></pre>
<p>Also make sure that the upstream Flume pipeline is configured to send the data to the Flume agent running this sink.</p>
</li>
</ol>
<p>See the <a href="https://flume.apache.org/documentation.html">Flume&#8217;s documentation</a> for more information about
configuring Flume agents.</p>
<h4 id="configuring-spark-streaming-application-1">Configuring Spark Streaming Application</h4>
<ol>
<li>
<p><strong>Linking:</strong> In your SBT/Maven project definition, link your streaming application against the <code>spark-streaming-flume_2.10</code> (see <a href="streaming-programming-guide.html#linking">Linking section</a> in the main programming guide).</p>
</li>
<li>
<p><strong>Programming:</strong> In the streaming application code, import <code>FlumeUtils</code> and create input DStream as follows.</p>
<div class="codetabs">
<div data-lang="scala">
<pre><code> import org.apache.spark.streaming.flume._
val flumeStream = FlumeUtils.createPollingStream(streamingContext, [sink machine hostname], [sink port])
</code></pre>
</div>
<div data-lang="java">
<pre><code> import org.apache.spark.streaming.flume.*;
JavaReceiverInputDStream&lt;SparkFlumeEvent&gt;flumeStream =
FlumeUtils.createPollingStream(streamingContext, [sink machine hostname], [sink port]);
</code></pre>
</div>
<div data-lang="python">
<pre><code> from pyspark.streaming.flume import FlumeUtils
addresses = [([sink machine hostname 1], [sink port 1]), ([sink machine hostname 2], [sink port 2])]
flumeStream = FlumeUtils.createPollingStream(streamingContext, addresses)
</code></pre>
<p>By default, the Python API will decode Flume event body as UTF8 encoded strings. You can specify your custom decoding function to decode the body byte arrays in Flume events to any arbitrary data type.
See the <a href="api/python/pyspark.streaming.html#pyspark.streaming.flume.FlumeUtils">API docs</a>.</p>
</div>
</div>
<p>See the Scala example <a href="https://github.com/apache/spark/tree/master/examples/src/main/scala/org/apache/spark/examples/streaming/FlumePollingEventCount.scala">FlumePollingEventCount</a>.</p>
<p>Note that each input DStream can be configured to receive data from multiple sinks.</p>
</li>
<li>
<p><strong>Deploying:</strong> This is same as the first approach.</p>
</li>
</ol>
</div>
<!-- /container -->
</div>
<script src="js/vendor/jquery-1.8.0.min.js"></script>
<script src="js/vendor/bootstrap.min.js"></script>
<script src="js/vendor/anchor.min.js"></script>
<script src="js/main.js"></script>
<!-- MathJax Section -->
<script type="text/x-mathjax-config">
MathJax.Hub.Config({
TeX: { equationNumbers: { autoNumber: "AMS" } }
});
</script>
<script>
// Note that we load MathJax this way to work with local file (file://), HTTP and HTTPS.
// We could use "//cdn.mathjax...", but that won't support "file://".
(function(d, script) {
script = d.createElement('script');
script.type = 'text/javascript';
script.async = true;
script.onload = function(){
MathJax.Hub.Config({
tex2jax: {
inlineMath: [ ["$", "$"], ["\\\\(","\\\\)"] ],
displayMath: [ ["$$","$$"], ["\\[", "\\]"] ],
processEscapes: true,
skipTags: ['script', 'noscript', 'style', 'textarea', 'pre']
}
});
};
script.src = ('https:' == document.location.protocol ? 'https://' : 'http://') +
'cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML';
d.getElementsByTagName('head')[0].appendChild(script);
}(document));
</script>
</body>
</html>