blob: 232d1f01b93dc68c2e4cd8f64c902da9b6251e38 [file] [log] [blame]
<!DOCTYPE html >
<html>
<head>
<meta http-equiv="X-UA-Compatible" content="IE=edge" />
<meta name="viewport" content="width=device-width, initial-scale=1.0, maximum-scale=1.0, user-scalable=no" />
<title>Spark 3.3.4 ScalaDoc - org.apache.spark.SparkContext</title>
<meta name="description" content="Spark 3.3.4 ScalaDoc - org.apache.spark.SparkContext" />
<meta name="keywords" content="Spark 3.3.4 ScalaDoc org.apache.spark.SparkContext" />
<meta http-equiv="content-type" content="text/html; charset=UTF-8" />
<link href="../../../lib/index.css" media="screen" type="text/css" rel="stylesheet" />
<link href="../../../lib/template.css" media="screen" type="text/css" rel="stylesheet" />
<link href="../../../lib/diagrams.css" media="screen" type="text/css" rel="stylesheet" id="diagrams-css" />
<script type="text/javascript" src="../../../lib/jquery.min.js"></script>
<script type="text/javascript" src="../../../lib/jquery.panzoom.min.js"></script>
<script type="text/javascript" src="../../../lib/jquery.mousewheel.min.js"></script>
<script type="text/javascript" src="../../../lib/index.js"></script>
<script type="text/javascript" src="../../../index.js"></script>
<script type="text/javascript" src="../../../lib/scheduler.js"></script>
<script type="text/javascript" src="../../../lib/template.js"></script>
<script type="text/javascript">
/* this variable can be used by the JS to determine the path to the root document */
var toRoot = '../../../';
</script>
</head>
<body>
<div id="search">
<span id="doc-title">Spark 3.3.4 ScalaDoc<span id="doc-version"></span></span>
<span class="close-results"><span class="left">&lt;</span> Back</span>
<div id="textfilter">
<span class="input">
<input autocapitalize="none" placeholder="Search" id="index-input" type="text" accesskey="/" />
<i class="clear material-icons"></i>
<i id="search-icon" class="material-icons"></i>
</span>
</div>
</div>
<div id="search-results">
<div id="search-progress">
<div id="progress-fill"></div>
</div>
<div id="results-content">
<div id="entity-results"></div>
<div id="member-results"></div>
</div>
</div>
<div id="content-scroll-container" style="-webkit-overflow-scrolling: touch;">
<div id="content-container" style="-webkit-overflow-scrolling: touch;">
<div id="subpackage-spacer">
<div id="packages">
<h1>Packages</h1>
<ul>
<li name="_root_.root" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="_root_"></a><a id="root:_root_"></a>
<span class="permalink">
<a href="../../../index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="../../../index.html"><span class="name">root</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="../../../index.html" class="extype" name="_root_">root</a></dd></dl></div>
</li><li name="_root_.org" visbl="pub" class="indented1 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="org"></a><a id="org:org"></a>
<span class="permalink">
<a href="../../../org/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="../../index.html"><span class="name">org</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="../../../index.html" class="extype" name="_root_">root</a></dd></dl></div>
</li><li name="org.apache" visbl="pub" class="indented2 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="apache"></a><a id="apache:apache"></a>
<span class="permalink">
<a href="../../../org/apache/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="../index.html"><span class="name">apache</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="../../index.html" class="extype" name="org">org</a></dd></dl></div>
</li><li name="org.apache.spark" visbl="pub" class="indented3 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="spark"></a><a id="spark:spark"></a>
<span class="permalink">
<a href="../../../org/apache/spark/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="Core Spark functionality." href="index.html"><span class="name">spark</span></a>
</span>
<p class="shortcomment cmt">Core Spark functionality.</p><div class="fullcomment"><div class="comment cmt"><p>Core Spark functionality. <a href="" class="extype" name="org.apache.spark.SparkContext">org.apache.spark.SparkContext</a> serves as the main entry point to
Spark, while <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">org.apache.spark.rdd.RDD</a> is the data type representing a distributed collection,
and provides most parallel operations.</p><p>In addition, <a href="rdd/PairRDDFunctions.html" class="extype" name="org.apache.spark.rdd.PairRDDFunctions">org.apache.spark.rdd.PairRDDFunctions</a> contains operations available only on RDDs
of key-value pairs, such as <code>groupByKey</code> and <code>join</code>; <a href="rdd/DoubleRDDFunctions.html" class="extype" name="org.apache.spark.rdd.DoubleRDDFunctions">org.apache.spark.rdd.DoubleRDDFunctions</a>
contains operations available only on RDDs of Doubles; and
<a href="rdd/SequenceFileRDDFunctions.html" class="extype" name="org.apache.spark.rdd.SequenceFileRDDFunctions">org.apache.spark.rdd.SequenceFileRDDFunctions</a> contains operations available on RDDs that can
be saved as SequenceFiles. These operations are automatically available on any RDD of the right
type (e.g. RDD[(Int, Int)] through implicit conversions.</p><p>Java programmers should reference the <a href="api/java/index.html" class="extype" name="org.apache.spark.api.java">org.apache.spark.api.java</a> package
for Spark programming APIs in Java.</p><p>Classes and methods marked with <span class="experimental badge" style="float: none;">
Experimental</span> are user-facing features which have not been officially adopted by the
Spark project. These are subject to change or removal in minor releases.</p><p>Classes and methods marked with <span class="developer badge" style="float: none;">
Developer API</span> are intended for advanced users want to extend Spark through lower
level interfaces. These are subject to changes or removal in minor releases.
</p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="../index.html" class="extype" name="org.apache">apache</a></dd></dl></div>
</li><li name="org.apache.spark.api" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="api"></a><a id="api:api"></a>
<span class="permalink">
<a href="../../../org/apache/spark/api/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="api/index.html"><span class="name">api</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.broadcast" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="broadcast"></a><a id="broadcast:broadcast"></a>
<span class="permalink">
<a href="../../../org/apache/spark/broadcast/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="Spark's broadcast variables, used to broadcast immutable datasets to all nodes." href="broadcast/index.html"><span class="name">broadcast</span></a>
</span>
<p class="shortcomment cmt">Spark's broadcast variables, used to broadcast immutable datasets to all nodes.</p><div class="fullcomment"><div class="comment cmt"><p>Spark's broadcast variables, used to broadcast immutable datasets to all nodes.
</p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.graphx" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="graphx"></a><a id="graphx:graphx"></a>
<span class="permalink">
<a href="../../../org/apache/spark/graphx/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="ALPHA COMPONENT GraphX is a graph processing framework built on top of Spark." href="graphx/index.html"><span class="name">graphx</span></a>
</span>
<p class="shortcomment cmt"><span class="badge" style="float: right;">ALPHA COMPONENT</span>
GraphX is a graph processing framework built on top of Spark.</p><div class="fullcomment"><div class="comment cmt"><p><span class="badge" style="float: right;">ALPHA COMPONENT</span>
GraphX is a graph processing framework built on top of Spark.
</p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.input" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="input"></a><a id="input:input"></a>
<span class="permalink">
<a href="../../../org/apache/spark/input/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="input/index.html"><span class="name">input</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.io" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="io"></a><a id="io:io"></a>
<span class="permalink">
<a href="../../../org/apache/spark/io/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="IO codecs used for compression." href="io/index.html"><span class="name">io</span></a>
</span>
<p class="shortcomment cmt">IO codecs used for compression.</p><div class="fullcomment"><div class="comment cmt"><p>IO codecs used for compression. See <a href="io/CompressionCodec.html" class="extype" name="org.apache.spark.io.CompressionCodec">org.apache.spark.io.CompressionCodec</a>.
</p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.launcher" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="launcher"></a><a id="launcher:launcher"></a>
<span class="permalink">
<a href="../../../org/apache/spark/launcher/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="launcher/index.html"><span class="name">launcher</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.mapred" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="mapred"></a><a id="mapred:mapred"></a>
<span class="permalink">
<a href="../../../org/apache/spark/mapred/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="mapred/index.html"><span class="name">mapred</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.metrics" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="metrics"></a><a id="metrics:metrics"></a>
<span class="permalink">
<a href="../../../org/apache/spark/metrics/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="metrics/index.html"><span class="name">metrics</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.ml" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="ml"></a><a id="ml:ml"></a>
<span class="permalink">
<a href="../../../org/apache/spark/ml/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="DataFrame-based machine learning APIs to let users quickly assemble and configure practical machine learning pipelines." href="ml/index.html"><span class="name">ml</span></a>
</span>
<p class="shortcomment cmt">DataFrame-based machine learning APIs to let users quickly assemble and configure practical
machine learning pipelines.</p><div class="fullcomment"><div class="comment cmt"><p>DataFrame-based machine learning APIs to let users quickly assemble and configure practical
machine learning pipelines.
</p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.mllib" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="mllib"></a><a id="mllib:mllib"></a>
<span class="permalink">
<a href="../../../org/apache/spark/mllib/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="RDD-based machine learning APIs (in maintenance mode)." href="mllib/index.html"><span class="name">mllib</span></a>
</span>
<p class="shortcomment cmt">RDD-based machine learning APIs (in maintenance mode).</p><div class="fullcomment"><div class="comment cmt"><p>RDD-based machine learning APIs (in maintenance mode).</p><p>The <code>spark.mllib</code> package is in maintenance mode as of the Spark 2.0.0 release to encourage
migration to the DataFrame-based APIs under the <a href="ml/index.html" class="extype" name="org.apache.spark.ml">org.apache.spark.ml</a> package.
While in maintenance mode,</p><ul><li>no new features in the RDD-based <code>spark.mllib</code> package will be accepted, unless they block
implementing new features in the DataFrame-based <code>spark.ml</code> package;</li><li>bug fixes in the RDD-based APIs will still be accepted.</li></ul><p>The developers will continue adding more features to the DataFrame-based APIs in the 2.x series
to reach feature parity with the RDD-based APIs.
And once we reach feature parity, this package will be deprecated.
</p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd><dt>See also</dt><dd><span class="cmt"><p><a href="https://issues.apache.org/jira/browse/SPARK-4591">SPARK-4591</a> to track
the progress of feature parity</p></span></dd></dl></div>
</li><li name="org.apache.spark.partial" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="partial"></a><a id="partial:partial"></a>
<span class="permalink">
<a href="../../../org/apache/spark/partial/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="Support for approximate results." href="partial/index.html"><span class="name">partial</span></a>
</span>
<p class="shortcomment cmt">Support for approximate results.</p><div class="fullcomment"><div class="comment cmt"><p>Support for approximate results. This provides convenient api and also implementation for
approximate calculation.
</p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd><dt>See also</dt><dd><span class="cmt"><p><a href="rdd/RDD.html#countApprox(timeout:Long,confidence:Double):org.apache.spark.partial.PartialResult[org.apache.spark.partial.BoundedDouble]" class="extmbr" name="org.apache.spark.rdd.RDD#countApprox">org.apache.spark.rdd.RDD.countApprox</a></p></span></dd></dl></div>
</li><li name="org.apache.spark.rdd" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="rdd"></a><a id="rdd:rdd"></a>
<span class="permalink">
<a href="../../../org/apache/spark/rdd/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="Provides several RDD implementations." href="rdd/index.html"><span class="name">rdd</span></a>
</span>
<p class="shortcomment cmt">Provides several RDD implementations.</p><div class="fullcomment"><div class="comment cmt"><p>Provides several RDD implementations. See <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">org.apache.spark.rdd.RDD</a>.
</p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.resource" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="resource"></a><a id="resource:resource"></a>
<span class="permalink">
<a href="../../../org/apache/spark/resource/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="resource/index.html"><span class="name">resource</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.scheduler" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="scheduler"></a><a id="scheduler:scheduler"></a>
<span class="permalink">
<a href="../../../org/apache/spark/scheduler/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="Spark's scheduling components." href="scheduler/index.html"><span class="name">scheduler</span></a>
</span>
<p class="shortcomment cmt">Spark's scheduling components.</p><div class="fullcomment"><div class="comment cmt"><p>Spark's scheduling components. This includes the <code>org.apache.spark.scheduler.DAGScheduler</code> and
lower level <code>org.apache.spark.scheduler.TaskScheduler</code>.
</p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.security" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="security"></a><a id="security:security"></a>
<span class="permalink">
<a href="../../../org/apache/spark/security/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="security/index.html"><span class="name">security</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.serializer" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="serializer"></a><a id="serializer:serializer"></a>
<span class="permalink">
<a href="../../../org/apache/spark/serializer/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="Pluggable serializers for RDD and shuffle data." href="serializer/index.html"><span class="name">serializer</span></a>
</span>
<p class="shortcomment cmt">Pluggable serializers for RDD and shuffle data.</p><div class="fullcomment"><div class="comment cmt"><p>Pluggable serializers for RDD and shuffle data.
</p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd><dt>See also</dt><dd><span class="cmt"><p><a href="serializer/Serializer.html" class="extype" name="org.apache.spark.serializer.Serializer">org.apache.spark.serializer.Serializer</a></p></span></dd></dl></div>
</li><li name="org.apache.spark.shuffle" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="shuffle"></a><a id="shuffle:shuffle"></a>
<span class="permalink">
<a href="../../../org/apache/spark/shuffle/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="shuffle/index.html"><span class="name">shuffle</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.sql" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="sql"></a><a id="sql:sql"></a>
<span class="permalink">
<a href="../../../org/apache/spark/sql/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="Allows the execution of relational queries, including those expressed in SQL using Spark." href="sql/index.html"><span class="name">sql</span></a>
</span>
<p class="shortcomment cmt">Allows the execution of relational queries, including those expressed in SQL using Spark.</p><div class="fullcomment"><div class="comment cmt"><p>Allows the execution of relational queries, including those expressed in SQL using Spark.
</p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.status" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="status"></a><a id="status:status"></a>
<span class="permalink">
<a href="../../../org/apache/spark/status/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="status/index.html"><span class="name">status</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.storage" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="storage"></a><a id="storage:storage"></a>
<span class="permalink">
<a href="../../../org/apache/spark/storage/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="storage/index.html"><span class="name">storage</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.streaming" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="streaming"></a><a id="streaming:streaming"></a>
<span class="permalink">
<a href="../../../org/apache/spark/streaming/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="Spark Streaming functionality." href="streaming/index.html"><span class="name">streaming</span></a>
</span>
<p class="shortcomment cmt">Spark Streaming functionality.</p><div class="fullcomment"><div class="comment cmt"><p>Spark Streaming functionality. <a href="streaming/StreamingContext.html" class="extype" name="org.apache.spark.streaming.StreamingContext">org.apache.spark.streaming.StreamingContext</a> serves as the main
entry point to Spark Streaming, while <a href="streaming/dstream/DStream.html" class="extype" name="org.apache.spark.streaming.dstream.DStream">org.apache.spark.streaming.dstream.DStream</a> is the data
type representing a continuous sequence of RDDs, representing a continuous stream of data.</p><p>In addition, <a href="streaming/dstream/PairDStreamFunctions.html" class="extype" name="org.apache.spark.streaming.dstream.PairDStreamFunctions">org.apache.spark.streaming.dstream.PairDStreamFunctions</a> contains operations
available only on DStreams
of key-value pairs, such as <code>groupByKey</code> and <code>reduceByKey</code>. These operations are automatically
available on any DStream of the right type (e.g. DStream[(Int, Int)] through implicit
conversions.</p><p>For the Java API of Spark Streaming, take a look at the
<a href="streaming/api/java/JavaStreamingContext.html" class="extype" name="org.apache.spark.streaming.api.java.JavaStreamingContext">org.apache.spark.streaming.api.java.JavaStreamingContext</a> which serves as the entry point, and
the <a href="streaming/api/java/JavaDStream.html" class="extype" name="org.apache.spark.streaming.api.java.JavaDStream">org.apache.spark.streaming.api.java.JavaDStream</a> and the
<a href="streaming/api/java/JavaPairDStream.html" class="extype" name="org.apache.spark.streaming.api.java.JavaPairDStream">org.apache.spark.streaming.api.java.JavaPairDStream</a> which have the DStream functionality.
</p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.unsafe" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="unsafe"></a><a id="unsafe:unsafe"></a>
<span class="permalink">
<a href="../../../org/apache/spark/unsafe/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="" href="unsafe/index.html"><span class="name">unsafe</span></a>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li name="org.apache.spark.util" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="util"></a><a id="util:util"></a>
<span class="permalink">
<a href="../../../org/apache/spark/util/index.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">package</span>
</span>
<span class="symbol">
<a title="Spark utilities." href="util/index.html"><span class="name">util</span></a>
</span>
<p class="shortcomment cmt">Spark utilities.</p><div class="fullcomment"><div class="comment cmt"><p>Spark utilities.
</p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="Aggregator.html" title=":: DeveloperApi :: A set of functions used to aggregate data."></a>
<a href="Aggregator.html" title=":: DeveloperApi :: A set of functions used to aggregate data.">Aggregator</a>
</li><li class="current-entities indented3">
<a class="object" href="BarrierTaskContext$.html" title=""></a>
<a class="class" href="BarrierTaskContext.html" title=":: Experimental :: A TaskContext with extra contextual info and tooling for tasks in a barrier stage."></a>
<a href="BarrierTaskContext.html" title=":: Experimental :: A TaskContext with extra contextual info and tooling for tasks in a barrier stage.">BarrierTaskContext</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="BarrierTaskInfo.html" title=":: Experimental :: Carries all task infos of a barrier task."></a>
<a href="BarrierTaskInfo.html" title=":: Experimental :: Carries all task infos of a barrier task.">BarrierTaskInfo</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="ComplexFutureAction.html" title="A FutureAction for actions that could trigger multiple Spark jobs."></a>
<a href="ComplexFutureAction.html" title="A FutureAction for actions that could trigger multiple Spark jobs.">ComplexFutureAction</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="ContextAwareIterator.html" title=":: DeveloperApi :: A TaskContext aware iterator."></a>
<a href="ContextAwareIterator.html" title=":: DeveloperApi :: A TaskContext aware iterator.">ContextAwareIterator</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="Dependency.html" title=":: DeveloperApi :: Base class for dependencies."></a>
<a href="Dependency.html" title=":: DeveloperApi :: Base class for dependencies.">Dependency</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="ExceptionFailure.html" title=":: DeveloperApi :: Task failed due to a runtime exception."></a>
<a href="ExceptionFailure.html" title=":: DeveloperApi :: Task failed due to a runtime exception.">ExceptionFailure</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="ExecutorLostFailure.html" title=":: DeveloperApi :: The task failed because the executor that it was running on was lost."></a>
<a href="ExecutorLostFailure.html" title=":: DeveloperApi :: The task failed because the executor that it was running on was lost.">ExecutorLostFailure</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="FetchFailed.html" title=":: DeveloperApi :: Task failed to fetch shuffle data from a remote node."></a>
<a href="FetchFailed.html" title=":: DeveloperApi :: Task failed to fetch shuffle data from a remote node.">FetchFailed</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="trait" href="FutureAction.html" title="A future for the result of an action to support cancellation."></a>
<a href="FutureAction.html" title="A future for the result of an action to support cancellation.">FutureAction</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="HashPartitioner.html" title="A org.apache.spark.Partitioner that implements hash-based partitioning using Java's Object.hashCode."></a>
<a href="HashPartitioner.html" title="A org.apache.spark.Partitioner that implements hash-based partitioning using Java's Object.hashCode.">HashPartitioner</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="InterruptibleIterator.html" title=":: DeveloperApi :: An iterator that wraps around an existing iterator to provide task killing functionality."></a>
<a href="InterruptibleIterator.html" title=":: DeveloperApi :: An iterator that wraps around an existing iterator to provide task killing functionality.">InterruptibleIterator</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="JobExecutionStatus.html" title=""></a>
<a href="JobExecutionStatus.html" title="">JobExecutionStatus</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="trait" href="JobSubmitter.html" title="Handle via which a &quot;run&quot; function passed to a ComplexFutureAction can submit jobs for execution."></a>
<a href="JobSubmitter.html" title="Handle via which a &quot;run&quot; function passed to a ComplexFutureAction can submit jobs for execution.">JobSubmitter</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="NarrowDependency.html" title=":: DeveloperApi :: Base class for dependencies where each partition of the child RDD depends on a small number of partitions of the parent RDD."></a>
<a href="NarrowDependency.html" title=":: DeveloperApi :: Base class for dependencies where each partition of the child RDD depends on a small number of partitions of the parent RDD.">NarrowDependency</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="OneToOneDependency.html" title=":: DeveloperApi :: Represents a one-to-one dependency between partitions of the parent and child RDDs."></a>
<a href="OneToOneDependency.html" title=":: DeveloperApi :: Represents a one-to-one dependency between partitions of the parent and child RDDs.">OneToOneDependency</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="trait" href="Partition.html" title="An identifier for a partition in an RDD."></a>
<a href="Partition.html" title="An identifier for a partition in an RDD.">Partition</a>
</li><li class="current-entities indented3">
<a class="object" href="Partitioner$.html" title=""></a>
<a class="class" href="Partitioner.html" title="An object that defines how the elements in a key-value pair RDD are partitioned by key."></a>
<a href="Partitioner.html" title="An object that defines how the elements in a key-value pair RDD are partitioned by key.">Partitioner</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="RangeDependency.html" title=":: DeveloperApi :: Represents a one-to-one dependency between ranges of partitions in the parent and child RDDs."></a>
<a href="RangeDependency.html" title=":: DeveloperApi :: Represents a one-to-one dependency between ranges of partitions in the parent and child RDDs.">RangeDependency</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="RangePartitioner.html" title="A org.apache.spark.Partitioner that partitions sortable records by range into roughly equal ranges."></a>
<a href="RangePartitioner.html" title="A org.apache.spark.Partitioner that partitions sortable records by range into roughly equal ranges.">RangePartitioner</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="object" href="Resubmitted$.html" title=":: DeveloperApi :: A org.apache.spark.scheduler.ShuffleMapTask that completed successfully earlier, but we lost the executor before the stage completed."></a>
<a href="Resubmitted$.html" title=":: DeveloperApi :: A org.apache.spark.scheduler.ShuffleMapTask that completed successfully earlier, but we lost the executor before the stage completed.">Resubmitted</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="SerializableWritable.html" title=""></a>
<a href="SerializableWritable.html" title="">SerializableWritable</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="ShuffleDependency.html" title=":: DeveloperApi :: Represents a dependency on the output of a shuffle stage."></a>
<a href="ShuffleDependency.html" title=":: DeveloperApi :: Represents a dependency on the output of a shuffle stage.">ShuffleDependency</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="SimpleFutureAction.html" title="A FutureAction holding the result of an action that triggers a single job."></a>
<a href="SimpleFutureAction.html" title="A FutureAction holding the result of an action that triggers a single job.">SimpleFutureAction</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="SparkConf.html" title="Configuration for a Spark application."></a>
<a href="SparkConf.html" title="Configuration for a Spark application.">SparkConf</a>
</li><li class="current-entities indented3">
<a class="object" href="SparkContext$.html" title="The SparkContext object contains a number of implicit conversions and parameters for use with various Spark features."></a>
<a class="class" href="" title="Main entry point for Spark functionality."></a>
<a href="" title="Main entry point for Spark functionality.">SparkContext</a>
</li><li class="current-entities indented3">
<a class="object" href="SparkEnv$.html" title=""></a>
<a class="class" href="SparkEnv.html" title=":: DeveloperApi :: Holds all the runtime environment objects for a running Spark instance (either master or worker), including the serializer, RpcEnv, block manager, map output tracker, etc."></a>
<a href="SparkEnv.html" title=":: DeveloperApi :: Holds all the runtime environment objects for a running Spark instance (either master or worker), including the serializer, RpcEnv, block manager, map output tracker, etc.">SparkEnv</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="SparkException.html" title=""></a>
<a href="SparkException.html" title="">SparkException</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="trait" href="SparkExecutorInfo.html" title=""></a>
<a href="SparkExecutorInfo.html" title="">SparkExecutorInfo</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="object" href="SparkFiles$.html" title="Resolves paths to files added through SparkContext.addFile()."></a>
<a href="SparkFiles$.html" title="Resolves paths to files added through SparkContext.addFile().">SparkFiles</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="SparkFirehoseListener.html" title=""></a>
<a href="SparkFirehoseListener.html" title="">SparkFirehoseListener</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="trait" href="SparkJobInfo.html" title=""></a>
<a href="SparkJobInfo.html" title="">SparkJobInfo</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="trait" href="SparkStageInfo.html" title=""></a>
<a href="SparkStageInfo.html" title="">SparkStageInfo</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="SparkStatusTracker.html" title="Low-level status reporting APIs for monitoring job and stage progress."></a>
<a href="SparkStatusTracker.html" title="Low-level status reporting APIs for monitoring job and stage progress.">SparkStatusTracker</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="trait" href="SparkThrowable.html" title=""></a>
<a href="SparkThrowable.html" title="">SparkThrowable</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="object" href="Success$.html" title=":: DeveloperApi :: Task succeeded."></a>
<a href="Success$.html" title=":: DeveloperApi :: Task succeeded.">Success</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="TaskCommitDenied.html" title=":: DeveloperApi :: Task requested the driver to commit, but was denied."></a>
<a href="TaskCommitDenied.html" title=":: DeveloperApi :: Task requested the driver to commit, but was denied.">TaskCommitDenied</a>
</li><li class="current-entities indented3">
<a class="object" href="TaskContext$.html" title=""></a>
<a class="class" href="TaskContext.html" title="Contextual information about a task which can be read or mutated during execution."></a>
<a href="TaskContext.html" title="Contextual information about a task which can be read or mutated during execution.">TaskContext</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="trait" href="TaskEndReason.html" title=":: DeveloperApi :: Various possible reasons why a task ended."></a>
<a href="TaskEndReason.html" title=":: DeveloperApi :: Various possible reasons why a task ended.">TaskEndReason</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="trait" href="TaskFailedReason.html" title=":: DeveloperApi :: Various possible reasons why a task failed."></a>
<a href="TaskFailedReason.html" title=":: DeveloperApi :: Various possible reasons why a task failed.">TaskFailedReason</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="TaskKilled.html" title=":: DeveloperApi :: Task was killed intentionally and needs to be rescheduled."></a>
<a href="TaskKilled.html" title=":: DeveloperApi :: Task was killed intentionally and needs to be rescheduled.">TaskKilled</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="class" href="TaskKilledException.html" title=":: DeveloperApi :: Exception thrown when a task is explicitly killed (i.e., task failure is expected)."></a>
<a href="TaskKilledException.html" title=":: DeveloperApi :: Exception thrown when a task is explicitly killed (i.e., task failure is expected).">TaskKilledException</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="object" href="TaskResultLost$.html" title=":: DeveloperApi :: The task finished successfully, but the result was lost from the executor's block manager before it was fetched."></a>
<a href="TaskResultLost$.html" title=":: DeveloperApi :: The task finished successfully, but the result was lost from the executor's block manager before it was fetched.">TaskResultLost</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="object" href="UnknownReason$.html" title=":: DeveloperApi :: We don't know why the task ended -- for example, because of a ClassNotFound exception when deserializing the task result."></a>
<a href="UnknownReason$.html" title=":: DeveloperApi :: We don't know why the task ended -- for example, because of a ClassNotFound exception when deserializing the task result.">UnknownReason</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="object" href="WritableConverter$.html" title=""></a>
<a href="WritableConverter$.html" title="">WritableConverter</a>
</li><li class="current-entities indented3">
<span class="separator"></span>
<a class="object" href="WritableFactory$.html" title=""></a>
<a href="WritableFactory$.html" title="">WritableFactory</a>
</li>
</ul>
</div>
</div>
<div id="content">
<body class="class type">
<div id="definition">
<a href="SparkContext$.html" title="See companion object"><div class="big-circle class-companion-object">c</div></a>
<p id="owner"><a href="../../index.html" class="extype" name="org">org</a>.<a href="../index.html" class="extype" name="org.apache">apache</a>.<a href="index.html" class="extype" name="org.apache.spark">spark</a></p>
<h1><a href="SparkContext$.html" title="See companion object">SparkContext</a><span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html" title="Permalink">
<i class="material-icons"></i>
</a>
</span></h1>
<h3><span class="morelinks"><div>
Companion <a href="SparkContext$.html" title="See companion object">object SparkContext</a>
</div></span></h3>
</div>
<h4 id="signature" class="signature">
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">class</span>
</span>
<span class="symbol">
<span class="name">SparkContext</span><span class="result"> extends <span class="extype" name="org.apache.spark.internal.Logging">Logging</span></span>
</span>
</h4>
<div id="comment" class="fullcommenttop"><div class="comment cmt"><p>Main entry point for Spark functionality. A SparkContext represents the connection to a Spark
cluster, and can be used to create RDDs, accumulators and broadcast variables on that cluster.
</p></div><dl class="attributes block"> <dt>Source</dt><dd><a href="https://github.com/apache/spark/tree/v3.3.4/core/src/main/scala/org/apache/spark/SparkContext.scala" target="_blank">SparkContext.scala</a></dd><dt>Note</dt><dd><span class="cmt"><p>Only one <code>SparkContext</code> should be active per JVM. You must <code>stop()</code> the
active <code>SparkContext</code> before creating a new one.</p></span></dd></dl><div class="toggleContainer block">
<span class="toggle">
Linear Supertypes
</span>
<div class="superTypes hiddenContent"><span class="extype" name="org.apache.spark.internal.Logging">Logging</span>, <span class="extype" name="scala.AnyRef">AnyRef</span>, <span class="extype" name="scala.Any">Any</span></div>
</div></div>
<div id="mbrsel">
<div class="toggle"></div>
<div id="memberfilter">
<i class="material-icons arrow"></i>
<span class="input">
<input id="mbrsel-input" placeholder="Filter all members" type="text" accesskey="/" />
</span>
<i class="clear material-icons"></i>
</div>
<div id="filterby">
<div id="order">
<span class="filtertype">Ordering</span>
<ol>
<li class="alpha in"><span>Alphabetic</span></li>
<li class="inherit out"><span>By Inheritance</span></li>
</ol>
</div>
<div class="ancestors">
<span class="filtertype">Inherited<br />
</span>
<ol id="linearization">
<li class="in" name="org.apache.spark.SparkContext"><span>SparkContext</span></li><li class="in" name="org.apache.spark.internal.Logging"><span>Logging</span></li><li class="in" name="scala.AnyRef"><span>AnyRef</span></li><li class="in" name="scala.Any"><span>Any</span></li>
</ol>
</div><div class="ancestors">
<span class="filtertype"></span>
<ol>
<li class="hideall out"><span>Hide All</span></li>
<li class="showall in"><span>Show All</span></li>
</ol>
</div>
<div id="visbl">
<span class="filtertype">Visibility</span>
<ol><li class="public in"><span>Public</span></li><li class="all out"><span>All</span></li></ol>
</div>
</div>
</div>
<div id="template">
<div id="allMembers">
<div id="constructors" class="members">
<h3>Instance Constructors</h3>
<ol><li name="org.apache.spark.SparkContext#&lt;init&gt;" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="&lt;init&gt;(master:String,appName:String,sparkHome:String,jars:Seq[String],environment:scala.collection.Map[String,String]):org.apache.spark.SparkContext"></a><a id="&lt;init&gt;:SparkContext"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#&lt;init&gt;(master:String,appName:String,sparkHome:String,jars:Seq[String],environment:scala.collection.Map[String,String]):org.apache.spark.SparkContext" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">new</span>
</span>
<span class="symbol">
<span class="name">SparkContext</span><span class="params">(<span name="master">master: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="appName">appName: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="sparkHome">sparkHome: <span class="extype" name="scala.Predef.String">String</span> = <span class="symbol">null</span></span>, <span name="jars">jars: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>] = <span class="symbol">Nil</span></span>, <span name="environment">environment: <span class="extype" name="scala.collection.Map">Map</span>[<span class="extype" name="scala.Predef.String">String</span>, <span class="extype" name="scala.Predef.String">String</span>] = <span class="symbol">Map()</span></span>)</span>
</span>
<p class="shortcomment cmt">Alternative constructor that allows setting common Spark properties directly
</p><div class="fullcomment"><div class="comment cmt"><p>Alternative constructor that allows setting common Spark properties directly
</p></div><dl class="paramcmts block"><dt class="param">master</dt><dd class="cmt"><p>Cluster URL to connect to (e.g. mesos://host:port, spark://host:port, local[4]).</p></dd><dt class="param">appName</dt><dd class="cmt"><p>A name for your application, to display on the cluster web UI.</p></dd><dt class="param">sparkHome</dt><dd class="cmt"><p>Location where Spark is installed on cluster nodes.</p></dd><dt class="param">jars</dt><dd class="cmt"><p>Collection of JARs to send to the cluster. These can be paths on the local file
system or HDFS, HTTP, HTTPS, or FTP URLs.</p></dd><dt class="param">environment</dt><dd class="cmt"><p>Environment variables to set on worker nodes.</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#&lt;init&gt;" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="&lt;init&gt;(master:String,appName:String,conf:org.apache.spark.SparkConf):org.apache.spark.SparkContext"></a><a id="&lt;init&gt;:SparkContext"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#&lt;init&gt;(master:String,appName:String,conf:org.apache.spark.SparkConf):org.apache.spark.SparkContext" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">new</span>
</span>
<span class="symbol">
<span class="name">SparkContext</span><span class="params">(<span name="master">master: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="appName">appName: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="conf">conf: <a href="SparkConf.html" class="extype" name="org.apache.spark.SparkConf">SparkConf</a></span>)</span>
</span>
<p class="shortcomment cmt">Alternative constructor that allows setting common Spark properties directly
</p><div class="fullcomment"><div class="comment cmt"><p>Alternative constructor that allows setting common Spark properties directly
</p></div><dl class="paramcmts block"><dt class="param">master</dt><dd class="cmt"><p>Cluster URL to connect to (e.g. mesos://host:port, spark://host:port, local[4]).</p></dd><dt class="param">appName</dt><dd class="cmt"><p>A name for your application, to display on the cluster web UI</p></dd><dt class="param">conf</dt><dd class="cmt"><p>a <a href="SparkConf.html" class="extype" name="org.apache.spark.SparkConf">org.apache.spark.SparkConf</a> object specifying other Spark parameters</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#&lt;init&gt;" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="&lt;init&gt;():org.apache.spark.SparkContext"></a><a id="&lt;init&gt;:SparkContext"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#&lt;init&gt;():org.apache.spark.SparkContext" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">new</span>
</span>
<span class="symbol">
<span class="name">SparkContext</span><span class="params">()</span>
</span>
<p class="shortcomment cmt">Create a SparkContext that loads settings from system properties (for instance, when
launching with ./bin/spark-submit).</p>
</li><li name="org.apache.spark.SparkContext#&lt;init&gt;" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="&lt;init&gt;(config:org.apache.spark.SparkConf):org.apache.spark.SparkContext"></a><a id="&lt;init&gt;:SparkContext"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#&lt;init&gt;(config:org.apache.spark.SparkConf):org.apache.spark.SparkContext" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">new</span>
</span>
<span class="symbol">
<span class="name">SparkContext</span><span class="params">(<span name="config">config: <a href="SparkConf.html" class="extype" name="org.apache.spark.SparkConf">SparkConf</a></span>)</span>
</span>
<p class="shortcomment cmt"></p><div class="fullcomment"><div class="comment cmt"></div><dl class="paramcmts block"><dt class="param">config</dt><dd class="cmt"><p>a Spark Config object describing the application configuration. Any settings in
this config overrides the default configs as well as system properties.</p></dd></dl></div>
</li></ol>
</div>
<div class="values members">
<h3>Value Members</h3>
<ol>
<li name="scala.AnyRef#!=" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="!=(x$1:Any):Boolean"></a><a id="!=(Any):Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#!=(x$1:Any):Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span title="gt4s: $bang$eq" class="name">!=</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.Any">Any</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd></dl></div>
</li><li name="scala.AnyRef###" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="##():Int"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html###():Int" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span title="gt4s: $hash$hash" class="name">##</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Int">Int</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd></dl></div>
</li><li name="scala.AnyRef#==" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="==(x$1:Any):Boolean"></a><a id="==(Any):Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#==(x$1:Any):Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span title="gt4s: $eq$eq" class="name">==</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.Any">Any</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#addArchive" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="addArchive(path:String):Unit"></a><a id="addArchive(String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#addArchive(path:String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">addArchive</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">:: Experimental ::
Add an archive to be downloaded and unpacked with this Spark job on every node.</p><div class="fullcomment"><div class="comment cmt"><p>:: Experimental ::
Add an archive to be downloaded and unpacked with this Spark job on every node.</p><p>If an archive is added during execution, it will not be available until the next TaskSet
starts.
</p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>can be either a local file, a file in HDFS (or other Hadoop-supported
filesystems), or an HTTP, HTTPS or FTP URI. To access the file in Spark jobs,
use <code>SparkFiles.get(paths-to-files)</code> to find its download/unpacked location.
The given path should be one of .zip, .tar, .tar.gz, .tgz and .jar.</p></dd></dl><dl class="attributes block"> <dt>Annotations</dt><dd>
<span class="name">@Experimental</span><span class="args">()</span>
</dd><dt>Since</dt><dd><p>3.1.0</p></dd><dt>Note</dt><dd><span class="cmt"><p>A path can be added only once. Subsequent additions of the same path are ignored.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#addFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="addFile(path:String,recursive:Boolean):Unit"></a><a id="addFile(String,Boolean):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#addFile(path:String,recursive:Boolean):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">addFile</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="recursive">recursive: <span class="extype" name="scala.Boolean">Boolean</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Add a file to be downloaded with this Spark job on every node.</p><div class="fullcomment"><div class="comment cmt"><p>Add a file to be downloaded with this Spark job on every node.</p><p>If a file is added during execution, it will not be available until the next TaskSet starts.
</p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>can be either a local file, a file in HDFS (or other Hadoop-supported
filesystems), or an HTTP, HTTPS or FTP URI. To access the file in Spark jobs,
use <code>SparkFiles.get(fileName)</code> to find its download location.</p></dd><dt class="param">recursive</dt><dd class="cmt"><p>if true, a directory can be given in <code>path</code>. Currently directories are
only supported for Hadoop-supported filesystems.</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>A path can be added only once. Subsequent additions of the same path are ignored.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#addFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="addFile(path:String):Unit"></a><a id="addFile(String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#addFile(path:String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">addFile</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Add a file to be downloaded with this Spark job on every node.</p><div class="fullcomment"><div class="comment cmt"><p>Add a file to be downloaded with this Spark job on every node.</p><p>If a file is added during execution, it will not be available until the next TaskSet starts.
</p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>can be either a local file, a file in HDFS (or other Hadoop-supported
filesystems), or an HTTP, HTTPS or FTP URI. To access the file in Spark jobs,
use <code>SparkFiles.get(fileName)</code> to find its download location.</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>A path can be added only once. Subsequent additions of the same path are ignored.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#addJar" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="addJar(path:String):Unit"></a><a id="addJar(String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#addJar(path:String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">addJar</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Adds a JAR dependency for all tasks to be executed on this <code>SparkContext</code> in the future.</p><div class="fullcomment"><div class="comment cmt"><p>Adds a JAR dependency for all tasks to be executed on this <code>SparkContext</code> in the future.</p><p>If a jar is added during execution, it will not be available until the next TaskSet starts.
</p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>can be either a local file, a file in HDFS (or other Hadoop-supported filesystems),
an HTTP, HTTPS or FTP URI, or local:/path for a file on every worker node.</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>A path can be added only once. Subsequent additions of the same path are ignored.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#addSparkListener" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="addSparkListener(listener:org.apache.spark.scheduler.SparkListenerInterface):Unit"></a><a id="addSparkListener(SparkListenerInterface):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#addSparkListener(listener:org.apache.spark.scheduler.SparkListenerInterface):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">addSparkListener</span><span class="params">(<span name="listener">listener: <span class="extype" name="org.apache.spark.scheduler.SparkListenerInterface">SparkListenerInterface</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">:: DeveloperApi ::
Register a listener to receive up-calls from events that happen during execution.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi ::
Register a listener to receive up-calls from events that happen during execution.
</p></div><dl class="attributes block"> <dt>Annotations</dt><dd>
<span class="name">@DeveloperApi</span><span class="args">()</span>
</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#appName" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="appName:String"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#appName:String" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">appName</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span>
</span>
</li><li name="org.apache.spark.SparkContext#applicationAttemptId" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="applicationAttemptId:Option[String]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#applicationAttemptId:Option[String]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">applicationAttemptId</span><span class="result">: <span class="extype" name="scala.Option">Option</span>[<span class="extype" name="scala.Predef.String">String</span>]</span>
</span>
</li><li name="org.apache.spark.SparkContext#applicationId" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="applicationId:String"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#applicationId:String" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">applicationId</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span>
</span>
<p class="shortcomment cmt">A unique identifier for the Spark application.</p><div class="fullcomment"><div class="comment cmt"><p>A unique identifier for the Spark application.
Its format depends on the scheduler implementation.
(i.e.
in case of local spark app something like 'local-1433865536131'
in case of YARN something like 'application_1433865536131_34483'
in case of MESOS something like 'driver-20170926223339-0001'
)
</p></div></div>
</li><li name="org.apache.spark.SparkContext#archives" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="archives:Seq[String]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#archives:Seq[String]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">archives</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span>
</span>
</li><li name="scala.Any#asInstanceOf" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="asInstanceOf[T0]:T0"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#asInstanceOf[T0]:T0" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">asInstanceOf</span><span class="tparams">[<span name="T0">T0</span>]</span><span class="result">: <span class="extype" name="scala.Any.asInstanceOf.T0">T0</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>Any</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#binaryFiles" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="binaryFiles(path:String,minPartitions:Int):org.apache.spark.rdd.RDD[(String,org.apache.spark.input.PortableDataStream)]"></a><a id="binaryFiles(String,Int):RDD[(String,PortableDataStream)]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#binaryFiles(path:String,minPartitions:Int):org.apache.spark.rdd.RDD[(String,org.apache.spark.input.PortableDataStream)]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">binaryFiles</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="scala.Predef.String">String</span>, <a href="input/PortableDataStream.html" class="extype" name="org.apache.spark.input.PortableDataStream">PortableDataStream</a>)]</span>
</span>
<p class="shortcomment cmt">Get an RDD for a Hadoop-readable dataset as PortableDataStream for each file
(useful for binary data)</p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a Hadoop-readable dataset as PortableDataStream for each file
(useful for binary data)</p><p>For example, if you have the following files:</p><pre>hdfs:<span class="cmt">//a-hdfs-path/part-00000</span>
hdfs:<span class="cmt">//a-hdfs-path/part-00001</span>
...
hdfs:<span class="cmt">//a-hdfs-path/part-nnnnn</span></pre><p>Do
<code>val rdd = sparkContext.binaryFiles(&quot;hdfs://a-hdfs-path&quot;)</code>,</p><p>then <code>rdd</code> contains</p><pre>(a-hdfs-path/part-<span class="num">00000</span>, its content)
(a-hdfs-path/part-<span class="num">00001</span>, its content)
...
(a-hdfs-path/part-nnnnn, its content)</pre></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>Directory to the input data files, the path can be comma separated paths as the
list of inputs.</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>A suggestion value of the minimal splitting number for input data.</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing tuples of file path and corresponding file content</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Small files are preferred; very large files may cause bad performance.</p></span>, <span class="cmt"><p>On some filesystems, <code>.../path/&#42;</code> can be a more efficient way to read all files
in a directory rather than <code>.../path/</code> or <code>.../path</code></p></span>, <span class="cmt"><p>Partitioning is determined by data locality. This may result in too few partitions
by default.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#binaryRecords" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="binaryRecords(path:String,recordLength:Int,conf:org.apache.hadoop.conf.Configuration):org.apache.spark.rdd.RDD[Array[Byte]]"></a><a id="binaryRecords(String,Int,Configuration):RDD[Array[Byte]]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#binaryRecords(path:String,recordLength:Int,conf:org.apache.hadoop.conf.Configuration):org.apache.spark.rdd.RDD[Array[Byte]]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">binaryRecords</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="recordLength">recordLength: <span class="extype" name="scala.Int">Int</span></span>, <span name="conf">conf: <span class="extype" name="org.apache.hadoop.conf.Configuration">Configuration</span> = <span class="symbol"><span class="name"><a href="#hadoopConfiguration:org.apache.hadoop.conf.Configuration">hadoopConfiguration</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="scala.Array">Array</span>[<span class="extype" name="scala.Byte">Byte</span>]]</span>
</span>
<p class="shortcomment cmt">Load data from a flat binary file, assuming the length of each record is constant.</p><div class="fullcomment"><div class="comment cmt"><p>Load data from a flat binary file, assuming the length of each record is constant.
</p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>Directory to the input data files, the path can be comma separated paths as the
list of inputs.</p></dd><dt class="param">recordLength</dt><dd class="cmt"><p>The length at which to split the records</p></dd><dt class="param">conf</dt><dd class="cmt"><p>Configuration for setting up the dataset.</p></dd><dt>returns</dt><dd class="cmt"><p>An RDD of data with values, represented as byte arrays</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>We ensure that the byte array for each record in the resulting RDD
has the provided record length.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#broadcast" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="broadcast[T](value:T)(implicitevidence$9:scala.reflect.ClassTag[T]):org.apache.spark.broadcast.Broadcast[T]"></a><a id="broadcast[T](T)(ClassTag[T]):Broadcast[T]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#broadcast[T](value:T)(implicitevidence$9:scala.reflect.ClassTag[T]):org.apache.spark.broadcast.Broadcast[T]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">broadcast</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="value">value: <span class="extype" name="org.apache.spark.SparkContext.broadcast.T">T</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.broadcast.T">T</span>]</span>)</span><span class="result">: <a href="broadcast/Broadcast.html" class="extype" name="org.apache.spark.broadcast.Broadcast">Broadcast</a>[<span class="extype" name="org.apache.spark.SparkContext.broadcast.T">T</span>]</span>
</span>
<p class="shortcomment cmt">Broadcast a read-only variable to the cluster, returning a
<a href="broadcast/Broadcast.html" class="extype" name="org.apache.spark.broadcast.Broadcast">org.apache.spark.broadcast.Broadcast</a> object for reading it in distributed functions.</p><div class="fullcomment"><div class="comment cmt"><p>Broadcast a read-only variable to the cluster, returning a
<a href="broadcast/Broadcast.html" class="extype" name="org.apache.spark.broadcast.Broadcast">org.apache.spark.broadcast.Broadcast</a> object for reading it in distributed functions.
The variable will be sent to each cluster only once.
</p></div><dl class="paramcmts block"><dt class="param">value</dt><dd class="cmt"><p>value to broadcast to the Spark nodes</p></dd><dt>returns</dt><dd class="cmt"><p><code>Broadcast</code> object, a read-only variable cached on each machine</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#cancelAllJobs" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="cancelAllJobs():Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#cancelAllJobs():Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">cancelAllJobs</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Cancel all jobs that have been scheduled or are running.</p>
</li><li name="org.apache.spark.SparkContext#cancelJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="cancelJob(jobId:Int):Unit"></a><a id="cancelJob(Int):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#cancelJob(jobId:Int):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">cancelJob</span><span class="params">(<span name="jobId">jobId: <span class="extype" name="scala.Int">Int</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Cancel a given job if it's scheduled or running.</p><div class="fullcomment"><div class="comment cmt"><p>Cancel a given job if it's scheduled or running.
</p></div><dl class="paramcmts block"><dt class="param">jobId</dt><dd class="cmt"><p>the job ID to cancel</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Throws <code>InterruptedException</code> if the cancel message cannot be sent</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#cancelJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="cancelJob(jobId:Int,reason:String):Unit"></a><a id="cancelJob(Int,String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#cancelJob(jobId:Int,reason:String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">cancelJob</span><span class="params">(<span name="jobId">jobId: <span class="extype" name="scala.Int">Int</span></span>, <span name="reason">reason: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Cancel a given job if it's scheduled or running.</p><div class="fullcomment"><div class="comment cmt"><p>Cancel a given job if it's scheduled or running.
</p></div><dl class="paramcmts block"><dt class="param">jobId</dt><dd class="cmt"><p>the job ID to cancel</p></dd><dt class="param">reason</dt><dd class="cmt"><p>optional reason for cancellation</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Throws <code>InterruptedException</code> if the cancel message cannot be sent</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#cancelJobGroup" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="cancelJobGroup(groupId:String):Unit"></a><a id="cancelJobGroup(String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#cancelJobGroup(groupId:String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">cancelJobGroup</span><span class="params">(<span name="groupId">groupId: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Cancel active jobs for the specified group.</p><div class="fullcomment"><div class="comment cmt"><p>Cancel active jobs for the specified group. See <code>org.apache.spark.SparkContext.setJobGroup</code>
for more information.
</p></div></div>
</li><li name="org.apache.spark.SparkContext#cancelStage" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="cancelStage(stageId:Int):Unit"></a><a id="cancelStage(Int):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#cancelStage(stageId:Int):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">cancelStage</span><span class="params">(<span name="stageId">stageId: <span class="extype" name="scala.Int">Int</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Cancel a given stage and all jobs associated with it.</p><div class="fullcomment"><div class="comment cmt"><p>Cancel a given stage and all jobs associated with it.
</p></div><dl class="paramcmts block"><dt class="param">stageId</dt><dd class="cmt"><p>the stage ID to cancel</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Throws <code>InterruptedException</code> if the cancel message cannot be sent</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#cancelStage" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="cancelStage(stageId:Int,reason:String):Unit"></a><a id="cancelStage(Int,String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#cancelStage(stageId:Int,reason:String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">cancelStage</span><span class="params">(<span name="stageId">stageId: <span class="extype" name="scala.Int">Int</span></span>, <span name="reason">reason: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Cancel a given stage and all jobs associated with it.</p><div class="fullcomment"><div class="comment cmt"><p>Cancel a given stage and all jobs associated with it.
</p></div><dl class="paramcmts block"><dt class="param">stageId</dt><dd class="cmt"><p>the stage ID to cancel</p></dd><dt class="param">reason</dt><dd class="cmt"><p>reason for cancellation</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Throws <code>InterruptedException</code> if the cancel message cannot be sent</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#checkpointFile" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="checkpointFile[T](path:String)(implicitevidence$5:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="checkpointFile[T](String)(ClassTag[T]):RDD[T]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#checkpointFile[T](path:String)(implicitevidence$5:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">checkpointFile</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.checkpointFile.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.checkpointFile.T">T</span>]</span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected[<a href="index.html" class="extype" name="org.apache.spark">spark</a>] </dd></dl></div>
</li><li name="org.apache.spark.SparkContext#clearCallSite" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="clearCallSite():Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#clearCallSite():Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">clearCallSite</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Clear the thread-local property for overriding the call sites
of actions and RDDs.</p>
</li><li name="org.apache.spark.SparkContext#clearJobGroup" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="clearJobGroup():Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#clearJobGroup():Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">clearJobGroup</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Clear the current thread's job group ID and its description.</p>
</li><li name="scala.AnyRef#clone" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="clone():Object"></a><a id="clone():AnyRef"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#clone():Object" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">clone</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.AnyRef">AnyRef</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected[<span class="extype" name="java.lang">lang</span>] </dd><dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd>
<span class="name">@throws</span><span class="args">(<span>
<span class="defval" name="classOf[java.lang.CloneNotSupportedException]">...</span>
</span>)</span>
<span class="name">@native</span><span class="args">()</span>
</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#collectionAccumulator" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="collectionAccumulator[T](name:String):org.apache.spark.util.CollectionAccumulator[T]"></a><a id="collectionAccumulator[T](String):CollectionAccumulator[T]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#collectionAccumulator[T](name:String):org.apache.spark.util.CollectionAccumulator[T]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">collectionAccumulator</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="name">name: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <a href="util/CollectionAccumulator.html" class="extype" name="org.apache.spark.util.CollectionAccumulator">CollectionAccumulator</a>[<span class="extype" name="org.apache.spark.SparkContext.collectionAccumulator.T">T</span>]</span>
</span>
<p class="shortcomment cmt">Create and register a <code>CollectionAccumulator</code>, which starts with empty list and accumulates
inputs by adding them into the list.</p>
</li><li name="org.apache.spark.SparkContext#collectionAccumulator" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="collectionAccumulator[T]:org.apache.spark.util.CollectionAccumulator[T]"></a><a id="collectionAccumulator[T]:CollectionAccumulator[T]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#collectionAccumulator[T]:org.apache.spark.util.CollectionAccumulator[T]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">collectionAccumulator</span><span class="tparams">[<span name="T">T</span>]</span><span class="result">: <a href="util/CollectionAccumulator.html" class="extype" name="org.apache.spark.util.CollectionAccumulator">CollectionAccumulator</a>[<span class="extype" name="org.apache.spark.SparkContext.collectionAccumulator.T">T</span>]</span>
</span>
<p class="shortcomment cmt">Create and register a <code>CollectionAccumulator</code>, which starts with empty list and accumulates
inputs by adding them into the list.</p>
</li><li name="org.apache.spark.SparkContext#defaultMinPartitions" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="defaultMinPartitions:Int"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#defaultMinPartitions:Int" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">defaultMinPartitions</span><span class="result">: <span class="extype" name="scala.Int">Int</span></span>
</span>
<p class="shortcomment cmt">Default min number of partitions for Hadoop RDDs when not given by user
Notice that we use math.min so the &quot;defaultMinPartitions&quot; cannot be higher than 2.</p><div class="fullcomment"><div class="comment cmt"><p>Default min number of partitions for Hadoop RDDs when not given by user
Notice that we use math.min so the &quot;defaultMinPartitions&quot; cannot be higher than 2.
The reasons for this are discussed in https://github.com/mesos/spark/pull/718
</p></div></div>
</li><li name="org.apache.spark.SparkContext#defaultParallelism" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="defaultParallelism:Int"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#defaultParallelism:Int" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">defaultParallelism</span><span class="result">: <span class="extype" name="scala.Int">Int</span></span>
</span>
<p class="shortcomment cmt">Default level of parallelism to use when not given by user (e.g.</p><div class="fullcomment"><div class="comment cmt"><p>Default level of parallelism to use when not given by user (e.g. parallelize and makeRDD).</p></div></div>
</li><li name="org.apache.spark.SparkContext#deployMode" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="deployMode:String"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#deployMode:String" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">deployMode</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span>
</span>
</li><li name="org.apache.spark.SparkContext#doubleAccumulator" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="doubleAccumulator(name:String):org.apache.spark.util.DoubleAccumulator"></a><a id="doubleAccumulator(String):DoubleAccumulator"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#doubleAccumulator(name:String):org.apache.spark.util.DoubleAccumulator" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">doubleAccumulator</span><span class="params">(<span name="name">name: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <a href="util/DoubleAccumulator.html" class="extype" name="org.apache.spark.util.DoubleAccumulator">DoubleAccumulator</a></span>
</span>
<p class="shortcomment cmt">Create and register a double accumulator, which starts with 0 and accumulates inputs by <code>add</code>.</p>
</li><li name="org.apache.spark.SparkContext#doubleAccumulator" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="doubleAccumulator:org.apache.spark.util.DoubleAccumulator"></a><a id="doubleAccumulator:DoubleAccumulator"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#doubleAccumulator:org.apache.spark.util.DoubleAccumulator" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">doubleAccumulator</span><span class="result">: <a href="util/DoubleAccumulator.html" class="extype" name="org.apache.spark.util.DoubleAccumulator">DoubleAccumulator</a></span>
</span>
<p class="shortcomment cmt">Create and register a double accumulator, which starts with 0 and accumulates inputs by <code>add</code>.</p>
</li><li name="org.apache.spark.SparkContext#emptyRDD" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="emptyRDD[T](implicitevidence$8:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="emptyRDD[T](ClassTag[T]):RDD[T]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#emptyRDD[T](implicitevidence$8:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">emptyRDD</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.emptyRDD.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.emptyRDD.T">T</span>]</span>
</span>
<p class="shortcomment cmt">Get an RDD that has no partitions or elements.</p>
</li><li name="scala.AnyRef#eq" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="eq(x$1:AnyRef):Boolean"></a><a id="eq(AnyRef):Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#eq(x$1:AnyRef):Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">eq</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.AnyRef">AnyRef</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd></dl></div>
</li><li name="scala.AnyRef#equals" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="equals(x$1:Any):Boolean"></a><a id="equals(Any):Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#equals(x$1:Any):Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">equals</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.Any">Any</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#files" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="files:Seq[String]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#files:Seq[String]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">files</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span>
</span>
</li><li name="scala.AnyRef#finalize" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="finalize():Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#finalize():Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">finalize</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected[<span class="extype" name="java.lang">lang</span>] </dd><dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd>
<span class="name">@throws</span><span class="args">(<span>
<span class="symbol">classOf[java.lang.Throwable]</span>
</span>)</span>
</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#getAllPools" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="getAllPools:Seq[org.apache.spark.scheduler.Schedulable]"></a><a id="getAllPools:Seq[Schedulable]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#getAllPools:Seq[org.apache.spark.scheduler.Schedulable]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">getAllPools</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="org.apache.spark.scheduler.Schedulable">Schedulable</span>]</span>
</span>
<p class="shortcomment cmt">:: DeveloperApi ::
Return pools for fair scheduler
</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi ::
Return pools for fair scheduler
</p></div><dl class="attributes block"> <dt>Annotations</dt><dd>
<span class="name">@DeveloperApi</span><span class="args">()</span>
</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#getCheckpointDir" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="getCheckpointDir:Option[String]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#getCheckpointDir:Option[String]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">getCheckpointDir</span><span class="result">: <span class="extype" name="scala.Option">Option</span>[<span class="extype" name="scala.Predef.String">String</span>]</span>
</span>
</li><li name="scala.AnyRef#getClass" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="getClass():Class[_]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#getClass():Class[_]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">getClass</span><span class="params">()</span><span class="result">: <span class="extype" name="java.lang.Class">Class</span>[_]</span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd><dt>Annotations</dt><dd>
<span class="name">@native</span><span class="args">()</span>
</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#getConf" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="getConf:org.apache.spark.SparkConf"></a><a id="getConf:SparkConf"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#getConf:org.apache.spark.SparkConf" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">getConf</span><span class="result">: <a href="SparkConf.html" class="extype" name="org.apache.spark.SparkConf">SparkConf</a></span>
</span>
<p class="shortcomment cmt">Return a copy of this SparkContext's configuration.</p><div class="fullcomment"><div class="comment cmt"><p>Return a copy of this SparkContext's configuration. The configuration <i>cannot</i> be
changed at runtime.
</p></div></div>
</li><li name="org.apache.spark.SparkContext#getExecutorMemoryStatus" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="getExecutorMemoryStatus:scala.collection.Map[String,(Long,Long)]"></a><a id="getExecutorMemoryStatus:Map[String,(Long,Long)]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#getExecutorMemoryStatus:scala.collection.Map[String,(Long,Long)]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">getExecutorMemoryStatus</span><span class="result">: <span class="extype" name="scala.collection.Map">Map</span>[<span class="extype" name="scala.Predef.String">String</span>, (<span class="extype" name="scala.Long">Long</span>, <span class="extype" name="scala.Long">Long</span>)]</span>
</span>
<p class="shortcomment cmt">Return a map from the block manager to the max memory available for caching and the remaining
memory available for caching.</p>
</li><li name="org.apache.spark.SparkContext#getLocalProperty" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="getLocalProperty(key:String):String"></a><a id="getLocalProperty(String):String"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#getLocalProperty(key:String):String" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">getLocalProperty</span><span class="params">(<span name="key">key: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span>
</span>
<p class="shortcomment cmt">Get a local property set in this thread, or null if it is missing.</p><div class="fullcomment"><div class="comment cmt"><p>Get a local property set in this thread, or null if it is missing. See
<code>org.apache.spark.SparkContext.setLocalProperty</code>.
</p></div></div>
</li><li name="org.apache.spark.SparkContext#getPersistentRDDs" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="getPersistentRDDs:scala.collection.Map[Int,org.apache.spark.rdd.RDD[_]]"></a><a id="getPersistentRDDs:Map[Int,RDD[_]]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#getPersistentRDDs:scala.collection.Map[Int,org.apache.spark.rdd.RDD[_]]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">getPersistentRDDs</span><span class="result">: <span class="extype" name="scala.collection.Map">Map</span>[<span class="extype" name="scala.Int">Int</span>, <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[_]]</span>
</span>
<p class="shortcomment cmt">Returns an immutable map of RDDs that have marked themselves as persistent via cache() call.</p><div class="fullcomment"><div class="comment cmt"><p>Returns an immutable map of RDDs that have marked themselves as persistent via cache() call.
</p></div><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>This does not necessarily mean the caching or computation was successful.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#getPoolForName" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="getPoolForName(pool:String):Option[org.apache.spark.scheduler.Schedulable]"></a><a id="getPoolForName(String):Option[Schedulable]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#getPoolForName(pool:String):Option[org.apache.spark.scheduler.Schedulable]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">getPoolForName</span><span class="params">(<span name="pool">pool: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Option">Option</span>[<span class="extype" name="org.apache.spark.scheduler.Schedulable">Schedulable</span>]</span>
</span>
<p class="shortcomment cmt">:: DeveloperApi ::
Return the pool associated with the given name, if one exists
</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi ::
Return the pool associated with the given name, if one exists
</p></div><dl class="attributes block"> <dt>Annotations</dt><dd>
<span class="name">@DeveloperApi</span><span class="args">()</span>
</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#getRDDStorageInfo" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="getRDDStorageInfo:Array[org.apache.spark.storage.RDDInfo]"></a><a id="getRDDStorageInfo:Array[RDDInfo]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#getRDDStorageInfo:Array[org.apache.spark.storage.RDDInfo]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">getRDDStorageInfo</span><span class="result">: <span class="extype" name="scala.Array">Array</span>[<a href="storage/RDDInfo.html" class="extype" name="org.apache.spark.storage.RDDInfo">RDDInfo</a>]</span>
</span>
<p class="shortcomment cmt">:: DeveloperApi ::
Return information about what RDDs are cached, if they are in mem or on disk, how much space
they take, etc.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi ::
Return information about what RDDs are cached, if they are in mem or on disk, how much space
they take, etc.
</p></div><dl class="attributes block"> <dt>Annotations</dt><dd>
<span class="name">@DeveloperApi</span><span class="args">()</span>
</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#getSchedulingMode" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="getSchedulingMode:org.apache.spark.scheduler.SchedulingMode.SchedulingMode"></a><a id="getSchedulingMode:SchedulingMode"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#getSchedulingMode:org.apache.spark.scheduler.SchedulingMode.SchedulingMode" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">getSchedulingMode</span><span class="result">: <a href="scheduler/SchedulingMode$.html#SchedulingMode=org.apache.spark.scheduler.SchedulingMode.Value" class="extmbr" name="org.apache.spark.scheduler.SchedulingMode.SchedulingMode">SchedulingMode</a></span>
</span>
<p class="shortcomment cmt">Return current scheduling mode
</p>
</li><li name="org.apache.spark.SparkContext#hadoopConfiguration" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="hadoopConfiguration:org.apache.hadoop.conf.Configuration"></a><a id="hadoopConfiguration:Configuration"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#hadoopConfiguration:org.apache.hadoop.conf.Configuration" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">hadoopConfiguration</span><span class="result">: <span class="extype" name="org.apache.hadoop.conf.Configuration">Configuration</span></span>
</span>
<p class="shortcomment cmt">A default Hadoop Configuration for the Hadoop code (e.g.</p><div class="fullcomment"><div class="comment cmt"><p>A default Hadoop Configuration for the Hadoop code (e.g. file systems) that we reuse.
</p></div><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>As it will be reused in all Hadoop RDDs, it's better not to modify it unless you
plan to set some global configurations for all Hadoop RDDs.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#hadoopFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="hadoopFile[K,V,F&lt;:org.apache.hadoop.mapred.InputFormat[K,V]](path:String)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="hadoopFile[K,V,F&lt;:InputFormat[K,V]](String)(ClassTag[K],ClassTag[V],ClassTag[F]):RDD[(K,V)]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#hadoopFile[K,V,F&lt;:org.apache.hadoop.mapred.InputFormat[K,V]](path:String)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">hadoopFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>, <span name="F">F &lt;: <span class="extype" name="org.apache.hadoop.mapred.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>]</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="km">km: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>]</span>, <span name="vm">vm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>]</span>, <span name="fm">fm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.F">F</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>)]</span>
</span>
<p class="shortcomment cmt">Smarter version of hadoopFile() that uses class tags to figure out the classes of keys,
values and the InputFormat so that users don't need to pass them directly.</p><div class="fullcomment"><div class="comment cmt"><p>Smarter version of hadoopFile() that uses class tags to figure out the classes of keys,
values and the InputFormat so that users don't need to pass them directly. Instead, callers
can just write, for example,</p><pre><span class="kw">val</span> file = sparkContext.hadoopFile[LongWritable, Text, TextInputFormat](path)</pre></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths as
a list of inputs</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each
record, directly caching the returned RDD or directly passing it to an aggregation or shuffle
operation will create many references to the same object.
If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first
copy them using a <code>map</code> function.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#hadoopFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="hadoopFile[K,V,F&lt;:org.apache.hadoop.mapred.InputFormat[K,V]](path:String,minPartitions:Int)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="hadoopFile[K,V,F&lt;:InputFormat[K,V]](String,Int)(ClassTag[K],ClassTag[V],ClassTag[F]):RDD[(K,V)]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#hadoopFile[K,V,F&lt;:org.apache.hadoop.mapred.InputFormat[K,V]](path:String,minPartitions:Int)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">hadoopFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>, <span name="F">F &lt;: <span class="extype" name="org.apache.hadoop.mapred.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>]</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="km">km: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>]</span>, <span name="vm">vm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>]</span>, <span name="fm">fm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.F">F</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>)]</span>
</span>
<p class="shortcomment cmt">Smarter version of hadoopFile() that uses class tags to figure out the classes of keys,
values and the InputFormat so that users don't need to pass them directly.</p><div class="fullcomment"><div class="comment cmt"><p>Smarter version of hadoopFile() that uses class tags to figure out the classes of keys,
values and the InputFormat so that users don't need to pass them directly. Instead, callers
can just write, for example,</p><pre><span class="kw">val</span> file = sparkContext.hadoopFile[LongWritable, Text, TextInputFormat](path, minPartitions)</pre></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths
as a list of inputs</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>suggested minimum number of partitions for the resulting RDD</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each
record, directly caching the returned RDD or directly passing it to an aggregation or shuffle
operation will create many references to the same object.
If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first
copy them using a <code>map</code> function.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#hadoopFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="hadoopFile[K,V](path:String,inputFormatClass:Class[_&lt;:org.apache.hadoop.mapred.InputFormat[K,V]],keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="hadoopFile[K,V](String,Class[_&lt;:InputFormat[K,V]],Class[K],Class[V],Int):RDD[(K,V)]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#hadoopFile[K,V](path:String,inputFormatClass:Class[_&lt;:org.apache.hadoop.mapred.InputFormat[K,V]],keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">hadoopFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="inputFormatClass">inputFormatClass: <span class="extype" name="scala.Predef.Class">Class</span>[_ &lt;: <span class="extype" name="org.apache.hadoop.mapred.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>]]</span>, <span name="keyClass">keyClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>]</span>, <span name="valueClass">valueClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>]</span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>)]</span>
</span>
<p class="shortcomment cmt">Get an RDD for a Hadoop file with an arbitrary InputFormat
</p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a Hadoop file with an arbitrary InputFormat
</p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths
as a list of inputs</p></dd><dt class="param">inputFormatClass</dt><dd class="cmt"><p>storage format of the data to be read</p></dd><dt class="param">keyClass</dt><dd class="cmt"><p><code>Class</code> of the key associated with the <code>inputFormatClass</code> parameter</p></dd><dt class="param">valueClass</dt><dd class="cmt"><p><code>Class</code> of the value associated with the <code>inputFormatClass</code> parameter</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>suggested minimum number of partitions for the resulting RDD</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each
record, directly caching the returned RDD or directly passing it to an aggregation or shuffle
operation will create many references to the same object.
If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first
copy them using a <code>map</code> function.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#hadoopRDD" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="hadoopRDD[K,V](conf:org.apache.hadoop.mapred.JobConf,inputFormatClass:Class[_&lt;:org.apache.hadoop.mapred.InputFormat[K,V]],keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="hadoopRDD[K,V](JobConf,Class[_&lt;:InputFormat[K,V]],Class[K],Class[V],Int):RDD[(K,V)]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#hadoopRDD[K,V](conf:org.apache.hadoop.mapred.JobConf,inputFormatClass:Class[_&lt;:org.apache.hadoop.mapred.InputFormat[K,V]],keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">hadoopRDD</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>]</span><span class="params">(<span name="conf">conf: <span class="extype" name="org.apache.hadoop.mapred.JobConf">JobConf</span></span>, <span name="inputFormatClass">inputFormatClass: <span class="extype" name="scala.Predef.Class">Class</span>[_ &lt;: <span class="extype" name="org.apache.hadoop.mapred.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopRDD.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopRDD.V">V</span>]]</span>, <span name="keyClass">keyClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopRDD.K">K</span>]</span>, <span name="valueClass">valueClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopRDD.V">V</span>]</span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.hadoopRDD.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopRDD.V">V</span>)]</span>
</span>
<p class="shortcomment cmt">Get an RDD for a Hadoop-readable dataset from a Hadoop JobConf given its InputFormat and other
necessary info (e.g.</p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a Hadoop-readable dataset from a Hadoop JobConf given its InputFormat and other
necessary info (e.g. file name for a filesystem-based dataset, table name for HyperTable),
using the older MapReduce API (<code>org.apache.hadoop.mapred</code>).
</p></div><dl class="paramcmts block"><dt class="param">conf</dt><dd class="cmt"><p>JobConf for setting up the dataset. Note: This will be put into a Broadcast.
Therefore if you plan to reuse this conf to create multiple RDDs, you need to make
sure you won't modify the conf. A safe approach is always creating a new conf for
a new RDD.</p></dd><dt class="param">inputFormatClass</dt><dd class="cmt"><p>storage format of the data to be read</p></dd><dt class="param">keyClass</dt><dd class="cmt"><p><code>Class</code> of the key associated with the <code>inputFormatClass</code> parameter</p></dd><dt class="param">valueClass</dt><dd class="cmt"><p><code>Class</code> of the value associated with the <code>inputFormatClass</code> parameter</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>Minimum number of Hadoop Splits to generate.</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each
record, directly caching the returned RDD or directly passing it to an aggregation or shuffle
operation will create many references to the same object.
If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first
copy them using a <code>map</code> function.</p></span></dd></dl></div>
</li><li name="scala.AnyRef#hashCode" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="hashCode():Int"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#hashCode():Int" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">hashCode</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Int">Int</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd><dt>Annotations</dt><dd>
<span class="name">@native</span><span class="args">()</span>
</dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#initializeLogIfNecessary" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="initializeLogIfNecessary(isInterpreter:Boolean,silent:Boolean):Boolean"></a><a id="initializeLogIfNecessary(Boolean,Boolean):Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#initializeLogIfNecessary(isInterpreter:Boolean,silent:Boolean):Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">initializeLogIfNecessary</span><span class="params">(<span name="isInterpreter">isInterpreter: <span class="extype" name="scala.Boolean">Boolean</span></span>, <span name="silent">silent: <span class="extype" name="scala.Boolean">Boolean</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#initializeLogIfNecessary" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="initializeLogIfNecessary(isInterpreter:Boolean):Unit"></a><a id="initializeLogIfNecessary(Boolean):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#initializeLogIfNecessary(isInterpreter:Boolean):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">initializeLogIfNecessary</span><span class="params">(<span name="isInterpreter">isInterpreter: <span class="extype" name="scala.Boolean">Boolean</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="scala.Any#isInstanceOf" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="isInstanceOf[T0]:Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#isInstanceOf[T0]:Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">isInstanceOf</span><span class="tparams">[<span name="T0">T0</span>]</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>Any</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#isLocal" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="isLocal:Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#isLocal:Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">isLocal</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
</li><li name="org.apache.spark.SparkContext#isStopped" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="isStopped:Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#isStopped:Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">isStopped</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<p class="shortcomment cmt"></p><div class="fullcomment"><div class="comment cmt"></div><dl class="paramcmts block"><dt>returns</dt><dd class="cmt"><p>true if context is stopped or in the midst of stopping.</p></dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#isTraceEnabled" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="isTraceEnabled():Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#isTraceEnabled():Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">isTraceEnabled</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#jars" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="jars:Seq[String]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#jars:Seq[String]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">jars</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span>
</span>
</li><li name="org.apache.spark.SparkContext#killExecutor" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="killExecutor(executorId:String):Boolean"></a><a id="killExecutor(String):Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#killExecutor(executorId:String):Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">killExecutor</span><span class="params">(<span name="executorId">executorId: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<p class="shortcomment cmt">:: DeveloperApi ::
Request that the cluster manager kill the specified executor.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi ::
Request that the cluster manager kill the specified executor.
</p></div><dl class="paramcmts block"><dt>returns</dt><dd class="cmt"><p>whether the request is received.</p></dd></dl><dl class="attributes block"> <dt>Annotations</dt><dd>
<span class="name">@DeveloperApi</span><span class="args">()</span>
</dd><dt>Note</dt><dd><span class="cmt"><p>This is an indication to the cluster manager that the application wishes to adjust
its resource usage downwards. If the application wishes to replace the executor it kills
through this method with a new one, it should follow up explicitly with a call to
{{SparkContext#requestExecutors}}.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#killExecutors" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="killExecutors(executorIds:Seq[String]):Boolean"></a><a id="killExecutors(Seq[String]):Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#killExecutors(executorIds:Seq[String]):Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">killExecutors</span><span class="params">(<span name="executorIds">executorIds: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<p class="shortcomment cmt">:: DeveloperApi ::
Request that the cluster manager kill the specified executors.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi ::
Request that the cluster manager kill the specified executors.</p><p>This is not supported when dynamic allocation is turned on.
</p></div><dl class="paramcmts block"><dt>returns</dt><dd class="cmt"><p>whether the request is received.</p></dd></dl><dl class="attributes block"> <dt>Annotations</dt><dd>
<span class="name">@DeveloperApi</span><span class="args">()</span>
</dd><dt>Note</dt><dd><span class="cmt"><p>This is an indication to the cluster manager that the application wishes to adjust
its resource usage downwards. If the application wishes to replace the executors it kills
through this method with new ones, it should follow up explicitly with a call to
{{SparkContext#requestExecutors}}.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#killTaskAttempt" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="killTaskAttempt(taskId:Long,interruptThread:Boolean,reason:String):Boolean"></a><a id="killTaskAttempt(Long,Boolean,String):Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#killTaskAttempt(taskId:Long,interruptThread:Boolean,reason:String):Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">killTaskAttempt</span><span class="params">(<span name="taskId">taskId: <span class="extype" name="scala.Long">Long</span></span>, <span name="interruptThread">interruptThread: <span class="extype" name="scala.Boolean">Boolean</span> = <span class="symbol">true</span></span>, <span name="reason">reason: <span class="extype" name="scala.Predef.String">String</span> = <span class="defval" name="&quot;killed via SparkContext.killTaskAttempt&quot;">...</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<p class="shortcomment cmt">Kill and reschedule the given task attempt.</p><div class="fullcomment"><div class="comment cmt"><p>Kill and reschedule the given task attempt. Task ids can be obtained from the Spark UI
or through SparkListener.onTaskStart.
</p></div><dl class="paramcmts block"><dt class="param">taskId</dt><dd class="cmt"><p>the task ID to kill. This id uniquely identifies the task attempt.</p></dd><dt class="param">interruptThread</dt><dd class="cmt"><p>whether to interrupt the thread running the task.</p></dd><dt class="param">reason</dt><dd class="cmt"><p>the reason for killing the task, which should be a short string. If a task
is killed multiple times with different reasons, only one reason will be reported.</p></dd><dt>returns</dt><dd class="cmt"><p>Whether the task was successfully killed.</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#listArchives" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="listArchives():Seq[String]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#listArchives():Seq[String]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">listArchives</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span>
</span>
<p class="shortcomment cmt">:: Experimental ::
Returns a list of archive paths that are added to resources.</p><div class="fullcomment"><div class="comment cmt"><p>:: Experimental ::
Returns a list of archive paths that are added to resources.
</p></div><dl class="attributes block"> <dt>Annotations</dt><dd>
<span class="name">@Experimental</span><span class="args">()</span>
</dd><dt>Since</dt><dd><p>3.1.0</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#listFiles" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="listFiles():Seq[String]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#listFiles():Seq[String]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">listFiles</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span>
</span>
<p class="shortcomment cmt">Returns a list of file paths that are added to resources.</p>
</li><li name="org.apache.spark.SparkContext#listJars" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="listJars():Seq[String]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#listJars():Seq[String]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">listJars</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span>
</span>
<p class="shortcomment cmt">Returns a list of jar files that are added to resources.</p>
</li><li name="org.apache.spark.SparkContext#localProperties" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="localProperties:InheritableThreadLocal[java.util.Properties]"></a><a id="localProperties:InheritableThreadLocal[Properties]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#localProperties:InheritableThreadLocal[java.util.Properties]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">val</span>
</span>
<span class="symbol">
<span class="name">localProperties</span><span class="result">: <span class="extype" name="java.lang.InheritableThreadLocal">InheritableThreadLocal</span>[<span class="extype" name="java.util.Properties">Properties</span>]</span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected[<a href="index.html" class="extype" name="org.apache.spark">spark</a>] </dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#log" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="log:org.slf4j.Logger"></a><a id="log:Logger"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#log:org.slf4j.Logger" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">log</span><span class="result">: <span class="extype" name="org.slf4j.Logger">Logger</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#logDebug" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="logDebug(msg:=&gt;String,throwable:Throwable):Unit"></a><a id="logDebug(⇒String,Throwable):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#logDebug(msg:=&gt;String,throwable:Throwable):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">logDebug</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>, <span name="throwable">throwable: <span class="extype" name="scala.Throwable">Throwable</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#logDebug" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="logDebug(msg:=&gt;String):Unit"></a><a id="logDebug(⇒String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#logDebug(msg:=&gt;String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">logDebug</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#logError" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="logError(msg:=&gt;String,throwable:Throwable):Unit"></a><a id="logError(⇒String,Throwable):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#logError(msg:=&gt;String,throwable:Throwable):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">logError</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>, <span name="throwable">throwable: <span class="extype" name="scala.Throwable">Throwable</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#logError" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="logError(msg:=&gt;String):Unit"></a><a id="logError(⇒String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#logError(msg:=&gt;String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">logError</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#logInfo" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="logInfo(msg:=&gt;String,throwable:Throwable):Unit"></a><a id="logInfo(⇒String,Throwable):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#logInfo(msg:=&gt;String,throwable:Throwable):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">logInfo</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>, <span name="throwable">throwable: <span class="extype" name="scala.Throwable">Throwable</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#logInfo" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="logInfo(msg:=&gt;String):Unit"></a><a id="logInfo(⇒String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#logInfo(msg:=&gt;String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">logInfo</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#logName" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="logName:String"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#logName:String" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">logName</span><span class="result">: <span class="extype" name="java.lang.String">String</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#logTrace" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="logTrace(msg:=&gt;String,throwable:Throwable):Unit"></a><a id="logTrace(⇒String,Throwable):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#logTrace(msg:=&gt;String,throwable:Throwable):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">logTrace</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>, <span name="throwable">throwable: <span class="extype" name="scala.Throwable">Throwable</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#logTrace" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="logTrace(msg:=&gt;String):Unit"></a><a id="logTrace(⇒String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#logTrace(msg:=&gt;String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">logTrace</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#logWarning" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="logWarning(msg:=&gt;String,throwable:Throwable):Unit"></a><a id="logWarning(⇒String,Throwable):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#logWarning(msg:=&gt;String,throwable:Throwable):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">logWarning</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>, <span name="throwable">throwable: <span class="extype" name="scala.Throwable">Throwable</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.internal.Logging#logWarning" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="logWarning(msg:=&gt;String):Unit"></a><a id="logWarning(⇒String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#logWarning(msg:=&gt;String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">logWarning</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#longAccumulator" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="longAccumulator(name:String):org.apache.spark.util.LongAccumulator"></a><a id="longAccumulator(String):LongAccumulator"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#longAccumulator(name:String):org.apache.spark.util.LongAccumulator" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">longAccumulator</span><span class="params">(<span name="name">name: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <a href="util/LongAccumulator.html" class="extype" name="org.apache.spark.util.LongAccumulator">LongAccumulator</a></span>
</span>
<p class="shortcomment cmt">Create and register a long accumulator, which starts with 0 and accumulates inputs by <code>add</code>.</p>
</li><li name="org.apache.spark.SparkContext#longAccumulator" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="longAccumulator:org.apache.spark.util.LongAccumulator"></a><a id="longAccumulator:LongAccumulator"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#longAccumulator:org.apache.spark.util.LongAccumulator" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">longAccumulator</span><span class="result">: <a href="util/LongAccumulator.html" class="extype" name="org.apache.spark.util.LongAccumulator">LongAccumulator</a></span>
</span>
<p class="shortcomment cmt">Create and register a long accumulator, which starts with 0 and accumulates inputs by <code>add</code>.</p>
</li><li name="org.apache.spark.SparkContext#makeRDD" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="makeRDD[T](seq:Seq[(T,Seq[String])])(implicitevidence$3:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="makeRDD[T](Seq[(T,Seq[String])])(ClassTag[T]):RDD[T]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#makeRDD[T](seq:Seq[(T,Seq[String])])(implicitevidence$3:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">makeRDD</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="seq">seq: <span class="extype" name="scala.Seq">Seq</span>[(<span class="extype" name="org.apache.spark.SparkContext.makeRDD.T">T</span>, <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>])]</span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.makeRDD.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.makeRDD.T">T</span>]</span>
</span>
<p class="shortcomment cmt">Distribute a local Scala collection to form an RDD, with one or more
location preferences (hostnames of Spark nodes) for each object.</p><div class="fullcomment"><div class="comment cmt"><p>Distribute a local Scala collection to form an RDD, with one or more
location preferences (hostnames of Spark nodes) for each object.
Create a new partition for each collection item.</p></div><dl class="paramcmts block"><dt class="param">seq</dt><dd class="cmt"><p>list of tuples of data and location preferences (hostnames of Spark nodes)</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing data partitioned according to location preferences</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#makeRDD" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="makeRDD[T](seq:Seq[T],numSlices:Int)(implicitevidence$2:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="makeRDD[T](Seq[T],Int)(ClassTag[T]):RDD[T]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#makeRDD[T](seq:Seq[T],numSlices:Int)(implicitevidence$2:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">makeRDD</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="seq">seq: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="org.apache.spark.SparkContext.makeRDD.T">T</span>]</span>, <span name="numSlices">numSlices: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultParallelism:Int">defaultParallelism</a></span></span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.makeRDD.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.makeRDD.T">T</span>]</span>
</span>
<p class="shortcomment cmt">Distribute a local Scala collection to form an RDD.</p><div class="fullcomment"><div class="comment cmt"><p>Distribute a local Scala collection to form an RDD.</p><p>This method is identical to <code>parallelize</code>.</p></div><dl class="paramcmts block"><dt class="param">seq</dt><dd class="cmt"><p>Scala collection to distribute</p></dd><dt class="param">numSlices</dt><dd class="cmt"><p>number of partitions to divide the collection into</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing distributed collection</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#master" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="master:String"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#master:String" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">master</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span>
</span>
</li><li name="scala.AnyRef#ne" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="ne(x$1:AnyRef):Boolean"></a><a id="ne(AnyRef):Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#ne(x$1:AnyRef):Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">ne</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.AnyRef">AnyRef</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#newAPIHadoopFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="newAPIHadoopFile[K,V,F&lt;:org.apache.hadoop.mapreduce.InputFormat[K,V]](path:String,fClass:Class[F],kClass:Class[K],vClass:Class[V],conf:org.apache.hadoop.conf.Configuration):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="newAPIHadoopFile[K,V,F&lt;:InputFormat[K,V]](String,Class[F],Class[K],Class[V],Configuration):RDD[(K,V)]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#newAPIHadoopFile[K,V,F&lt;:org.apache.hadoop.mapreduce.InputFormat[K,V]](path:String,fClass:Class[F],kClass:Class[K],vClass:Class[V],conf:org.apache.hadoop.conf.Configuration):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">newAPIHadoopFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>, <span name="F">F &lt;: <span class="extype" name="org.apache.hadoop.mapreduce.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.V">V</span>]</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="fClass">fClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.F">F</span>]</span>, <span name="kClass">kClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.K">K</span>]</span>, <span name="vClass">vClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.V">V</span>]</span>, <span name="conf">conf: <span class="extype" name="org.apache.hadoop.conf.Configuration">Configuration</span> = <span class="symbol"><span class="name"><a href="#hadoopConfiguration:org.apache.hadoop.conf.Configuration">hadoopConfiguration</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.V">V</span>)]</span>
</span>
<p class="shortcomment cmt">Get an RDD for a given Hadoop file with an arbitrary new API InputFormat
and extra configuration options to pass to the input format.</p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a given Hadoop file with an arbitrary new API InputFormat
and extra configuration options to pass to the input format.
</p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths
as a list of inputs</p></dd><dt class="param">fClass</dt><dd class="cmt"><p>storage format of the data to be read</p></dd><dt class="param">kClass</dt><dd class="cmt"><p><code>Class</code> of the key associated with the <code>fClass</code> parameter</p></dd><dt class="param">vClass</dt><dd class="cmt"><p><code>Class</code> of the value associated with the <code>fClass</code> parameter</p></dd><dt class="param">conf</dt><dd class="cmt"><p>Hadoop configuration</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each
record, directly caching the returned RDD or directly passing it to an aggregation or shuffle
operation will create many references to the same object.
If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first
copy them using a <code>map</code> function.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#newAPIHadoopFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="newAPIHadoopFile[K,V,F&lt;:org.apache.hadoop.mapreduce.InputFormat[K,V]](path:String)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="newAPIHadoopFile[K,V,F&lt;:InputFormat[K,V]](String)(ClassTag[K],ClassTag[V],ClassTag[F]):RDD[(K,V)]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#newAPIHadoopFile[K,V,F&lt;:org.apache.hadoop.mapreduce.InputFormat[K,V]](path:String)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">newAPIHadoopFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>, <span name="F">F &lt;: <span class="extype" name="org.apache.hadoop.mapreduce.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.V">V</span>]</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="km">km: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.K">K</span>]</span>, <span name="vm">vm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.V">V</span>]</span>, <span name="fm">fm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.F">F</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.V">V</span>)]</span>
</span>
<p class="shortcomment cmt">Smarter version of <code>newApiHadoopFile</code> that uses class tags to figure out the classes of keys,
values and the <code>org.apache.hadoop.mapreduce.InputFormat</code> (new MapReduce API) so that user
don't need to pass them directly.</p><div class="fullcomment"><div class="comment cmt"><p>Smarter version of <code>newApiHadoopFile</code> that uses class tags to figure out the classes of keys,
values and the <code>org.apache.hadoop.mapreduce.InputFormat</code> (new MapReduce API) so that user
don't need to pass them directly. Instead, callers can just write, for example:
<code><code><code>
val file = sparkContext.hadoopFile[LongWritable, Text, TextInputFormat](path)
<code><code><code>
</code></code></code></code></code></code></p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths
as a list of inputs</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each
record, directly caching the returned RDD or directly passing it to an aggregation or shuffle
operation will create many references to the same object.
If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first
copy them using a <code>map</code> function.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#newAPIHadoopRDD" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="newAPIHadoopRDD[K,V,F&lt;:org.apache.hadoop.mapreduce.InputFormat[K,V]](conf:org.apache.hadoop.conf.Configuration,fClass:Class[F],kClass:Class[K],vClass:Class[V]):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="newAPIHadoopRDD[K,V,F&lt;:InputFormat[K,V]](Configuration,Class[F],Class[K],Class[V]):RDD[(K,V)]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#newAPIHadoopRDD[K,V,F&lt;:org.apache.hadoop.mapreduce.InputFormat[K,V]](conf:org.apache.hadoop.conf.Configuration,fClass:Class[F],kClass:Class[K],vClass:Class[V]):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">newAPIHadoopRDD</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>, <span name="F">F &lt;: <span class="extype" name="org.apache.hadoop.mapreduce.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.V">V</span>]</span>]</span><span class="params">(<span name="conf">conf: <span class="extype" name="org.apache.hadoop.conf.Configuration">Configuration</span> = <span class="symbol"><span class="name"><a href="#hadoopConfiguration:org.apache.hadoop.conf.Configuration">hadoopConfiguration</a></span></span></span>, <span name="fClass">fClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.F">F</span>]</span>, <span name="kClass">kClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.K">K</span>]</span>, <span name="vClass">vClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.V">V</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.V">V</span>)]</span>
</span>
<p class="shortcomment cmt">Get an RDD for a given Hadoop file with an arbitrary new API InputFormat
and extra configuration options to pass to the input format.</p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a given Hadoop file with an arbitrary new API InputFormat
and extra configuration options to pass to the input format.
</p></div><dl class="paramcmts block"><dt class="param">conf</dt><dd class="cmt"><p>Configuration for setting up the dataset. Note: This will be put into a Broadcast.
Therefore if you plan to reuse this conf to create multiple RDDs, you need to make
sure you won't modify the conf. A safe approach is always creating a new conf for
a new RDD.</p></dd><dt class="param">fClass</dt><dd class="cmt"><p>storage format of the data to be read</p></dd><dt class="param">kClass</dt><dd class="cmt"><p><code>Class</code> of the key associated with the <code>fClass</code> parameter</p></dd><dt class="param">vClass</dt><dd class="cmt"><p><code>Class</code> of the value associated with the <code>fClass</code> parameter</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each
record, directly caching the returned RDD or directly passing it to an aggregation or shuffle
operation will create many references to the same object.
If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first
copy them using a <code>map</code> function.</p></span></dd></dl></div>
</li><li name="scala.AnyRef#notify" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="notify():Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#notify():Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">notify</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd>
<span class="name">@native</span><span class="args">()</span>
</dd></dl></div>
</li><li name="scala.AnyRef#notifyAll" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="notifyAll():Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#notifyAll():Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">notifyAll</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd>
<span class="name">@native</span><span class="args">()</span>
</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#objectFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="objectFile[T](path:String,minPartitions:Int)(implicitevidence$4:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="objectFile[T](String,Int)(ClassTag[T]):RDD[T]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#objectFile[T](path:String,minPartitions:Int)(implicitevidence$4:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">objectFile</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.objectFile.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.objectFile.T">T</span>]</span>
</span>
<p class="shortcomment cmt">Load an RDD saved as a SequenceFile containing serialized objects, with NullWritable keys and
BytesWritable values that contain a serialized partition.</p><div class="fullcomment"><div class="comment cmt"><p>Load an RDD saved as a SequenceFile containing serialized objects, with NullWritable keys and
BytesWritable values that contain a serialized partition. This is still an experimental
storage format and may not be supported exactly as is in future Spark releases. It will also
be pretty slow if you use the default serializer (Java serialization),
though the nice thing about it is that there's very little effort required to save arbitrary
objects.
</p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths
as a list of inputs</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>suggested minimum number of partitions for the resulting RDD</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing deserialized data from the file(s)</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#parallelize" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="parallelize[T](seq:Seq[T],numSlices:Int)(implicitevidence$1:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="parallelize[T](Seq[T],Int)(ClassTag[T]):RDD[T]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#parallelize[T](seq:Seq[T],numSlices:Int)(implicitevidence$1:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">parallelize</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="seq">seq: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="org.apache.spark.SparkContext.parallelize.T">T</span>]</span>, <span name="numSlices">numSlices: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultParallelism:Int">defaultParallelism</a></span></span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.parallelize.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.parallelize.T">T</span>]</span>
</span>
<p class="shortcomment cmt">Distribute a local Scala collection to form an RDD.</p><div class="fullcomment"><div class="comment cmt"><p>Distribute a local Scala collection to form an RDD.
</p></div><dl class="paramcmts block"><dt class="param">seq</dt><dd class="cmt"><p>Scala collection to distribute</p></dd><dt class="param">numSlices</dt><dd class="cmt"><p>number of partitions to divide the collection into</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing distributed collection</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Parallelize acts lazily. If <code>seq</code> is a mutable collection and is altered after the call
to parallelize and before the first action on the RDD, the resultant RDD will reflect the
modified collection. Pass a copy of the argument to avoid this.</p></span>, <span class="cmt"><p>avoid using <code>parallelize(Seq())</code> to create an empty <code>RDD</code>. Consider <code>emptyRDD</code> for an
RDD with no partitions, or <code>parallelize(Seq[T]())</code> for an RDD of <code>T</code> with empty partitions.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#range" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="range(start:Long,end:Long,step:Long,numSlices:Int):org.apache.spark.rdd.RDD[Long]"></a><a id="range(Long,Long,Long,Int):RDD[Long]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#range(start:Long,end:Long,step:Long,numSlices:Int):org.apache.spark.rdd.RDD[Long]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">range</span><span class="params">(<span name="start">start: <span class="extype" name="scala.Long">Long</span></span>, <span name="end">end: <span class="extype" name="scala.Long">Long</span></span>, <span name="step">step: <span class="extype" name="scala.Long">Long</span> = <span class="symbol">1</span></span>, <span name="numSlices">numSlices: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultParallelism:Int">defaultParallelism</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="scala.Long">Long</span>]</span>
</span>
<p class="shortcomment cmt">Creates a new RDD[Long] containing elements from <code>start</code> to <code>end</code>(exclusive), increased by
<code>step</code> every element.</p><div class="fullcomment"><div class="comment cmt"><p>Creates a new RDD[Long] containing elements from <code>start</code> to <code>end</code>(exclusive), increased by
<code>step</code> every element.
</p></div><dl class="paramcmts block"><dt class="param">start</dt><dd class="cmt"><p>the start value.</p></dd><dt class="param">end</dt><dd class="cmt"><p>the end value.</p></dd><dt class="param">step</dt><dd class="cmt"><p>the incremental step</p></dd><dt class="param">numSlices</dt><dd class="cmt"><p>number of partitions to divide the collection into</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing distributed range</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>if we need to cache this RDD, we should make sure each partition does not exceed limit.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#register" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="register(acc:org.apache.spark.util.AccumulatorV2[_,_],name:String):Unit"></a><a id="register(AccumulatorV2[_,_],String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#register(acc:org.apache.spark.util.AccumulatorV2[_,_],name:String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">register</span><span class="params">(<span name="acc">acc: <a href="util/AccumulatorV2.html" class="extype" name="org.apache.spark.util.AccumulatorV2">AccumulatorV2</a>[_, _]</span>, <span name="name">name: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Register the given accumulator with given name.</p><div class="fullcomment"><div class="comment cmt"><p>Register the given accumulator with given name.
</p></div><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Accumulators must be registered before use, or it will throw exception.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#register" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="register(acc:org.apache.spark.util.AccumulatorV2[_,_]):Unit"></a><a id="register(AccumulatorV2[_,_]):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#register(acc:org.apache.spark.util.AccumulatorV2[_,_]):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">register</span><span class="params">(<span name="acc">acc: <a href="util/AccumulatorV2.html" class="extype" name="org.apache.spark.util.AccumulatorV2">AccumulatorV2</a>[_, _]</span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Register the given accumulator.</p><div class="fullcomment"><div class="comment cmt"><p>Register the given accumulator.
</p></div><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Accumulators must be registered before use, or it will throw exception.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#removeSparkListener" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="removeSparkListener(listener:org.apache.spark.scheduler.SparkListenerInterface):Unit"></a><a id="removeSparkListener(SparkListenerInterface):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#removeSparkListener(listener:org.apache.spark.scheduler.SparkListenerInterface):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">removeSparkListener</span><span class="params">(<span name="listener">listener: <span class="extype" name="org.apache.spark.scheduler.SparkListenerInterface">SparkListenerInterface</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">:: DeveloperApi ::
Deregister the listener from Spark's listener bus.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi ::
Deregister the listener from Spark's listener bus.
</p></div><dl class="attributes block"> <dt>Annotations</dt><dd>
<span class="name">@DeveloperApi</span><span class="args">()</span>
</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#requestExecutors" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="requestExecutors(numAdditionalExecutors:Int):Boolean"></a><a id="requestExecutors(Int):Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#requestExecutors(numAdditionalExecutors:Int):Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">requestExecutors</span><span class="params">(<span name="numAdditionalExecutors">numAdditionalExecutors: <span class="extype" name="scala.Int">Int</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<p class="shortcomment cmt">:: DeveloperApi ::
Request an additional number of executors from the cluster manager.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi ::
Request an additional number of executors from the cluster manager.</p></div><dl class="paramcmts block"><dt>returns</dt><dd class="cmt"><p>whether the request is received.</p></dd></dl><dl class="attributes block"> <dt>Annotations</dt><dd>
<span class="name">@DeveloperApi</span><span class="args">()</span>
</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#requestTotalExecutors" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="requestTotalExecutors(numExecutors:Int,localityAwareTasks:Int,hostToLocalTaskCount:scala.collection.immutable.Map[String,Int]):Boolean"></a><a id="requestTotalExecutors(Int,Int,Map[String,Int]):Boolean"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#requestTotalExecutors(numExecutors:Int,localityAwareTasks:Int,hostToLocalTaskCount:scala.collection.immutable.Map[String,Int]):Boolean" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">requestTotalExecutors</span><span class="params">(<span name="numExecutors">numExecutors: <span class="extype" name="scala.Int">Int</span></span>, <span name="localityAwareTasks">localityAwareTasks: <span class="extype" name="scala.Int">Int</span></span>, <span name="hostToLocalTaskCount">hostToLocalTaskCount: <span class="extype" name="scala.collection.immutable.Map">Map</span>[<span class="extype" name="scala.Predef.String">String</span>, <span class="extype" name="scala.Int">Int</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span>
</span>
<p class="shortcomment cmt">Update the cluster manager on our scheduling needs.</p><div class="fullcomment"><div class="comment cmt"><p>Update the cluster manager on our scheduling needs. Three bits of information are included
to help it make decisions. This applies to the default ResourceProfile.</p></div><dl class="paramcmts block"><dt class="param">numExecutors</dt><dd class="cmt"><p>The total number of executors we'd like to have. The cluster manager
shouldn't kill any running executor to reach this number, but,
if all existing executors were to die, this is the number of executors
we'd want to be allocated.</p></dd><dt class="param">localityAwareTasks</dt><dd class="cmt"><p>The number of tasks in all active stages that have a locality
preferences. This includes running, pending, and completed tasks.</p></dd><dt class="param">hostToLocalTaskCount</dt><dd class="cmt"><p>A map of hosts to the number of tasks from all active stages
that would like to like to run on that host.
This includes running, pending, and completed tasks.</p></dd><dt>returns</dt><dd class="cmt"><p>whether the request is acknowledged by the cluster manager.</p></dd></dl><dl class="attributes block"> <dt>Annotations</dt><dd>
<span class="name">@DeveloperApi</span><span class="args">()</span>
</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#resources" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="resources:scala.collection.Map[String,org.apache.spark.resource.ResourceInformation]"></a><a id="resources:Map[String,ResourceInformation]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#resources:scala.collection.Map[String,org.apache.spark.resource.ResourceInformation]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">resources</span><span class="result">: <span class="extype" name="scala.collection.Map">Map</span>[<span class="extype" name="scala.Predef.String">String</span>, <a href="resource/ResourceInformation.html" class="extype" name="org.apache.spark.resource.ResourceInformation">ResourceInformation</a>]</span>
</span>
</li><li name="org.apache.spark.SparkContext#runApproximateJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="runApproximateJob[T,U,R](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=&gt;U,evaluator:org.apache.spark.partial.ApproximateEvaluator[U,R],timeout:Long):org.apache.spark.partial.PartialResult[R]"></a><a id="runApproximateJob[T,U,R](RDD[T],(TaskContext,Iterator[T])⇒U,ApproximateEvaluator[U,R],Long):PartialResult[R]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#runApproximateJob[T,U,R](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=&gt;U,evaluator:org.apache.spark.partial.ApproximateEvaluator[U,R],timeout:Long):org.apache.spark.partial.PartialResult[R]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">runApproximateJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>, <span name="R">R</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runApproximateJob.T">T</span>]</span>, <span name="func">func: (<a href="TaskContext.html" class="extype" name="org.apache.spark.TaskContext">TaskContext</a>, <span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runApproximateJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runApproximateJob.U">U</span></span>, <span name="evaluator">evaluator: <span class="extype" name="org.apache.spark.partial.ApproximateEvaluator">ApproximateEvaluator</span>[<span class="extype" name="org.apache.spark.SparkContext.runApproximateJob.U">U</span>, <span class="extype" name="org.apache.spark.SparkContext.runApproximateJob.R">R</span>]</span>, <span name="timeout">timeout: <span class="extype" name="scala.Long">Long</span></span>)</span><span class="result">: <a href="partial/PartialResult.html" class="extype" name="org.apache.spark.partial.PartialResult">PartialResult</a>[<span class="extype" name="org.apache.spark.SparkContext.runApproximateJob.R">R</span>]</span>
</span>
<p class="shortcomment cmt">:: DeveloperApi ::
Run a job that can return approximate results.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi ::
Run a job that can return approximate results.
</p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">func</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">evaluator</dt><dd class="cmt"><p><code>ApproximateEvaluator</code> to receive the partial results</p></dd><dt class="param">timeout</dt><dd class="cmt"><p>maximum time to wait for the job, in milliseconds</p></dd><dt>returns</dt><dd class="cmt"><p>partial result (how partial depends on whether the job was finished before or
after timeout)</p></dd></dl><dl class="attributes block"> <dt>Annotations</dt><dd>
<span class="name">@DeveloperApi</span><span class="args">()</span>
</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],processPartition:Iterator[T]=&gt;U,resultHandler:(Int,U)=&gt;Unit)(implicitevidence$16:scala.reflect.ClassTag[U]):Unit"></a><a id="runJob[T,U](RDD[T],(Iterator[T])⇒U,(Int,U)⇒Unit)(ClassTag[U]):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],processPartition:Iterator[T]=&gt;U,resultHandler:(Int,U)=&gt;Unit)(implicitevidence$16:scala.reflect.ClassTag[U]):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="processPartition">processPartition: (<span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>, <span name="resultHandler">resultHandler: (<span class="extype" name="scala.Int">Int</span>, <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>) ⇒ <span class="extype" name="scala.Unit">Unit</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Run a job on all partitions in an RDD and pass the results to a handler function.</p><div class="fullcomment"><div class="comment cmt"><p>Run a job on all partitions in an RDD and pass the results to a handler function.
</p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">processPartition</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">resultHandler</dt><dd class="cmt"><p>callback to pass each result to</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],processPartition:(org.apache.spark.TaskContext,Iterator[T])=&gt;U,resultHandler:(Int,U)=&gt;Unit)(implicitevidence$15:scala.reflect.ClassTag[U]):Unit"></a><a id="runJob[T,U](RDD[T],(TaskContext,Iterator[T])⇒U,(Int,U)⇒Unit)(ClassTag[U]):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],processPartition:(org.apache.spark.TaskContext,Iterator[T])=&gt;U,resultHandler:(Int,U)=&gt;Unit)(implicitevidence$15:scala.reflect.ClassTag[U]):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="processPartition">processPartition: (<a href="TaskContext.html" class="extype" name="org.apache.spark.TaskContext">TaskContext</a>, <span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>, <span name="resultHandler">resultHandler: (<span class="extype" name="scala.Int">Int</span>, <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>) ⇒ <span class="extype" name="scala.Unit">Unit</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Run a job on all partitions in an RDD and pass the results to a handler function.</p><div class="fullcomment"><div class="comment cmt"><p>Run a job on all partitions in an RDD and pass the results to a handler function. The function
that is run against each partition additionally takes <code>TaskContext</code> argument.
</p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">processPartition</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">resultHandler</dt><dd class="cmt"><p>callback to pass each result to</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:Iterator[T]=&gt;U)(implicitevidence$14:scala.reflect.ClassTag[U]):Array[U]"></a><a id="runJob[T,U](RDD[T],(Iterator[T])⇒U)(ClassTag[U]):Array[U]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:Iterator[T]=&gt;U)(implicitevidence$14:scala.reflect.ClassTag[U]):Array[U]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="func">func: (<span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Array">Array</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>
</span>
<p class="shortcomment cmt">Run a job on all partitions in an RDD and return the results in an array.</p><div class="fullcomment"><div class="comment cmt"><p>Run a job on all partitions in an RDD and return the results in an array.
</p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">func</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt>returns</dt><dd class="cmt"><p>in-memory collection with a result of the job (each collection element will contain
a result from one partition)</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=&gt;U)(implicitevidence$13:scala.reflect.ClassTag[U]):Array[U]"></a><a id="runJob[T,U](RDD[T],(TaskContext,Iterator[T])⇒U)(ClassTag[U]):Array[U]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=&gt;U)(implicitevidence$13:scala.reflect.ClassTag[U]):Array[U]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="func">func: (<a href="TaskContext.html" class="extype" name="org.apache.spark.TaskContext">TaskContext</a>, <span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Array">Array</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>
</span>
<p class="shortcomment cmt">Run a job on all partitions in an RDD and return the results in an array.</p><div class="fullcomment"><div class="comment cmt"><p>Run a job on all partitions in an RDD and return the results in an array. The function
that is run against each partition additionally takes <code>TaskContext</code> argument.
</p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">func</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt>returns</dt><dd class="cmt"><p>in-memory collection with a result of the job (each collection element will contain
a result from one partition)</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:Iterator[T]=&gt;U,partitions:Seq[Int])(implicitevidence$12:scala.reflect.ClassTag[U]):Array[U]"></a><a id="runJob[T,U](RDD[T],(Iterator[T])⇒U,Seq[Int])(ClassTag[U]):Array[U]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:Iterator[T]=&gt;U,partitions:Seq[Int])(implicitevidence$12:scala.reflect.ClassTag[U]):Array[U]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="func">func: (<span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>, <span name="partitions">partitions: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Int">Int</span>]</span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Array">Array</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>
</span>
<p class="shortcomment cmt">Run a function on a given set of partitions in an RDD and return the results as an array.</p><div class="fullcomment"><div class="comment cmt"><p>Run a function on a given set of partitions in an RDD and return the results as an array.
</p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">func</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">partitions</dt><dd class="cmt"><p>set of partitions to run on; some jobs may not want to compute on all
partitions of the target RDD, e.g. for operations like <code>first()</code></p></dd><dt>returns</dt><dd class="cmt"><p>in-memory collection with a result of the job (each collection element will contain
a result from one partition)</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=&gt;U,partitions:Seq[Int])(implicitevidence$11:scala.reflect.ClassTag[U]):Array[U]"></a><a id="runJob[T,U](RDD[T],(TaskContext,Iterator[T])⇒U,Seq[Int])(ClassTag[U]):Array[U]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=&gt;U,partitions:Seq[Int])(implicitevidence$11:scala.reflect.ClassTag[U]):Array[U]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="func">func: (<a href="TaskContext.html" class="extype" name="org.apache.spark.TaskContext">TaskContext</a>, <span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>, <span name="partitions">partitions: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Int">Int</span>]</span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Array">Array</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>
</span>
<p class="shortcomment cmt">Run a function on a given set of partitions in an RDD and return the results as an array.</p><div class="fullcomment"><div class="comment cmt"><p>Run a function on a given set of partitions in an RDD and return the results as an array.
The function that is run against each partition additionally takes <code>TaskContext</code> argument.
</p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">func</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">partitions</dt><dd class="cmt"><p>set of partitions to run on; some jobs may not want to compute on all
partitions of the target RDD, e.g. for operations like <code>first()</code></p></dd><dt>returns</dt><dd class="cmt"><p>in-memory collection with a result of the job (each collection element will contain
a result from one partition)</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=&gt;U,partitions:Seq[Int],resultHandler:(Int,U)=&gt;Unit)(implicitevidence$10:scala.reflect.ClassTag[U]):Unit"></a><a id="runJob[T,U](RDD[T],(TaskContext,Iterator[T])⇒U,Seq[Int],(Int,U)⇒Unit)(ClassTag[U]):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=&gt;U,partitions:Seq[Int],resultHandler:(Int,U)=&gt;Unit)(implicitevidence$10:scala.reflect.ClassTag[U]):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="func">func: (<a href="TaskContext.html" class="extype" name="org.apache.spark.TaskContext">TaskContext</a>, <span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>, <span name="partitions">partitions: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Int">Int</span>]</span>, <span name="resultHandler">resultHandler: (<span class="extype" name="scala.Int">Int</span>, <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>) ⇒ <span class="extype" name="scala.Unit">Unit</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Run a function on a given set of partitions in an RDD and pass the results to the given
handler function.</p><div class="fullcomment"><div class="comment cmt"><p>Run a function on a given set of partitions in an RDD and pass the results to the given
handler function. This is the main entry point for all actions in Spark.
</p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">func</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">partitions</dt><dd class="cmt"><p>set of partitions to run on; some jobs may not want to compute on all
partitions of the target RDD, e.g. for operations like <code>first()</code></p></dd><dt class="param">resultHandler</dt><dd class="cmt"><p>callback to pass each result to</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#sequenceFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="sequenceFile[K,V](path:String,minPartitions:Int)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitkcf:()=&gt;org.apache.spark.WritableConverter[K],implicitvcf:()=&gt;org.apache.spark.WritableConverter[V]):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="sequenceFile[K,V](String,Int)(ClassTag[K],ClassTag[V],()⇒WritableConverter[K],()⇒WritableConverter[V]):RDD[(K,V)]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#sequenceFile[K,V](path:String,minPartitions:Int)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitkcf:()=&gt;org.apache.spark.WritableConverter[K],implicitvcf:()=&gt;org.apache.spark.WritableConverter[V]):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">sequenceFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="km">km: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>]</span>, <span name="vm">vm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>]</span>, <span name="kcf">kcf: () ⇒ <span class="extype" name="org.apache.spark.WritableConverter">WritableConverter</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>]</span>, <span name="vcf">vcf: () ⇒ <span class="extype" name="org.apache.spark.WritableConverter">WritableConverter</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>)]</span>
</span>
<p class="shortcomment cmt">Version of sequenceFile() for types implicitly convertible to Writables through a
WritableConverter.</p><div class="fullcomment"><div class="comment cmt"><p>Version of sequenceFile() for types implicitly convertible to Writables through a
WritableConverter. For example, to access a SequenceFile where the keys are Text and the
values are IntWritable, you could simply write</p><pre>sparkContext.sequenceFile[<span class="std">String</span>, <span class="std">Int</span>](path, ...)</pre><p>WritableConverters are provided in a somewhat strange way (by an implicit function) to support
both subclasses of Writable and types for which we define a converter (e.g. Int to
IntWritable). The most natural thing would've been to have implicit objects for the
converters, but then we couldn't have an object for every subclass of Writable (you can't
have a parameterized singleton object). We use functions instead to create a new converter
for the appropriate type. In addition, we pass the converter a ClassTag of its type to
allow it to figure out the Writable class to use in the subclass case.
</p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths
as a list of inputs</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>suggested minimum number of partitions for the resulting RDD</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each
record, directly caching the returned RDD or directly passing it to an aggregation or shuffle
operation will create many references to the same object.
If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first
copy them using a <code>map</code> function.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#sequenceFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="sequenceFile[K,V](path:String,keyClass:Class[K],valueClass:Class[V]):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="sequenceFile[K,V](String,Class[K],Class[V]):RDD[(K,V)]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#sequenceFile[K,V](path:String,keyClass:Class[K],valueClass:Class[V]):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">sequenceFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="keyClass">keyClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>]</span>, <span name="valueClass">valueClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>)]</span>
</span>
<p class="shortcomment cmt">Get an RDD for a Hadoop SequenceFile with given key and value types.</p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a Hadoop SequenceFile with given key and value types.
</p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths
as a list of inputs</p></dd><dt class="param">keyClass</dt><dd class="cmt"><p><code>Class</code> of the key associated with <code>SequenceFileInputFormat</code></p></dd><dt class="param">valueClass</dt><dd class="cmt"><p><code>Class</code> of the value associated with <code>SequenceFileInputFormat</code></p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each
record, directly caching the returned RDD or directly passing it to an aggregation or shuffle
operation will create many references to the same object.
If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first
copy them using a <code>map</code> function.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#sequenceFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="sequenceFile[K,V](path:String,keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="sequenceFile[K,V](String,Class[K],Class[V],Int):RDD[(K,V)]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#sequenceFile[K,V](path:String,keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">sequenceFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="keyClass">keyClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>]</span>, <span name="valueClass">valueClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>]</span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>)]</span>
</span>
<p class="shortcomment cmt">Get an RDD for a Hadoop SequenceFile with given key and value types.</p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a Hadoop SequenceFile with given key and value types.
</p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths
as a list of inputs</p></dd><dt class="param">keyClass</dt><dd class="cmt"><p><code>Class</code> of the key associated with <code>SequenceFileInputFormat</code></p></dd><dt class="param">valueClass</dt><dd class="cmt"><p><code>Class</code> of the value associated with <code>SequenceFileInputFormat</code></p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>suggested minimum number of partitions for the resulting RDD</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each
record, directly caching the returned RDD or directly passing it to an aggregation or shuffle
operation will create many references to the same object.
If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first
copy them using a <code>map</code> function.</p></span></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#setCallSite" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="setCallSite(shortCallSite:String):Unit"></a><a id="setCallSite(String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#setCallSite(shortCallSite:String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">setCallSite</span><span class="params">(<span name="shortCallSite">shortCallSite: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Set the thread-local property for overriding the call sites
of actions and RDDs.</p>
</li><li name="org.apache.spark.SparkContext#setCheckpointDir" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="setCheckpointDir(directory:String):Unit"></a><a id="setCheckpointDir(String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#setCheckpointDir(directory:String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">setCheckpointDir</span><span class="params">(<span name="directory">directory: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Set the directory under which RDDs are going to be checkpointed.</p><div class="fullcomment"><div class="comment cmt"><p>Set the directory under which RDDs are going to be checkpointed.</p></div><dl class="paramcmts block"><dt class="param">directory</dt><dd class="cmt"><p>path to the directory where checkpoint files will be stored
(must be HDFS path if running in cluster)</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#setJobDescription" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="setJobDescription(value:String):Unit"></a><a id="setJobDescription(String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#setJobDescription(value:String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">setJobDescription</span><span class="params">(<span name="value">value: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Set a human readable description of the current job.</p>
</li><li name="org.apache.spark.SparkContext#setJobGroup" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="setJobGroup(groupId:String,description:String,interruptOnCancel:Boolean):Unit"></a><a id="setJobGroup(String,String,Boolean):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#setJobGroup(groupId:String,description:String,interruptOnCancel:Boolean):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">setJobGroup</span><span class="params">(<span name="groupId">groupId: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="description">description: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="interruptOnCancel">interruptOnCancel: <span class="extype" name="scala.Boolean">Boolean</span> = <span class="symbol">false</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Assigns a group ID to all the jobs started by this thread until the group ID is set to a
different value or cleared.</p><div class="fullcomment"><div class="comment cmt"><p>Assigns a group ID to all the jobs started by this thread until the group ID is set to a
different value or cleared.</p><p>Often, a unit of execution in an application consists of multiple Spark actions or jobs.
Application programmers can use this method to group all those jobs together and give a
group description. Once set, the Spark web UI will associate such jobs with this group.</p><p>The application can also use <code>org.apache.spark.SparkContext.cancelJobGroup</code> to cancel all
running jobs in this group. For example,</p><pre><span class="cmt">// In the main thread:</span>
sc.setJobGroup(<span class="lit">"some_job_to_cancel"</span>, <span class="lit">"some job description"</span>)
sc.parallelize(<span class="num">1</span> to <span class="num">10000</span>, <span class="num">2</span>).map { i <span class="kw">=&gt;</span> Thread.sleep(<span class="num">10</span>); i }.count()
<span class="cmt">// In a separate thread:</span>
sc.cancelJobGroup(<span class="lit">"some_job_to_cancel"</span>)</pre></div><dl class="paramcmts block"><dt class="param">interruptOnCancel</dt><dd class="cmt"><p>If true, then job cancellation will result in <code>Thread.interrupt()</code>
being called on the job's executor threads. This is useful to help ensure that the tasks
are actually stopped in a timely manner, but is off by default due to HDFS-1208, where HDFS
may respond to Thread.interrupt() by marking nodes as dead.</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#setLocalProperty" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="setLocalProperty(key:String,value:String):Unit"></a><a id="setLocalProperty(String,String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#setLocalProperty(key:String,value:String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">setLocalProperty</span><span class="params">(<span name="key">key: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="value">value: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Set a local property that affects jobs submitted from this thread, such as the Spark fair
scheduler pool.</p><div class="fullcomment"><div class="comment cmt"><p>Set a local property that affects jobs submitted from this thread, such as the Spark fair
scheduler pool. User-defined properties may also be set here. These properties are propagated
through to worker tasks and can be accessed there via
<a href="TaskContext.html#getLocalProperty(key:String):String" class="extmbr" name="org.apache.spark.TaskContext#getLocalProperty">org.apache.spark.TaskContext#getLocalProperty</a>.</p><p>These properties are inherited by child threads spawned from this thread. This
may have unexpected consequences when working with thread pools. The standard java
implementation of thread pools have worker threads spawn other worker threads.
As a result, local properties may propagate unpredictably.
</p></div></div>
</li><li name="org.apache.spark.SparkContext#setLogLevel" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="setLogLevel(logLevel:String):Unit"></a><a id="setLogLevel(String):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#setLogLevel(logLevel:String):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">setLogLevel</span><span class="params">(<span name="logLevel">logLevel: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Control our logLevel.</p><div class="fullcomment"><div class="comment cmt"><p>Control our logLevel. This overrides any user-defined log settings.</p></div><dl class="paramcmts block"><dt class="param">logLevel</dt><dd class="cmt"><p>The desired log level as a string.
Valid log levels include: ALL, DEBUG, ERROR, FATAL, INFO, OFF, TRACE, WARN</p></dd></dl></div>
</li><li name="org.apache.spark.SparkContext#sparkUser" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="sparkUser:String"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#sparkUser:String" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">val</span>
</span>
<span class="symbol">
<span class="name">sparkUser</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span>
</span>
</li><li name="org.apache.spark.SparkContext#startTime" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="startTime:Long"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#startTime:Long" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">val</span>
</span>
<span class="symbol">
<span class="name">startTime</span><span class="result">: <span class="extype" name="scala.Long">Long</span></span>
</span>
</li><li name="org.apache.spark.SparkContext#statusTracker" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="statusTracker:org.apache.spark.SparkStatusTracker"></a><a id="statusTracker:SparkStatusTracker"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#statusTracker:org.apache.spark.SparkStatusTracker" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">statusTracker</span><span class="result">: <a href="SparkStatusTracker.html" class="extype" name="org.apache.spark.SparkStatusTracker">SparkStatusTracker</a></span>
</span>
</li><li name="org.apache.spark.SparkContext#stop" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="stop():Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#stop():Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">stop</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<p class="shortcomment cmt">Shut down the SparkContext.</p>
</li><li name="org.apache.spark.SparkContext#submitJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="submitJob[T,U,R](rdd:org.apache.spark.rdd.RDD[T],processPartition:Iterator[T]=&gt;U,partitions:Seq[Int],resultHandler:(Int,U)=&gt;Unit,resultFunc:=&gt;R):org.apache.spark.SimpleFutureAction[R]"></a><a id="submitJob[T,U,R](RDD[T],(Iterator[T])⇒U,Seq[Int],(Int,U)⇒Unit,⇒R):SimpleFutureAction[R]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#submitJob[T,U,R](rdd:org.apache.spark.rdd.RDD[T],processPartition:Iterator[T]=&gt;U,partitions:Seq[Int],resultHandler:(Int,U)=&gt;Unit,resultFunc:=&gt;R):org.apache.spark.SimpleFutureAction[R]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">submitJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>, <span name="R">R</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.submitJob.T">T</span>]</span>, <span name="processPartition">processPartition: (<span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.submitJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.submitJob.U">U</span></span>, <span name="partitions">partitions: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Int">Int</span>]</span>, <span name="resultHandler">resultHandler: (<span class="extype" name="scala.Int">Int</span>, <span class="extype" name="org.apache.spark.SparkContext.submitJob.U">U</span>) ⇒ <span class="extype" name="scala.Unit">Unit</span></span>, <span name="resultFunc">resultFunc: ⇒ <span class="extype" name="org.apache.spark.SparkContext.submitJob.R">R</span></span>)</span><span class="result">: <a href="SimpleFutureAction.html" class="extype" name="org.apache.spark.SimpleFutureAction">SimpleFutureAction</a>[<span class="extype" name="org.apache.spark.SparkContext.submitJob.R">R</span>]</span>
</span>
<p class="shortcomment cmt">Submit a job for execution and return a FutureJob holding the result.</p><div class="fullcomment"><div class="comment cmt"><p>Submit a job for execution and return a FutureJob holding the result.
</p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">processPartition</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">partitions</dt><dd class="cmt"><p>set of partitions to run on; some jobs may not want to compute on all
partitions of the target RDD, e.g. for operations like <code>first()</code></p></dd><dt class="param">resultHandler</dt><dd class="cmt"><p>callback to pass each result to</p></dd><dt class="param">resultFunc</dt><dd class="cmt"><p>function to be executed when the result is ready</p></dd></dl></div>
</li><li name="scala.AnyRef#synchronized" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="synchronized[T0](x$1:=&gt;T0):T0"></a><a id="synchronized[T0](⇒T0):T0"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#synchronized[T0](x$1:=&gt;T0):T0" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">synchronized</span><span class="tparams">[<span name="T0">T0</span>]</span><span class="params">(<span name="arg0">arg0: ⇒ <span class="extype" name="java.lang.AnyRef.synchronized.T0">T0</span></span>)</span><span class="result">: <span class="extype" name="java.lang.AnyRef.synchronized.T0">T0</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#textFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="textFile(path:String,minPartitions:Int):org.apache.spark.rdd.RDD[String]"></a><a id="textFile(String,Int):RDD[String]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#textFile(path:String,minPartitions:Int):org.apache.spark.rdd.RDD[String]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">textFile</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="scala.Predef.String">String</span>]</span>
</span>
<p class="shortcomment cmt">Read a text file from HDFS, a local file system (available on all nodes), or any
Hadoop-supported file system URI, and return it as an RDD of Strings.</p><div class="fullcomment"><div class="comment cmt"><p>Read a text file from HDFS, a local file system (available on all nodes), or any
Hadoop-supported file system URI, and return it as an RDD of Strings.
The text files must be encoded as UTF-8.
</p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>path to the text file on a supported file system</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>suggested minimum number of partitions for the resulting RDD</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of lines of the text file</p></dd></dl></div>
</li><li name="scala.AnyRef#toString" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="toString():String"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#toString():String" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">toString</span><span class="params">()</span><span class="result">: <span class="extype" name="java.lang.String">String</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#uiWebUrl" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="uiWebUrl:Option[String]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#uiWebUrl:Option[String]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">uiWebUrl</span><span class="result">: <span class="extype" name="scala.Option">Option</span>[<span class="extype" name="scala.Predef.String">String</span>]</span>
</span>
</li><li name="org.apache.spark.SparkContext#union" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="union[T](first:org.apache.spark.rdd.RDD[T],rest:org.apache.spark.rdd.RDD[T]*)(implicitevidence$7:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="union[T](RDD[T],RDD[T]*)(ClassTag[T]):RDD[T]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#union[T](first:org.apache.spark.rdd.RDD[T],rest:org.apache.spark.rdd.RDD[T]*)(implicitevidence$7:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">union</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="first">first: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]</span>, <span name="rest">rest: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]*</span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]</span>
</span>
<p class="shortcomment cmt">Build the union of a list of RDDs passed as variable-length arguments.</p>
</li><li name="org.apache.spark.SparkContext#union" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="union[T](rdds:Seq[org.apache.spark.rdd.RDD[T]])(implicitevidence$6:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="union[T](Seq[RDD[T]])(ClassTag[T]):RDD[T]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#union[T](rdds:Seq[org.apache.spark.rdd.RDD[T]])(implicitevidence$6:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">union</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="rdds">rdds: <span class="extype" name="scala.Seq">Seq</span>[<a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]]</span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]</span>
</span>
<p class="shortcomment cmt">Build the union of a list of RDDs.</p>
</li><li name="org.apache.spark.SparkContext#version" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped">
<a id="version:String"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#version:String" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">version</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span>
</span>
<p class="shortcomment cmt">The version of Spark on which this application is running.</p>
</li><li name="scala.AnyRef#wait" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="wait():Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#wait():Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">wait</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd>
<span class="name">@throws</span><span class="args">(<span>
<span class="defval" name="classOf[java.lang.InterruptedException]">...</span>
</span>)</span>
</dd></dl></div>
</li><li name="scala.AnyRef#wait" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="wait(x$1:Long,x$2:Int):Unit"></a><a id="wait(Long,Int):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#wait(x$1:Long,x$2:Int):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">wait</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.Long">Long</span></span>, <span name="arg1">arg1: <span class="extype" name="scala.Int">Int</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd>
<span class="name">@throws</span><span class="args">(<span>
<span class="defval" name="classOf[java.lang.InterruptedException]">...</span>
</span>)</span>
</dd></dl></div>
</li><li name="scala.AnyRef#wait" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="wait(x$1:Long):Unit"></a><a id="wait(Long):Unit"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#wait(x$1:Long):Unit" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier">final </span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">wait</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.Long">Long</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span>
</span>
<div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd>
<span class="name">@throws</span><span class="args">(<span>
<span class="defval" name="classOf[java.lang.InterruptedException]">...</span>
</span>)</span>
<span class="name">@native</span><span class="args">()</span>
</dd></dl></div>
</li><li name="org.apache.spark.SparkContext#wholeTextFiles" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped">
<a id="wholeTextFiles(path:String,minPartitions:Int):org.apache.spark.rdd.RDD[(String,String)]"></a><a id="wholeTextFiles(String,Int):RDD[(String,String)]"></a>
<span class="permalink">
<a href="../../../org/apache/spark/SparkContext.html#wholeTextFiles(path:String,minPartitions:Int):org.apache.spark.rdd.RDD[(String,String)]" title="Permalink">
<i class="material-icons"></i>
</a>
</span>
<span class="modifier_kind">
<span class="modifier"></span>
<span class="kind">def</span>
</span>
<span class="symbol">
<span class="name">wholeTextFiles</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="scala.Predef.String">String</span>, <span class="extype" name="scala.Predef.String">String</span>)]</span>
</span>
<p class="shortcomment cmt">Read a directory of text files from HDFS, a local file system (available on all nodes), or any
Hadoop-supported file system URI.</p><div class="fullcomment"><div class="comment cmt"><p>Read a directory of text files from HDFS, a local file system (available on all nodes), or any
Hadoop-supported file system URI. Each file is read as a single record and returned in a
key-value pair, where the key is the path of each file, the value is the content of each file.
The text files must be encoded as UTF-8.</p><p> For example, if you have the following files:</p><pre>hdfs:<span class="cmt">//a-hdfs-path/part-00000</span>
hdfs:<span class="cmt">//a-hdfs-path/part-00001</span>
...
hdfs:<span class="cmt">//a-hdfs-path/part-nnnnn</span></pre><p>Do <code>val rdd = sparkContext.wholeTextFile(&quot;hdfs://a-hdfs-path&quot;)</code>,</p><p> then <code>rdd</code> contains</p><pre>(a-hdfs-path/part-<span class="num">00000</span>, its content)
(a-hdfs-path/part-<span class="num">00001</span>, its content)
...
(a-hdfs-path/part-nnnnn, its content)</pre></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>Directory to the input data files, the path can be comma separated paths as the
list of inputs.</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>A suggestion value of the minimal splitting number for input data.</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing tuples of file path and the corresponding file content</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Small files are preferred, large file is also allowable, but may cause bad performance.</p></span>, <span class="cmt"><p>On some filesystems, <code>.../path/&#42;</code> can be a more efficient way to read all files
in a directory rather than <code>.../path/</code> or <code>.../path</code></p></span>, <span class="cmt"><p>Partitioning is determined by data locality. This may result in too few partitions
by default.</p></span></dd></dl></div>
</li>
</ol>
</div>
</div>
<div id="inheritedMembers">
<div class="parent" name="org.apache.spark.internal.Logging">
<h3>Inherited from <span class="extype" name="org.apache.spark.internal.Logging">Logging</span></h3>
</div><div class="parent" name="scala.AnyRef">
<h3>Inherited from <span class="extype" name="scala.AnyRef">AnyRef</span></h3>
</div><div class="parent" name="scala.Any">
<h3>Inherited from <span class="extype" name="scala.Any">Any</span></h3>
</div>
</div>
<div id="groupedMembers">
<div class="group" name="Ungrouped">
<h3>Ungrouped</h3>
</div>
</div>
</div>
<div id="tooltip"></div>
<div id="footer"> </div>
</body>
</div>
</div>
</div>
</body>
</html>