| <!DOCTYPE html > |
| <html> |
| <head> |
| <meta http-equiv="X-UA-Compatible" content="IE=edge" /> |
| <meta name="viewport" content="width=device-width, initial-scale=1.0, maximum-scale=1.0, user-scalable=no" /> |
| <title>Spark 3.3.4 ScalaDoc - org.apache.spark.SparkContext</title> |
| <meta name="description" content="Spark 3.3.4 ScalaDoc - org.apache.spark.SparkContext" /> |
| <meta name="keywords" content="Spark 3.3.4 ScalaDoc org.apache.spark.SparkContext" /> |
| <meta http-equiv="content-type" content="text/html; charset=UTF-8" /> |
| |
| |
| <link href="../../../lib/index.css" media="screen" type="text/css" rel="stylesheet" /> |
| <link href="../../../lib/template.css" media="screen" type="text/css" rel="stylesheet" /> |
| <link href="../../../lib/diagrams.css" media="screen" type="text/css" rel="stylesheet" id="diagrams-css" /> |
| <script type="text/javascript" src="../../../lib/jquery.min.js"></script> |
| <script type="text/javascript" src="../../../lib/jquery.panzoom.min.js"></script> |
| <script type="text/javascript" src="../../../lib/jquery.mousewheel.min.js"></script> |
| <script type="text/javascript" src="../../../lib/index.js"></script> |
| <script type="text/javascript" src="../../../index.js"></script> |
| <script type="text/javascript" src="../../../lib/scheduler.js"></script> |
| <script type="text/javascript" src="../../../lib/template.js"></script> |
| |
| <script type="text/javascript"> |
| /* this variable can be used by the JS to determine the path to the root document */ |
| var toRoot = '../../../'; |
| </script> |
| |
| </head> |
| <body> |
| <div id="search"> |
| <span id="doc-title">Spark 3.3.4 ScalaDoc<span id="doc-version"></span></span> |
| <span class="close-results"><span class="left"><</span> Back</span> |
| <div id="textfilter"> |
| <span class="input"> |
| <input autocapitalize="none" placeholder="Search" id="index-input" type="text" accesskey="/" /> |
| <i class="clear material-icons"></i> |
| <i id="search-icon" class="material-icons"></i> |
| </span> |
| </div> |
| </div> |
| <div id="search-results"> |
| <div id="search-progress"> |
| <div id="progress-fill"></div> |
| </div> |
| <div id="results-content"> |
| <div id="entity-results"></div> |
| <div id="member-results"></div> |
| </div> |
| </div> |
| <div id="content-scroll-container" style="-webkit-overflow-scrolling: touch;"> |
| <div id="content-container" style="-webkit-overflow-scrolling: touch;"> |
| <div id="subpackage-spacer"> |
| <div id="packages"> |
| <h1>Packages</h1> |
| <ul> |
| <li name="_root_.root" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="_root_"></a><a id="root:_root_"></a> |
| <span class="permalink"> |
| <a href="../../../index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="../../../index.html"><span class="name">root</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="../../../index.html" class="extype" name="_root_">root</a></dd></dl></div> |
| </li><li name="_root_.org" visbl="pub" class="indented1 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="org"></a><a id="org:org"></a> |
| <span class="permalink"> |
| <a href="../../../org/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="../../index.html"><span class="name">org</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="../../../index.html" class="extype" name="_root_">root</a></dd></dl></div> |
| </li><li name="org.apache" visbl="pub" class="indented2 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="apache"></a><a id="apache:apache"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="../index.html"><span class="name">apache</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="../../index.html" class="extype" name="org">org</a></dd></dl></div> |
| </li><li name="org.apache.spark" visbl="pub" class="indented3 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="spark"></a><a id="spark:spark"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="Core Spark functionality." href="index.html"><span class="name">spark</span></a> |
| </span> |
| |
| <p class="shortcomment cmt">Core Spark functionality.</p><div class="fullcomment"><div class="comment cmt"><p>Core Spark functionality. <a href="" class="extype" name="org.apache.spark.SparkContext">org.apache.spark.SparkContext</a> serves as the main entry point to |
| Spark, while <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">org.apache.spark.rdd.RDD</a> is the data type representing a distributed collection, |
| and provides most parallel operations.</p><p>In addition, <a href="rdd/PairRDDFunctions.html" class="extype" name="org.apache.spark.rdd.PairRDDFunctions">org.apache.spark.rdd.PairRDDFunctions</a> contains operations available only on RDDs |
| of key-value pairs, such as <code>groupByKey</code> and <code>join</code>; <a href="rdd/DoubleRDDFunctions.html" class="extype" name="org.apache.spark.rdd.DoubleRDDFunctions">org.apache.spark.rdd.DoubleRDDFunctions</a> |
| contains operations available only on RDDs of Doubles; and |
| <a href="rdd/SequenceFileRDDFunctions.html" class="extype" name="org.apache.spark.rdd.SequenceFileRDDFunctions">org.apache.spark.rdd.SequenceFileRDDFunctions</a> contains operations available on RDDs that can |
| be saved as SequenceFiles. These operations are automatically available on any RDD of the right |
| type (e.g. RDD[(Int, Int)] through implicit conversions.</p><p>Java programmers should reference the <a href="api/java/index.html" class="extype" name="org.apache.spark.api.java">org.apache.spark.api.java</a> package |
| for Spark programming APIs in Java.</p><p>Classes and methods marked with <span class="experimental badge" style="float: none;"> |
| Experimental</span> are user-facing features which have not been officially adopted by the |
| Spark project. These are subject to change or removal in minor releases.</p><p>Classes and methods marked with <span class="developer badge" style="float: none;"> |
| Developer API</span> are intended for advanced users want to extend Spark through lower |
| level interfaces. These are subject to changes or removal in minor releases. |
| </p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="../index.html" class="extype" name="org.apache">apache</a></dd></dl></div> |
| </li><li name="org.apache.spark.api" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="api"></a><a id="api:api"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/api/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="api/index.html"><span class="name">api</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.broadcast" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="broadcast"></a><a id="broadcast:broadcast"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/broadcast/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="Spark's broadcast variables, used to broadcast immutable datasets to all nodes." href="broadcast/index.html"><span class="name">broadcast</span></a> |
| </span> |
| |
| <p class="shortcomment cmt">Spark's broadcast variables, used to broadcast immutable datasets to all nodes.</p><div class="fullcomment"><div class="comment cmt"><p>Spark's broadcast variables, used to broadcast immutable datasets to all nodes. |
| </p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.graphx" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="graphx"></a><a id="graphx:graphx"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/graphx/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="ALPHA COMPONENT GraphX is a graph processing framework built on top of Spark." href="graphx/index.html"><span class="name">graphx</span></a> |
| </span> |
| |
| <p class="shortcomment cmt"><span class="badge" style="float: right;">ALPHA COMPONENT</span> |
| GraphX is a graph processing framework built on top of Spark.</p><div class="fullcomment"><div class="comment cmt"><p><span class="badge" style="float: right;">ALPHA COMPONENT</span> |
| GraphX is a graph processing framework built on top of Spark. |
| </p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.input" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="input"></a><a id="input:input"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/input/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="input/index.html"><span class="name">input</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.io" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="io"></a><a id="io:io"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/io/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="IO codecs used for compression." href="io/index.html"><span class="name">io</span></a> |
| </span> |
| |
| <p class="shortcomment cmt">IO codecs used for compression.</p><div class="fullcomment"><div class="comment cmt"><p>IO codecs used for compression. See <a href="io/CompressionCodec.html" class="extype" name="org.apache.spark.io.CompressionCodec">org.apache.spark.io.CompressionCodec</a>. |
| </p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.launcher" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="launcher"></a><a id="launcher:launcher"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/launcher/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="launcher/index.html"><span class="name">launcher</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.mapred" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="mapred"></a><a id="mapred:mapred"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/mapred/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="mapred/index.html"><span class="name">mapred</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.metrics" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="metrics"></a><a id="metrics:metrics"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/metrics/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="metrics/index.html"><span class="name">metrics</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.ml" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="ml"></a><a id="ml:ml"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/ml/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="DataFrame-based machine learning APIs to let users quickly assemble and configure practical machine learning pipelines." href="ml/index.html"><span class="name">ml</span></a> |
| </span> |
| |
| <p class="shortcomment cmt">DataFrame-based machine learning APIs to let users quickly assemble and configure practical |
| machine learning pipelines.</p><div class="fullcomment"><div class="comment cmt"><p>DataFrame-based machine learning APIs to let users quickly assemble and configure practical |
| machine learning pipelines. |
| </p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.mllib" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="mllib"></a><a id="mllib:mllib"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/mllib/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="RDD-based machine learning APIs (in maintenance mode)." href="mllib/index.html"><span class="name">mllib</span></a> |
| </span> |
| |
| <p class="shortcomment cmt">RDD-based machine learning APIs (in maintenance mode).</p><div class="fullcomment"><div class="comment cmt"><p>RDD-based machine learning APIs (in maintenance mode).</p><p>The <code>spark.mllib</code> package is in maintenance mode as of the Spark 2.0.0 release to encourage |
| migration to the DataFrame-based APIs under the <a href="ml/index.html" class="extype" name="org.apache.spark.ml">org.apache.spark.ml</a> package. |
| While in maintenance mode,</p><ul><li>no new features in the RDD-based <code>spark.mllib</code> package will be accepted, unless they block |
| implementing new features in the DataFrame-based <code>spark.ml</code> package;</li><li>bug fixes in the RDD-based APIs will still be accepted.</li></ul><p>The developers will continue adding more features to the DataFrame-based APIs in the 2.x series |
| to reach feature parity with the RDD-based APIs. |
| And once we reach feature parity, this package will be deprecated. |
| </p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd><dt>See also</dt><dd><span class="cmt"><p><a href="https://issues.apache.org/jira/browse/SPARK-4591">SPARK-4591</a> to track |
| the progress of feature parity</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.partial" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="partial"></a><a id="partial:partial"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/partial/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="Support for approximate results." href="partial/index.html"><span class="name">partial</span></a> |
| </span> |
| |
| <p class="shortcomment cmt">Support for approximate results.</p><div class="fullcomment"><div class="comment cmt"><p>Support for approximate results. This provides convenient api and also implementation for |
| approximate calculation. |
| </p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd><dt>See also</dt><dd><span class="cmt"><p><a href="rdd/RDD.html#countApprox(timeout:Long,confidence:Double):org.apache.spark.partial.PartialResult[org.apache.spark.partial.BoundedDouble]" class="extmbr" name="org.apache.spark.rdd.RDD#countApprox">org.apache.spark.rdd.RDD.countApprox</a></p></span></dd></dl></div> |
| </li><li name="org.apache.spark.rdd" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="rdd"></a><a id="rdd:rdd"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/rdd/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="Provides several RDD implementations." href="rdd/index.html"><span class="name">rdd</span></a> |
| </span> |
| |
| <p class="shortcomment cmt">Provides several RDD implementations.</p><div class="fullcomment"><div class="comment cmt"><p>Provides several RDD implementations. See <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">org.apache.spark.rdd.RDD</a>. |
| </p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.resource" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="resource"></a><a id="resource:resource"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/resource/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="resource/index.html"><span class="name">resource</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.scheduler" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="scheduler"></a><a id="scheduler:scheduler"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/scheduler/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="Spark's scheduling components." href="scheduler/index.html"><span class="name">scheduler</span></a> |
| </span> |
| |
| <p class="shortcomment cmt">Spark's scheduling components.</p><div class="fullcomment"><div class="comment cmt"><p>Spark's scheduling components. This includes the <code>org.apache.spark.scheduler.DAGScheduler</code> and |
| lower level <code>org.apache.spark.scheduler.TaskScheduler</code>. |
| </p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.security" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="security"></a><a id="security:security"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/security/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="security/index.html"><span class="name">security</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.serializer" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="serializer"></a><a id="serializer:serializer"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/serializer/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="Pluggable serializers for RDD and shuffle data." href="serializer/index.html"><span class="name">serializer</span></a> |
| </span> |
| |
| <p class="shortcomment cmt">Pluggable serializers for RDD and shuffle data.</p><div class="fullcomment"><div class="comment cmt"><p>Pluggable serializers for RDD and shuffle data. |
| </p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd><dt>See also</dt><dd><span class="cmt"><p><a href="serializer/Serializer.html" class="extype" name="org.apache.spark.serializer.Serializer">org.apache.spark.serializer.Serializer</a></p></span></dd></dl></div> |
| </li><li name="org.apache.spark.shuffle" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="shuffle"></a><a id="shuffle:shuffle"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/shuffle/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="shuffle/index.html"><span class="name">shuffle</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.sql" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="sql"></a><a id="sql:sql"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/sql/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="Allows the execution of relational queries, including those expressed in SQL using Spark." href="sql/index.html"><span class="name">sql</span></a> |
| </span> |
| |
| <p class="shortcomment cmt">Allows the execution of relational queries, including those expressed in SQL using Spark.</p><div class="fullcomment"><div class="comment cmt"><p>Allows the execution of relational queries, including those expressed in SQL using Spark. |
| </p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.status" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="status"></a><a id="status:status"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/status/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="status/index.html"><span class="name">status</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.storage" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="storage"></a><a id="storage:storage"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/storage/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="storage/index.html"><span class="name">storage</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.streaming" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="streaming"></a><a id="streaming:streaming"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/streaming/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="Spark Streaming functionality." href="streaming/index.html"><span class="name">streaming</span></a> |
| </span> |
| |
| <p class="shortcomment cmt">Spark Streaming functionality.</p><div class="fullcomment"><div class="comment cmt"><p>Spark Streaming functionality. <a href="streaming/StreamingContext.html" class="extype" name="org.apache.spark.streaming.StreamingContext">org.apache.spark.streaming.StreamingContext</a> serves as the main |
| entry point to Spark Streaming, while <a href="streaming/dstream/DStream.html" class="extype" name="org.apache.spark.streaming.dstream.DStream">org.apache.spark.streaming.dstream.DStream</a> is the data |
| type representing a continuous sequence of RDDs, representing a continuous stream of data.</p><p>In addition, <a href="streaming/dstream/PairDStreamFunctions.html" class="extype" name="org.apache.spark.streaming.dstream.PairDStreamFunctions">org.apache.spark.streaming.dstream.PairDStreamFunctions</a> contains operations |
| available only on DStreams |
| of key-value pairs, such as <code>groupByKey</code> and <code>reduceByKey</code>. These operations are automatically |
| available on any DStream of the right type (e.g. DStream[(Int, Int)] through implicit |
| conversions.</p><p>For the Java API of Spark Streaming, take a look at the |
| <a href="streaming/api/java/JavaStreamingContext.html" class="extype" name="org.apache.spark.streaming.api.java.JavaStreamingContext">org.apache.spark.streaming.api.java.JavaStreamingContext</a> which serves as the entry point, and |
| the <a href="streaming/api/java/JavaDStream.html" class="extype" name="org.apache.spark.streaming.api.java.JavaDStream">org.apache.spark.streaming.api.java.JavaDStream</a> and the |
| <a href="streaming/api/java/JavaPairDStream.html" class="extype" name="org.apache.spark.streaming.api.java.JavaPairDStream">org.apache.spark.streaming.api.java.JavaPairDStream</a> which have the DStream functionality. |
| </p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.unsafe" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="unsafe"></a><a id="unsafe:unsafe"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/unsafe/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="" href="unsafe/index.html"><span class="name">unsafe</span></a> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li name="org.apache.spark.util" visbl="pub" class="indented4 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="util"></a><a id="util:util"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/util/index.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">package</span> |
| </span> |
| <span class="symbol"> |
| <a title="Spark utilities." href="util/index.html"><span class="name">util</span></a> |
| </span> |
| |
| <p class="shortcomment cmt">Spark utilities.</p><div class="fullcomment"><div class="comment cmt"><p>Spark utilities. |
| </p></div><dl class="attributes block"> <dt>Definition Classes</dt><dd><a href="index.html" class="extype" name="org.apache.spark">spark</a></dd></dl></div> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="Aggregator.html" title=":: DeveloperApi :: A set of functions used to aggregate data."></a> |
| <a href="Aggregator.html" title=":: DeveloperApi :: A set of functions used to aggregate data.">Aggregator</a> |
| </li><li class="current-entities indented3"> |
| <a class="object" href="BarrierTaskContext$.html" title=""></a> |
| <a class="class" href="BarrierTaskContext.html" title=":: Experimental :: A TaskContext with extra contextual info and tooling for tasks in a barrier stage."></a> |
| <a href="BarrierTaskContext.html" title=":: Experimental :: A TaskContext with extra contextual info and tooling for tasks in a barrier stage.">BarrierTaskContext</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="BarrierTaskInfo.html" title=":: Experimental :: Carries all task infos of a barrier task."></a> |
| <a href="BarrierTaskInfo.html" title=":: Experimental :: Carries all task infos of a barrier task.">BarrierTaskInfo</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="ComplexFutureAction.html" title="A FutureAction for actions that could trigger multiple Spark jobs."></a> |
| <a href="ComplexFutureAction.html" title="A FutureAction for actions that could trigger multiple Spark jobs.">ComplexFutureAction</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="ContextAwareIterator.html" title=":: DeveloperApi :: A TaskContext aware iterator."></a> |
| <a href="ContextAwareIterator.html" title=":: DeveloperApi :: A TaskContext aware iterator.">ContextAwareIterator</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="Dependency.html" title=":: DeveloperApi :: Base class for dependencies."></a> |
| <a href="Dependency.html" title=":: DeveloperApi :: Base class for dependencies.">Dependency</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="ExceptionFailure.html" title=":: DeveloperApi :: Task failed due to a runtime exception."></a> |
| <a href="ExceptionFailure.html" title=":: DeveloperApi :: Task failed due to a runtime exception.">ExceptionFailure</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="ExecutorLostFailure.html" title=":: DeveloperApi :: The task failed because the executor that it was running on was lost."></a> |
| <a href="ExecutorLostFailure.html" title=":: DeveloperApi :: The task failed because the executor that it was running on was lost.">ExecutorLostFailure</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="FetchFailed.html" title=":: DeveloperApi :: Task failed to fetch shuffle data from a remote node."></a> |
| <a href="FetchFailed.html" title=":: DeveloperApi :: Task failed to fetch shuffle data from a remote node.">FetchFailed</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="trait" href="FutureAction.html" title="A future for the result of an action to support cancellation."></a> |
| <a href="FutureAction.html" title="A future for the result of an action to support cancellation.">FutureAction</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="HashPartitioner.html" title="A org.apache.spark.Partitioner that implements hash-based partitioning using Java's Object.hashCode."></a> |
| <a href="HashPartitioner.html" title="A org.apache.spark.Partitioner that implements hash-based partitioning using Java's Object.hashCode.">HashPartitioner</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="InterruptibleIterator.html" title=":: DeveloperApi :: An iterator that wraps around an existing iterator to provide task killing functionality."></a> |
| <a href="InterruptibleIterator.html" title=":: DeveloperApi :: An iterator that wraps around an existing iterator to provide task killing functionality.">InterruptibleIterator</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="JobExecutionStatus.html" title=""></a> |
| <a href="JobExecutionStatus.html" title="">JobExecutionStatus</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="trait" href="JobSubmitter.html" title="Handle via which a "run" function passed to a ComplexFutureAction can submit jobs for execution."></a> |
| <a href="JobSubmitter.html" title="Handle via which a "run" function passed to a ComplexFutureAction can submit jobs for execution.">JobSubmitter</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="NarrowDependency.html" title=":: DeveloperApi :: Base class for dependencies where each partition of the child RDD depends on a small number of partitions of the parent RDD."></a> |
| <a href="NarrowDependency.html" title=":: DeveloperApi :: Base class for dependencies where each partition of the child RDD depends on a small number of partitions of the parent RDD.">NarrowDependency</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="OneToOneDependency.html" title=":: DeveloperApi :: Represents a one-to-one dependency between partitions of the parent and child RDDs."></a> |
| <a href="OneToOneDependency.html" title=":: DeveloperApi :: Represents a one-to-one dependency between partitions of the parent and child RDDs.">OneToOneDependency</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="trait" href="Partition.html" title="An identifier for a partition in an RDD."></a> |
| <a href="Partition.html" title="An identifier for a partition in an RDD.">Partition</a> |
| </li><li class="current-entities indented3"> |
| <a class="object" href="Partitioner$.html" title=""></a> |
| <a class="class" href="Partitioner.html" title="An object that defines how the elements in a key-value pair RDD are partitioned by key."></a> |
| <a href="Partitioner.html" title="An object that defines how the elements in a key-value pair RDD are partitioned by key.">Partitioner</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="RangeDependency.html" title=":: DeveloperApi :: Represents a one-to-one dependency between ranges of partitions in the parent and child RDDs."></a> |
| <a href="RangeDependency.html" title=":: DeveloperApi :: Represents a one-to-one dependency between ranges of partitions in the parent and child RDDs.">RangeDependency</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="RangePartitioner.html" title="A org.apache.spark.Partitioner that partitions sortable records by range into roughly equal ranges."></a> |
| <a href="RangePartitioner.html" title="A org.apache.spark.Partitioner that partitions sortable records by range into roughly equal ranges.">RangePartitioner</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="object" href="Resubmitted$.html" title=":: DeveloperApi :: A org.apache.spark.scheduler.ShuffleMapTask that completed successfully earlier, but we lost the executor before the stage completed."></a> |
| <a href="Resubmitted$.html" title=":: DeveloperApi :: A org.apache.spark.scheduler.ShuffleMapTask that completed successfully earlier, but we lost the executor before the stage completed.">Resubmitted</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="SerializableWritable.html" title=""></a> |
| <a href="SerializableWritable.html" title="">SerializableWritable</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="ShuffleDependency.html" title=":: DeveloperApi :: Represents a dependency on the output of a shuffle stage."></a> |
| <a href="ShuffleDependency.html" title=":: DeveloperApi :: Represents a dependency on the output of a shuffle stage.">ShuffleDependency</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="SimpleFutureAction.html" title="A FutureAction holding the result of an action that triggers a single job."></a> |
| <a href="SimpleFutureAction.html" title="A FutureAction holding the result of an action that triggers a single job.">SimpleFutureAction</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="SparkConf.html" title="Configuration for a Spark application."></a> |
| <a href="SparkConf.html" title="Configuration for a Spark application.">SparkConf</a> |
| </li><li class="current-entities indented3"> |
| <a class="object" href="SparkContext$.html" title="The SparkContext object contains a number of implicit conversions and parameters for use with various Spark features."></a> |
| <a class="class" href="" title="Main entry point for Spark functionality."></a> |
| <a href="" title="Main entry point for Spark functionality.">SparkContext</a> |
| </li><li class="current-entities indented3"> |
| <a class="object" href="SparkEnv$.html" title=""></a> |
| <a class="class" href="SparkEnv.html" title=":: DeveloperApi :: Holds all the runtime environment objects for a running Spark instance (either master or worker), including the serializer, RpcEnv, block manager, map output tracker, etc."></a> |
| <a href="SparkEnv.html" title=":: DeveloperApi :: Holds all the runtime environment objects for a running Spark instance (either master or worker), including the serializer, RpcEnv, block manager, map output tracker, etc.">SparkEnv</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="SparkException.html" title=""></a> |
| <a href="SparkException.html" title="">SparkException</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="trait" href="SparkExecutorInfo.html" title=""></a> |
| <a href="SparkExecutorInfo.html" title="">SparkExecutorInfo</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="object" href="SparkFiles$.html" title="Resolves paths to files added through SparkContext.addFile()."></a> |
| <a href="SparkFiles$.html" title="Resolves paths to files added through SparkContext.addFile().">SparkFiles</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="SparkFirehoseListener.html" title=""></a> |
| <a href="SparkFirehoseListener.html" title="">SparkFirehoseListener</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="trait" href="SparkJobInfo.html" title=""></a> |
| <a href="SparkJobInfo.html" title="">SparkJobInfo</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="trait" href="SparkStageInfo.html" title=""></a> |
| <a href="SparkStageInfo.html" title="">SparkStageInfo</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="SparkStatusTracker.html" title="Low-level status reporting APIs for monitoring job and stage progress."></a> |
| <a href="SparkStatusTracker.html" title="Low-level status reporting APIs for monitoring job and stage progress.">SparkStatusTracker</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="trait" href="SparkThrowable.html" title=""></a> |
| <a href="SparkThrowable.html" title="">SparkThrowable</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="object" href="Success$.html" title=":: DeveloperApi :: Task succeeded."></a> |
| <a href="Success$.html" title=":: DeveloperApi :: Task succeeded.">Success</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="TaskCommitDenied.html" title=":: DeveloperApi :: Task requested the driver to commit, but was denied."></a> |
| <a href="TaskCommitDenied.html" title=":: DeveloperApi :: Task requested the driver to commit, but was denied.">TaskCommitDenied</a> |
| </li><li class="current-entities indented3"> |
| <a class="object" href="TaskContext$.html" title=""></a> |
| <a class="class" href="TaskContext.html" title="Contextual information about a task which can be read or mutated during execution."></a> |
| <a href="TaskContext.html" title="Contextual information about a task which can be read or mutated during execution.">TaskContext</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="trait" href="TaskEndReason.html" title=":: DeveloperApi :: Various possible reasons why a task ended."></a> |
| <a href="TaskEndReason.html" title=":: DeveloperApi :: Various possible reasons why a task ended.">TaskEndReason</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="trait" href="TaskFailedReason.html" title=":: DeveloperApi :: Various possible reasons why a task failed."></a> |
| <a href="TaskFailedReason.html" title=":: DeveloperApi :: Various possible reasons why a task failed.">TaskFailedReason</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="TaskKilled.html" title=":: DeveloperApi :: Task was killed intentionally and needs to be rescheduled."></a> |
| <a href="TaskKilled.html" title=":: DeveloperApi :: Task was killed intentionally and needs to be rescheduled.">TaskKilled</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="class" href="TaskKilledException.html" title=":: DeveloperApi :: Exception thrown when a task is explicitly killed (i.e., task failure is expected)."></a> |
| <a href="TaskKilledException.html" title=":: DeveloperApi :: Exception thrown when a task is explicitly killed (i.e., task failure is expected).">TaskKilledException</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="object" href="TaskResultLost$.html" title=":: DeveloperApi :: The task finished successfully, but the result was lost from the executor's block manager before it was fetched."></a> |
| <a href="TaskResultLost$.html" title=":: DeveloperApi :: The task finished successfully, but the result was lost from the executor's block manager before it was fetched.">TaskResultLost</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="object" href="UnknownReason$.html" title=":: DeveloperApi :: We don't know why the task ended -- for example, because of a ClassNotFound exception when deserializing the task result."></a> |
| <a href="UnknownReason$.html" title=":: DeveloperApi :: We don't know why the task ended -- for example, because of a ClassNotFound exception when deserializing the task result.">UnknownReason</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="object" href="WritableConverter$.html" title=""></a> |
| <a href="WritableConverter$.html" title="">WritableConverter</a> |
| </li><li class="current-entities indented3"> |
| <span class="separator"></span> |
| <a class="object" href="WritableFactory$.html" title=""></a> |
| <a href="WritableFactory$.html" title="">WritableFactory</a> |
| </li> |
| </ul> |
| </div> |
| </div> |
| <div id="content"> |
| <body class="class type"> |
| <div id="definition"> |
| <a href="SparkContext$.html" title="See companion object"><div class="big-circle class-companion-object">c</div></a> |
| <p id="owner"><a href="../../index.html" class="extype" name="org">org</a>.<a href="../index.html" class="extype" name="org.apache">apache</a>.<a href="index.html" class="extype" name="org.apache.spark">spark</a></p> |
| <h1><a href="SparkContext$.html" title="See companion object">SparkContext</a><span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span></h1> |
| <h3><span class="morelinks"><div> |
| Companion <a href="SparkContext$.html" title="See companion object">object SparkContext</a> |
| </div></span></h3> |
| </div> |
| |
| <h4 id="signature" class="signature"> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">class</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">SparkContext</span><span class="result"> extends <span class="extype" name="org.apache.spark.internal.Logging">Logging</span></span> |
| </span> |
| </h4> |
| |
| |
| <div id="comment" class="fullcommenttop"><div class="comment cmt"><p>Main entry point for Spark functionality. A SparkContext represents the connection to a Spark |
| cluster, and can be used to create RDDs, accumulators and broadcast variables on that cluster. |
| </p></div><dl class="attributes block"> <dt>Source</dt><dd><a href="https://github.com/apache/spark/tree/v3.3.4/core/src/main/scala/org/apache/spark/SparkContext.scala" target="_blank">SparkContext.scala</a></dd><dt>Note</dt><dd><span class="cmt"><p>Only one <code>SparkContext</code> should be active per JVM. You must <code>stop()</code> the |
| active <code>SparkContext</code> before creating a new one.</p></span></dd></dl><div class="toggleContainer block"> |
| <span class="toggle"> |
| Linear Supertypes |
| </span> |
| <div class="superTypes hiddenContent"><span class="extype" name="org.apache.spark.internal.Logging">Logging</span>, <span class="extype" name="scala.AnyRef">AnyRef</span>, <span class="extype" name="scala.Any">Any</span></div> |
| </div></div> |
| |
| |
| <div id="mbrsel"> |
| <div class="toggle"></div> |
| <div id="memberfilter"> |
| <i class="material-icons arrow"></i> |
| <span class="input"> |
| <input id="mbrsel-input" placeholder="Filter all members" type="text" accesskey="/" /> |
| </span> |
| <i class="clear material-icons"></i> |
| </div> |
| <div id="filterby"> |
| <div id="order"> |
| <span class="filtertype">Ordering</span> |
| <ol> |
| |
| <li class="alpha in"><span>Alphabetic</span></li> |
| <li class="inherit out"><span>By Inheritance</span></li> |
| </ol> |
| </div> |
| <div class="ancestors"> |
| <span class="filtertype">Inherited<br /> |
| </span> |
| <ol id="linearization"> |
| <li class="in" name="org.apache.spark.SparkContext"><span>SparkContext</span></li><li class="in" name="org.apache.spark.internal.Logging"><span>Logging</span></li><li class="in" name="scala.AnyRef"><span>AnyRef</span></li><li class="in" name="scala.Any"><span>Any</span></li> |
| </ol> |
| </div><div class="ancestors"> |
| <span class="filtertype"></span> |
| <ol> |
| <li class="hideall out"><span>Hide All</span></li> |
| <li class="showall in"><span>Show All</span></li> |
| </ol> |
| </div> |
| <div id="visbl"> |
| <span class="filtertype">Visibility</span> |
| <ol><li class="public in"><span>Public</span></li><li class="all out"><span>All</span></li></ol> |
| </div> |
| </div> |
| </div> |
| |
| <div id="template"> |
| <div id="allMembers"> |
| <div id="constructors" class="members"> |
| <h3>Instance Constructors</h3> |
| <ol><li name="org.apache.spark.SparkContext#<init>" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="<init>(master:String,appName:String,sparkHome:String,jars:Seq[String],environment:scala.collection.Map[String,String]):org.apache.spark.SparkContext"></a><a id="<init>:SparkContext"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#<init>(master:String,appName:String,sparkHome:String,jars:Seq[String],environment:scala.collection.Map[String,String]):org.apache.spark.SparkContext" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">new</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">SparkContext</span><span class="params">(<span name="master">master: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="appName">appName: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="sparkHome">sparkHome: <span class="extype" name="scala.Predef.String">String</span> = <span class="symbol">null</span></span>, <span name="jars">jars: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>] = <span class="symbol">Nil</span></span>, <span name="environment">environment: <span class="extype" name="scala.collection.Map">Map</span>[<span class="extype" name="scala.Predef.String">String</span>, <span class="extype" name="scala.Predef.String">String</span>] = <span class="symbol">Map()</span></span>)</span> |
| </span> |
| |
| <p class="shortcomment cmt">Alternative constructor that allows setting common Spark properties directly |
| </p><div class="fullcomment"><div class="comment cmt"><p>Alternative constructor that allows setting common Spark properties directly |
| </p></div><dl class="paramcmts block"><dt class="param">master</dt><dd class="cmt"><p>Cluster URL to connect to (e.g. mesos://host:port, spark://host:port, local[4]).</p></dd><dt class="param">appName</dt><dd class="cmt"><p>A name for your application, to display on the cluster web UI.</p></dd><dt class="param">sparkHome</dt><dd class="cmt"><p>Location where Spark is installed on cluster nodes.</p></dd><dt class="param">jars</dt><dd class="cmt"><p>Collection of JARs to send to the cluster. These can be paths on the local file |
| system or HDFS, HTTP, HTTPS, or FTP URLs.</p></dd><dt class="param">environment</dt><dd class="cmt"><p>Environment variables to set on worker nodes.</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#<init>" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="<init>(master:String,appName:String,conf:org.apache.spark.SparkConf):org.apache.spark.SparkContext"></a><a id="<init>:SparkContext"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#<init>(master:String,appName:String,conf:org.apache.spark.SparkConf):org.apache.spark.SparkContext" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">new</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">SparkContext</span><span class="params">(<span name="master">master: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="appName">appName: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="conf">conf: <a href="SparkConf.html" class="extype" name="org.apache.spark.SparkConf">SparkConf</a></span>)</span> |
| </span> |
| |
| <p class="shortcomment cmt">Alternative constructor that allows setting common Spark properties directly |
| </p><div class="fullcomment"><div class="comment cmt"><p>Alternative constructor that allows setting common Spark properties directly |
| </p></div><dl class="paramcmts block"><dt class="param">master</dt><dd class="cmt"><p>Cluster URL to connect to (e.g. mesos://host:port, spark://host:port, local[4]).</p></dd><dt class="param">appName</dt><dd class="cmt"><p>A name for your application, to display on the cluster web UI</p></dd><dt class="param">conf</dt><dd class="cmt"><p>a <a href="SparkConf.html" class="extype" name="org.apache.spark.SparkConf">org.apache.spark.SparkConf</a> object specifying other Spark parameters</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#<init>" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="<init>():org.apache.spark.SparkContext"></a><a id="<init>:SparkContext"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#<init>():org.apache.spark.SparkContext" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">new</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">SparkContext</span><span class="params">()</span> |
| </span> |
| |
| <p class="shortcomment cmt">Create a SparkContext that loads settings from system properties (for instance, when |
| launching with ./bin/spark-submit).</p> |
| </li><li name="org.apache.spark.SparkContext#<init>" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="<init>(config:org.apache.spark.SparkConf):org.apache.spark.SparkContext"></a><a id="<init>:SparkContext"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#<init>(config:org.apache.spark.SparkConf):org.apache.spark.SparkContext" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">new</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">SparkContext</span><span class="params">(<span name="config">config: <a href="SparkConf.html" class="extype" name="org.apache.spark.SparkConf">SparkConf</a></span>)</span> |
| </span> |
| |
| <p class="shortcomment cmt"></p><div class="fullcomment"><div class="comment cmt"></div><dl class="paramcmts block"><dt class="param">config</dt><dd class="cmt"><p>a Spark Config object describing the application configuration. Any settings in |
| this config overrides the default configs as well as system properties.</p></dd></dl></div> |
| </li></ol> |
| </div> |
| |
| |
| |
| |
| |
| <div class="values members"> |
| <h3>Value Members</h3> |
| <ol> |
| <li name="scala.AnyRef#!=" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="!=(x$1:Any):Boolean"></a><a id="!=(Any):Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#!=(x$1:Any):Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span title="gt4s: $bang$eq" class="name">!=</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.Any">Any</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd></dl></div> |
| </li><li name="scala.AnyRef###" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="##():Int"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html###():Int" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span title="gt4s: $hash$hash" class="name">##</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Int">Int</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd></dl></div> |
| </li><li name="scala.AnyRef#==" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="==(x$1:Any):Boolean"></a><a id="==(Any):Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#==(x$1:Any):Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span title="gt4s: $eq$eq" class="name">==</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.Any">Any</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#addArchive" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="addArchive(path:String):Unit"></a><a id="addArchive(String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#addArchive(path:String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">addArchive</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">:: Experimental :: |
| Add an archive to be downloaded and unpacked with this Spark job on every node.</p><div class="fullcomment"><div class="comment cmt"><p>:: Experimental :: |
| Add an archive to be downloaded and unpacked with this Spark job on every node.</p><p>If an archive is added during execution, it will not be available until the next TaskSet |
| starts. |
| </p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>can be either a local file, a file in HDFS (or other Hadoop-supported |
| filesystems), or an HTTP, HTTPS or FTP URI. To access the file in Spark jobs, |
| use <code>SparkFiles.get(paths-to-files)</code> to find its download/unpacked location. |
| The given path should be one of .zip, .tar, .tar.gz, .tgz and .jar.</p></dd></dl><dl class="attributes block"> <dt>Annotations</dt><dd> |
| <span class="name">@Experimental</span><span class="args">()</span> |
| |
| </dd><dt>Since</dt><dd><p>3.1.0</p></dd><dt>Note</dt><dd><span class="cmt"><p>A path can be added only once. Subsequent additions of the same path are ignored.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#addFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="addFile(path:String,recursive:Boolean):Unit"></a><a id="addFile(String,Boolean):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#addFile(path:String,recursive:Boolean):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">addFile</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="recursive">recursive: <span class="extype" name="scala.Boolean">Boolean</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Add a file to be downloaded with this Spark job on every node.</p><div class="fullcomment"><div class="comment cmt"><p>Add a file to be downloaded with this Spark job on every node.</p><p>If a file is added during execution, it will not be available until the next TaskSet starts. |
| </p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>can be either a local file, a file in HDFS (or other Hadoop-supported |
| filesystems), or an HTTP, HTTPS or FTP URI. To access the file in Spark jobs, |
| use <code>SparkFiles.get(fileName)</code> to find its download location.</p></dd><dt class="param">recursive</dt><dd class="cmt"><p>if true, a directory can be given in <code>path</code>. Currently directories are |
| only supported for Hadoop-supported filesystems.</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>A path can be added only once. Subsequent additions of the same path are ignored.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#addFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="addFile(path:String):Unit"></a><a id="addFile(String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#addFile(path:String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">addFile</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Add a file to be downloaded with this Spark job on every node.</p><div class="fullcomment"><div class="comment cmt"><p>Add a file to be downloaded with this Spark job on every node.</p><p>If a file is added during execution, it will not be available until the next TaskSet starts. |
| </p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>can be either a local file, a file in HDFS (or other Hadoop-supported |
| filesystems), or an HTTP, HTTPS or FTP URI. To access the file in Spark jobs, |
| use <code>SparkFiles.get(fileName)</code> to find its download location.</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>A path can be added only once. Subsequent additions of the same path are ignored.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#addJar" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="addJar(path:String):Unit"></a><a id="addJar(String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#addJar(path:String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">addJar</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Adds a JAR dependency for all tasks to be executed on this <code>SparkContext</code> in the future.</p><div class="fullcomment"><div class="comment cmt"><p>Adds a JAR dependency for all tasks to be executed on this <code>SparkContext</code> in the future.</p><p>If a jar is added during execution, it will not be available until the next TaskSet starts. |
| </p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>can be either a local file, a file in HDFS (or other Hadoop-supported filesystems), |
| an HTTP, HTTPS or FTP URI, or local:/path for a file on every worker node.</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>A path can be added only once. Subsequent additions of the same path are ignored.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#addSparkListener" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="addSparkListener(listener:org.apache.spark.scheduler.SparkListenerInterface):Unit"></a><a id="addSparkListener(SparkListenerInterface):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#addSparkListener(listener:org.apache.spark.scheduler.SparkListenerInterface):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">addSparkListener</span><span class="params">(<span name="listener">listener: <span class="extype" name="org.apache.spark.scheduler.SparkListenerInterface">SparkListenerInterface</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">:: DeveloperApi :: |
| Register a listener to receive up-calls from events that happen during execution.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi :: |
| Register a listener to receive up-calls from events that happen during execution. |
| </p></div><dl class="attributes block"> <dt>Annotations</dt><dd> |
| <span class="name">@DeveloperApi</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#appName" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="appName:String"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#appName:String" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">appName</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span> |
| </span> |
| |
| |
| </li><li name="org.apache.spark.SparkContext#applicationAttemptId" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="applicationAttemptId:Option[String]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#applicationAttemptId:Option[String]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">applicationAttemptId</span><span class="result">: <span class="extype" name="scala.Option">Option</span>[<span class="extype" name="scala.Predef.String">String</span>]</span> |
| </span> |
| |
| |
| </li><li name="org.apache.spark.SparkContext#applicationId" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="applicationId:String"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#applicationId:String" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">applicationId</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">A unique identifier for the Spark application.</p><div class="fullcomment"><div class="comment cmt"><p>A unique identifier for the Spark application. |
| Its format depends on the scheduler implementation. |
| (i.e. |
| in case of local spark app something like 'local-1433865536131' |
| in case of YARN something like 'application_1433865536131_34483' |
| in case of MESOS something like 'driver-20170926223339-0001' |
| ) |
| </p></div></div> |
| </li><li name="org.apache.spark.SparkContext#archives" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="archives:Seq[String]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#archives:Seq[String]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">archives</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span> |
| </span> |
| |
| |
| </li><li name="scala.Any#asInstanceOf" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="asInstanceOf[T0]:T0"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#asInstanceOf[T0]:T0" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">asInstanceOf</span><span class="tparams">[<span name="T0">T0</span>]</span><span class="result">: <span class="extype" name="scala.Any.asInstanceOf.T0">T0</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>Any</dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#binaryFiles" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="binaryFiles(path:String,minPartitions:Int):org.apache.spark.rdd.RDD[(String,org.apache.spark.input.PortableDataStream)]"></a><a id="binaryFiles(String,Int):RDD[(String,PortableDataStream)]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#binaryFiles(path:String,minPartitions:Int):org.apache.spark.rdd.RDD[(String,org.apache.spark.input.PortableDataStream)]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">binaryFiles</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="scala.Predef.String">String</span>, <a href="input/PortableDataStream.html" class="extype" name="org.apache.spark.input.PortableDataStream">PortableDataStream</a>)]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Get an RDD for a Hadoop-readable dataset as PortableDataStream for each file |
| (useful for binary data)</p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a Hadoop-readable dataset as PortableDataStream for each file |
| (useful for binary data)</p><p>For example, if you have the following files:</p><pre>hdfs:<span class="cmt">//a-hdfs-path/part-00000</span> |
| hdfs:<span class="cmt">//a-hdfs-path/part-00001</span> |
| ... |
| hdfs:<span class="cmt">//a-hdfs-path/part-nnnnn</span></pre><p>Do |
| <code>val rdd = sparkContext.binaryFiles("hdfs://a-hdfs-path")</code>,</p><p>then <code>rdd</code> contains</p><pre>(a-hdfs-path/part-<span class="num">00000</span>, its content) |
| (a-hdfs-path/part-<span class="num">00001</span>, its content) |
| ... |
| (a-hdfs-path/part-nnnnn, its content)</pre></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>Directory to the input data files, the path can be comma separated paths as the |
| list of inputs.</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>A suggestion value of the minimal splitting number for input data.</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing tuples of file path and corresponding file content</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Small files are preferred; very large files may cause bad performance.</p></span>, <span class="cmt"><p>On some filesystems, <code>.../path/*</code> can be a more efficient way to read all files |
| in a directory rather than <code>.../path/</code> or <code>.../path</code></p></span>, <span class="cmt"><p>Partitioning is determined by data locality. This may result in too few partitions |
| by default.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#binaryRecords" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="binaryRecords(path:String,recordLength:Int,conf:org.apache.hadoop.conf.Configuration):org.apache.spark.rdd.RDD[Array[Byte]]"></a><a id="binaryRecords(String,Int,Configuration):RDD[Array[Byte]]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#binaryRecords(path:String,recordLength:Int,conf:org.apache.hadoop.conf.Configuration):org.apache.spark.rdd.RDD[Array[Byte]]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">binaryRecords</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="recordLength">recordLength: <span class="extype" name="scala.Int">Int</span></span>, <span name="conf">conf: <span class="extype" name="org.apache.hadoop.conf.Configuration">Configuration</span> = <span class="symbol"><span class="name"><a href="#hadoopConfiguration:org.apache.hadoop.conf.Configuration">hadoopConfiguration</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="scala.Array">Array</span>[<span class="extype" name="scala.Byte">Byte</span>]]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Load data from a flat binary file, assuming the length of each record is constant.</p><div class="fullcomment"><div class="comment cmt"><p>Load data from a flat binary file, assuming the length of each record is constant. |
| </p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>Directory to the input data files, the path can be comma separated paths as the |
| list of inputs.</p></dd><dt class="param">recordLength</dt><dd class="cmt"><p>The length at which to split the records</p></dd><dt class="param">conf</dt><dd class="cmt"><p>Configuration for setting up the dataset.</p></dd><dt>returns</dt><dd class="cmt"><p>An RDD of data with values, represented as byte arrays</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>We ensure that the byte array for each record in the resulting RDD |
| has the provided record length.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#broadcast" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="broadcast[T](value:T)(implicitevidence$9:scala.reflect.ClassTag[T]):org.apache.spark.broadcast.Broadcast[T]"></a><a id="broadcast[T](T)(ClassTag[T]):Broadcast[T]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#broadcast[T](value:T)(implicitevidence$9:scala.reflect.ClassTag[T]):org.apache.spark.broadcast.Broadcast[T]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">broadcast</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="value">value: <span class="extype" name="org.apache.spark.SparkContext.broadcast.T">T</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.broadcast.T">T</span>]</span>)</span><span class="result">: <a href="broadcast/Broadcast.html" class="extype" name="org.apache.spark.broadcast.Broadcast">Broadcast</a>[<span class="extype" name="org.apache.spark.SparkContext.broadcast.T">T</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Broadcast a read-only variable to the cluster, returning a |
| <a href="broadcast/Broadcast.html" class="extype" name="org.apache.spark.broadcast.Broadcast">org.apache.spark.broadcast.Broadcast</a> object for reading it in distributed functions.</p><div class="fullcomment"><div class="comment cmt"><p>Broadcast a read-only variable to the cluster, returning a |
| <a href="broadcast/Broadcast.html" class="extype" name="org.apache.spark.broadcast.Broadcast">org.apache.spark.broadcast.Broadcast</a> object for reading it in distributed functions. |
| The variable will be sent to each cluster only once. |
| </p></div><dl class="paramcmts block"><dt class="param">value</dt><dd class="cmt"><p>value to broadcast to the Spark nodes</p></dd><dt>returns</dt><dd class="cmt"><p><code>Broadcast</code> object, a read-only variable cached on each machine</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#cancelAllJobs" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="cancelAllJobs():Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#cancelAllJobs():Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">cancelAllJobs</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Cancel all jobs that have been scheduled or are running.</p> |
| </li><li name="org.apache.spark.SparkContext#cancelJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="cancelJob(jobId:Int):Unit"></a><a id="cancelJob(Int):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#cancelJob(jobId:Int):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">cancelJob</span><span class="params">(<span name="jobId">jobId: <span class="extype" name="scala.Int">Int</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Cancel a given job if it's scheduled or running.</p><div class="fullcomment"><div class="comment cmt"><p>Cancel a given job if it's scheduled or running. |
| </p></div><dl class="paramcmts block"><dt class="param">jobId</dt><dd class="cmt"><p>the job ID to cancel</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Throws <code>InterruptedException</code> if the cancel message cannot be sent</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#cancelJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="cancelJob(jobId:Int,reason:String):Unit"></a><a id="cancelJob(Int,String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#cancelJob(jobId:Int,reason:String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">cancelJob</span><span class="params">(<span name="jobId">jobId: <span class="extype" name="scala.Int">Int</span></span>, <span name="reason">reason: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Cancel a given job if it's scheduled or running.</p><div class="fullcomment"><div class="comment cmt"><p>Cancel a given job if it's scheduled or running. |
| </p></div><dl class="paramcmts block"><dt class="param">jobId</dt><dd class="cmt"><p>the job ID to cancel</p></dd><dt class="param">reason</dt><dd class="cmt"><p>optional reason for cancellation</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Throws <code>InterruptedException</code> if the cancel message cannot be sent</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#cancelJobGroup" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="cancelJobGroup(groupId:String):Unit"></a><a id="cancelJobGroup(String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#cancelJobGroup(groupId:String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">cancelJobGroup</span><span class="params">(<span name="groupId">groupId: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Cancel active jobs for the specified group.</p><div class="fullcomment"><div class="comment cmt"><p>Cancel active jobs for the specified group. See <code>org.apache.spark.SparkContext.setJobGroup</code> |
| for more information. |
| </p></div></div> |
| </li><li name="org.apache.spark.SparkContext#cancelStage" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="cancelStage(stageId:Int):Unit"></a><a id="cancelStage(Int):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#cancelStage(stageId:Int):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">cancelStage</span><span class="params">(<span name="stageId">stageId: <span class="extype" name="scala.Int">Int</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Cancel a given stage and all jobs associated with it.</p><div class="fullcomment"><div class="comment cmt"><p>Cancel a given stage and all jobs associated with it. |
| </p></div><dl class="paramcmts block"><dt class="param">stageId</dt><dd class="cmt"><p>the stage ID to cancel</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Throws <code>InterruptedException</code> if the cancel message cannot be sent</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#cancelStage" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="cancelStage(stageId:Int,reason:String):Unit"></a><a id="cancelStage(Int,String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#cancelStage(stageId:Int,reason:String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">cancelStage</span><span class="params">(<span name="stageId">stageId: <span class="extype" name="scala.Int">Int</span></span>, <span name="reason">reason: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Cancel a given stage and all jobs associated with it.</p><div class="fullcomment"><div class="comment cmt"><p>Cancel a given stage and all jobs associated with it. |
| </p></div><dl class="paramcmts block"><dt class="param">stageId</dt><dd class="cmt"><p>the stage ID to cancel</p></dd><dt class="param">reason</dt><dd class="cmt"><p>reason for cancellation</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Throws <code>InterruptedException</code> if the cancel message cannot be sent</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#checkpointFile" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="checkpointFile[T](path:String)(implicitevidence$5:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="checkpointFile[T](String)(ClassTag[T]):RDD[T]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#checkpointFile[T](path:String)(implicitevidence$5:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">checkpointFile</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.checkpointFile.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.checkpointFile.T">T</span>]</span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected[<a href="index.html" class="extype" name="org.apache.spark">spark</a>] </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#clearCallSite" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="clearCallSite():Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#clearCallSite():Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">clearCallSite</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Clear the thread-local property for overriding the call sites |
| of actions and RDDs.</p> |
| </li><li name="org.apache.spark.SparkContext#clearJobGroup" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="clearJobGroup():Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#clearJobGroup():Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">clearJobGroup</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Clear the current thread's job group ID and its description.</p> |
| </li><li name="scala.AnyRef#clone" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="clone():Object"></a><a id="clone():AnyRef"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#clone():Object" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">clone</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.AnyRef">AnyRef</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected[<span class="extype" name="java.lang">lang</span>] </dd><dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd> |
| <span class="name">@throws</span><span class="args">(<span> |
| |
| <span class="defval" name="classOf[java.lang.CloneNotSupportedException]">...</span> |
| </span>)</span> |
| |
| <span class="name">@native</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#collectionAccumulator" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="collectionAccumulator[T](name:String):org.apache.spark.util.CollectionAccumulator[T]"></a><a id="collectionAccumulator[T](String):CollectionAccumulator[T]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#collectionAccumulator[T](name:String):org.apache.spark.util.CollectionAccumulator[T]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">collectionAccumulator</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="name">name: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <a href="util/CollectionAccumulator.html" class="extype" name="org.apache.spark.util.CollectionAccumulator">CollectionAccumulator</a>[<span class="extype" name="org.apache.spark.SparkContext.collectionAccumulator.T">T</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Create and register a <code>CollectionAccumulator</code>, which starts with empty list and accumulates |
| inputs by adding them into the list.</p> |
| </li><li name="org.apache.spark.SparkContext#collectionAccumulator" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="collectionAccumulator[T]:org.apache.spark.util.CollectionAccumulator[T]"></a><a id="collectionAccumulator[T]:CollectionAccumulator[T]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#collectionAccumulator[T]:org.apache.spark.util.CollectionAccumulator[T]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">collectionAccumulator</span><span class="tparams">[<span name="T">T</span>]</span><span class="result">: <a href="util/CollectionAccumulator.html" class="extype" name="org.apache.spark.util.CollectionAccumulator">CollectionAccumulator</a>[<span class="extype" name="org.apache.spark.SparkContext.collectionAccumulator.T">T</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Create and register a <code>CollectionAccumulator</code>, which starts with empty list and accumulates |
| inputs by adding them into the list.</p> |
| </li><li name="org.apache.spark.SparkContext#defaultMinPartitions" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="defaultMinPartitions:Int"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#defaultMinPartitions:Int" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">defaultMinPartitions</span><span class="result">: <span class="extype" name="scala.Int">Int</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Default min number of partitions for Hadoop RDDs when not given by user |
| Notice that we use math.min so the "defaultMinPartitions" cannot be higher than 2.</p><div class="fullcomment"><div class="comment cmt"><p>Default min number of partitions for Hadoop RDDs when not given by user |
| Notice that we use math.min so the "defaultMinPartitions" cannot be higher than 2. |
| The reasons for this are discussed in https://github.com/mesos/spark/pull/718 |
| </p></div></div> |
| </li><li name="org.apache.spark.SparkContext#defaultParallelism" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="defaultParallelism:Int"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#defaultParallelism:Int" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">defaultParallelism</span><span class="result">: <span class="extype" name="scala.Int">Int</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Default level of parallelism to use when not given by user (e.g.</p><div class="fullcomment"><div class="comment cmt"><p>Default level of parallelism to use when not given by user (e.g. parallelize and makeRDD).</p></div></div> |
| </li><li name="org.apache.spark.SparkContext#deployMode" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="deployMode:String"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#deployMode:String" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">deployMode</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span> |
| </span> |
| |
| |
| </li><li name="org.apache.spark.SparkContext#doubleAccumulator" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="doubleAccumulator(name:String):org.apache.spark.util.DoubleAccumulator"></a><a id="doubleAccumulator(String):DoubleAccumulator"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#doubleAccumulator(name:String):org.apache.spark.util.DoubleAccumulator" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">doubleAccumulator</span><span class="params">(<span name="name">name: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <a href="util/DoubleAccumulator.html" class="extype" name="org.apache.spark.util.DoubleAccumulator">DoubleAccumulator</a></span> |
| </span> |
| |
| <p class="shortcomment cmt">Create and register a double accumulator, which starts with 0 and accumulates inputs by <code>add</code>.</p> |
| </li><li name="org.apache.spark.SparkContext#doubleAccumulator" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="doubleAccumulator:org.apache.spark.util.DoubleAccumulator"></a><a id="doubleAccumulator:DoubleAccumulator"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#doubleAccumulator:org.apache.spark.util.DoubleAccumulator" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">doubleAccumulator</span><span class="result">: <a href="util/DoubleAccumulator.html" class="extype" name="org.apache.spark.util.DoubleAccumulator">DoubleAccumulator</a></span> |
| </span> |
| |
| <p class="shortcomment cmt">Create and register a double accumulator, which starts with 0 and accumulates inputs by <code>add</code>.</p> |
| </li><li name="org.apache.spark.SparkContext#emptyRDD" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="emptyRDD[T](implicitevidence$8:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="emptyRDD[T](ClassTag[T]):RDD[T]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#emptyRDD[T](implicitevidence$8:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">emptyRDD</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.emptyRDD.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.emptyRDD.T">T</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Get an RDD that has no partitions or elements.</p> |
| </li><li name="scala.AnyRef#eq" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="eq(x$1:AnyRef):Boolean"></a><a id="eq(AnyRef):Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#eq(x$1:AnyRef):Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">eq</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.AnyRef">AnyRef</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd></dl></div> |
| </li><li name="scala.AnyRef#equals" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="equals(x$1:Any):Boolean"></a><a id="equals(Any):Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#equals(x$1:Any):Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">equals</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.Any">Any</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#files" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="files:Seq[String]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#files:Seq[String]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">files</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span> |
| </span> |
| |
| |
| </li><li name="scala.AnyRef#finalize" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="finalize():Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#finalize():Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">finalize</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected[<span class="extype" name="java.lang">lang</span>] </dd><dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd> |
| <span class="name">@throws</span><span class="args">(<span> |
| |
| <span class="symbol">classOf[java.lang.Throwable]</span> |
| </span>)</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#getAllPools" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="getAllPools:Seq[org.apache.spark.scheduler.Schedulable]"></a><a id="getAllPools:Seq[Schedulable]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#getAllPools:Seq[org.apache.spark.scheduler.Schedulable]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">getAllPools</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="org.apache.spark.scheduler.Schedulable">Schedulable</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">:: DeveloperApi :: |
| Return pools for fair scheduler |
| </p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi :: |
| Return pools for fair scheduler |
| </p></div><dl class="attributes block"> <dt>Annotations</dt><dd> |
| <span class="name">@DeveloperApi</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#getCheckpointDir" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="getCheckpointDir:Option[String]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#getCheckpointDir:Option[String]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">getCheckpointDir</span><span class="result">: <span class="extype" name="scala.Option">Option</span>[<span class="extype" name="scala.Predef.String">String</span>]</span> |
| </span> |
| |
| |
| </li><li name="scala.AnyRef#getClass" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="getClass():Class[_]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#getClass():Class[_]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">getClass</span><span class="params">()</span><span class="result">: <span class="extype" name="java.lang.Class">Class</span>[_]</span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd><dt>Annotations</dt><dd> |
| <span class="name">@native</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#getConf" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="getConf:org.apache.spark.SparkConf"></a><a id="getConf:SparkConf"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#getConf:org.apache.spark.SparkConf" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">getConf</span><span class="result">: <a href="SparkConf.html" class="extype" name="org.apache.spark.SparkConf">SparkConf</a></span> |
| </span> |
| |
| <p class="shortcomment cmt">Return a copy of this SparkContext's configuration.</p><div class="fullcomment"><div class="comment cmt"><p>Return a copy of this SparkContext's configuration. The configuration <i>cannot</i> be |
| changed at runtime. |
| </p></div></div> |
| </li><li name="org.apache.spark.SparkContext#getExecutorMemoryStatus" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="getExecutorMemoryStatus:scala.collection.Map[String,(Long,Long)]"></a><a id="getExecutorMemoryStatus:Map[String,(Long,Long)]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#getExecutorMemoryStatus:scala.collection.Map[String,(Long,Long)]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">getExecutorMemoryStatus</span><span class="result">: <span class="extype" name="scala.collection.Map">Map</span>[<span class="extype" name="scala.Predef.String">String</span>, (<span class="extype" name="scala.Long">Long</span>, <span class="extype" name="scala.Long">Long</span>)]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Return a map from the block manager to the max memory available for caching and the remaining |
| memory available for caching.</p> |
| </li><li name="org.apache.spark.SparkContext#getLocalProperty" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="getLocalProperty(key:String):String"></a><a id="getLocalProperty(String):String"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#getLocalProperty(key:String):String" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">getLocalProperty</span><span class="params">(<span name="key">key: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Get a local property set in this thread, or null if it is missing.</p><div class="fullcomment"><div class="comment cmt"><p>Get a local property set in this thread, or null if it is missing. See |
| <code>org.apache.spark.SparkContext.setLocalProperty</code>. |
| </p></div></div> |
| </li><li name="org.apache.spark.SparkContext#getPersistentRDDs" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="getPersistentRDDs:scala.collection.Map[Int,org.apache.spark.rdd.RDD[_]]"></a><a id="getPersistentRDDs:Map[Int,RDD[_]]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#getPersistentRDDs:scala.collection.Map[Int,org.apache.spark.rdd.RDD[_]]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">getPersistentRDDs</span><span class="result">: <span class="extype" name="scala.collection.Map">Map</span>[<span class="extype" name="scala.Int">Int</span>, <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[_]]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Returns an immutable map of RDDs that have marked themselves as persistent via cache() call.</p><div class="fullcomment"><div class="comment cmt"><p>Returns an immutable map of RDDs that have marked themselves as persistent via cache() call. |
| </p></div><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>This does not necessarily mean the caching or computation was successful.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#getPoolForName" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="getPoolForName(pool:String):Option[org.apache.spark.scheduler.Schedulable]"></a><a id="getPoolForName(String):Option[Schedulable]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#getPoolForName(pool:String):Option[org.apache.spark.scheduler.Schedulable]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">getPoolForName</span><span class="params">(<span name="pool">pool: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Option">Option</span>[<span class="extype" name="org.apache.spark.scheduler.Schedulable">Schedulable</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">:: DeveloperApi :: |
| Return the pool associated with the given name, if one exists |
| </p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi :: |
| Return the pool associated with the given name, if one exists |
| </p></div><dl class="attributes block"> <dt>Annotations</dt><dd> |
| <span class="name">@DeveloperApi</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#getRDDStorageInfo" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="getRDDStorageInfo:Array[org.apache.spark.storage.RDDInfo]"></a><a id="getRDDStorageInfo:Array[RDDInfo]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#getRDDStorageInfo:Array[org.apache.spark.storage.RDDInfo]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">getRDDStorageInfo</span><span class="result">: <span class="extype" name="scala.Array">Array</span>[<a href="storage/RDDInfo.html" class="extype" name="org.apache.spark.storage.RDDInfo">RDDInfo</a>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">:: DeveloperApi :: |
| Return information about what RDDs are cached, if they are in mem or on disk, how much space |
| they take, etc.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi :: |
| Return information about what RDDs are cached, if they are in mem or on disk, how much space |
| they take, etc. |
| </p></div><dl class="attributes block"> <dt>Annotations</dt><dd> |
| <span class="name">@DeveloperApi</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#getSchedulingMode" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="getSchedulingMode:org.apache.spark.scheduler.SchedulingMode.SchedulingMode"></a><a id="getSchedulingMode:SchedulingMode"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#getSchedulingMode:org.apache.spark.scheduler.SchedulingMode.SchedulingMode" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">getSchedulingMode</span><span class="result">: <a href="scheduler/SchedulingMode$.html#SchedulingMode=org.apache.spark.scheduler.SchedulingMode.Value" class="extmbr" name="org.apache.spark.scheduler.SchedulingMode.SchedulingMode">SchedulingMode</a></span> |
| </span> |
| |
| <p class="shortcomment cmt">Return current scheduling mode |
| </p> |
| </li><li name="org.apache.spark.SparkContext#hadoopConfiguration" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="hadoopConfiguration:org.apache.hadoop.conf.Configuration"></a><a id="hadoopConfiguration:Configuration"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#hadoopConfiguration:org.apache.hadoop.conf.Configuration" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">hadoopConfiguration</span><span class="result">: <span class="extype" name="org.apache.hadoop.conf.Configuration">Configuration</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">A default Hadoop Configuration for the Hadoop code (e.g.</p><div class="fullcomment"><div class="comment cmt"><p>A default Hadoop Configuration for the Hadoop code (e.g. file systems) that we reuse. |
| </p></div><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>As it will be reused in all Hadoop RDDs, it's better not to modify it unless you |
| plan to set some global configurations for all Hadoop RDDs.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#hadoopFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="hadoopFile[K,V,F<:org.apache.hadoop.mapred.InputFormat[K,V]](path:String)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="hadoopFile[K,V,F<:InputFormat[K,V]](String)(ClassTag[K],ClassTag[V],ClassTag[F]):RDD[(K,V)]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#hadoopFile[K,V,F<:org.apache.hadoop.mapred.InputFormat[K,V]](path:String)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">hadoopFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>, <span name="F">F <: <span class="extype" name="org.apache.hadoop.mapred.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>]</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="km">km: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>]</span>, <span name="vm">vm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>]</span>, <span name="fm">fm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.F">F</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>)]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Smarter version of hadoopFile() that uses class tags to figure out the classes of keys, |
| values and the InputFormat so that users don't need to pass them directly.</p><div class="fullcomment"><div class="comment cmt"><p>Smarter version of hadoopFile() that uses class tags to figure out the classes of keys, |
| values and the InputFormat so that users don't need to pass them directly. Instead, callers |
| can just write, for example,</p><pre><span class="kw">val</span> file = sparkContext.hadoopFile[LongWritable, Text, TextInputFormat](path)</pre></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths as |
| a list of inputs</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each |
| record, directly caching the returned RDD or directly passing it to an aggregation or shuffle |
| operation will create many references to the same object. |
| If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first |
| copy them using a <code>map</code> function.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#hadoopFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="hadoopFile[K,V,F<:org.apache.hadoop.mapred.InputFormat[K,V]](path:String,minPartitions:Int)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="hadoopFile[K,V,F<:InputFormat[K,V]](String,Int)(ClassTag[K],ClassTag[V],ClassTag[F]):RDD[(K,V)]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#hadoopFile[K,V,F<:org.apache.hadoop.mapred.InputFormat[K,V]](path:String,minPartitions:Int)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">hadoopFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>, <span name="F">F <: <span class="extype" name="org.apache.hadoop.mapred.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>]</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="km">km: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>]</span>, <span name="vm">vm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>]</span>, <span name="fm">fm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.F">F</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>)]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Smarter version of hadoopFile() that uses class tags to figure out the classes of keys, |
| values and the InputFormat so that users don't need to pass them directly.</p><div class="fullcomment"><div class="comment cmt"><p>Smarter version of hadoopFile() that uses class tags to figure out the classes of keys, |
| values and the InputFormat so that users don't need to pass them directly. Instead, callers |
| can just write, for example,</p><pre><span class="kw">val</span> file = sparkContext.hadoopFile[LongWritable, Text, TextInputFormat](path, minPartitions)</pre></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths |
| as a list of inputs</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>suggested minimum number of partitions for the resulting RDD</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each |
| record, directly caching the returned RDD or directly passing it to an aggregation or shuffle |
| operation will create many references to the same object. |
| If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first |
| copy them using a <code>map</code> function.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#hadoopFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="hadoopFile[K,V](path:String,inputFormatClass:Class[_<:org.apache.hadoop.mapred.InputFormat[K,V]],keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="hadoopFile[K,V](String,Class[_<:InputFormat[K,V]],Class[K],Class[V],Int):RDD[(K,V)]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#hadoopFile[K,V](path:String,inputFormatClass:Class[_<:org.apache.hadoop.mapred.InputFormat[K,V]],keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">hadoopFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="inputFormatClass">inputFormatClass: <span class="extype" name="scala.Predef.Class">Class</span>[_ <: <span class="extype" name="org.apache.hadoop.mapred.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>]]</span>, <span name="keyClass">keyClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>]</span>, <span name="valueClass">valueClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>]</span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.hadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopFile.V">V</span>)]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Get an RDD for a Hadoop file with an arbitrary InputFormat |
| </p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a Hadoop file with an arbitrary InputFormat |
| </p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths |
| as a list of inputs</p></dd><dt class="param">inputFormatClass</dt><dd class="cmt"><p>storage format of the data to be read</p></dd><dt class="param">keyClass</dt><dd class="cmt"><p><code>Class</code> of the key associated with the <code>inputFormatClass</code> parameter</p></dd><dt class="param">valueClass</dt><dd class="cmt"><p><code>Class</code> of the value associated with the <code>inputFormatClass</code> parameter</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>suggested minimum number of partitions for the resulting RDD</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each |
| record, directly caching the returned RDD or directly passing it to an aggregation or shuffle |
| operation will create many references to the same object. |
| If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first |
| copy them using a <code>map</code> function.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#hadoopRDD" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="hadoopRDD[K,V](conf:org.apache.hadoop.mapred.JobConf,inputFormatClass:Class[_<:org.apache.hadoop.mapred.InputFormat[K,V]],keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="hadoopRDD[K,V](JobConf,Class[_<:InputFormat[K,V]],Class[K],Class[V],Int):RDD[(K,V)]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#hadoopRDD[K,V](conf:org.apache.hadoop.mapred.JobConf,inputFormatClass:Class[_<:org.apache.hadoop.mapred.InputFormat[K,V]],keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">hadoopRDD</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>]</span><span class="params">(<span name="conf">conf: <span class="extype" name="org.apache.hadoop.mapred.JobConf">JobConf</span></span>, <span name="inputFormatClass">inputFormatClass: <span class="extype" name="scala.Predef.Class">Class</span>[_ <: <span class="extype" name="org.apache.hadoop.mapred.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopRDD.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopRDD.V">V</span>]]</span>, <span name="keyClass">keyClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopRDD.K">K</span>]</span>, <span name="valueClass">valueClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.hadoopRDD.V">V</span>]</span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.hadoopRDD.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.hadoopRDD.V">V</span>)]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Get an RDD for a Hadoop-readable dataset from a Hadoop JobConf given its InputFormat and other |
| necessary info (e.g.</p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a Hadoop-readable dataset from a Hadoop JobConf given its InputFormat and other |
| necessary info (e.g. file name for a filesystem-based dataset, table name for HyperTable), |
| using the older MapReduce API (<code>org.apache.hadoop.mapred</code>). |
| </p></div><dl class="paramcmts block"><dt class="param">conf</dt><dd class="cmt"><p>JobConf for setting up the dataset. Note: This will be put into a Broadcast. |
| Therefore if you plan to reuse this conf to create multiple RDDs, you need to make |
| sure you won't modify the conf. A safe approach is always creating a new conf for |
| a new RDD.</p></dd><dt class="param">inputFormatClass</dt><dd class="cmt"><p>storage format of the data to be read</p></dd><dt class="param">keyClass</dt><dd class="cmt"><p><code>Class</code> of the key associated with the <code>inputFormatClass</code> parameter</p></dd><dt class="param">valueClass</dt><dd class="cmt"><p><code>Class</code> of the value associated with the <code>inputFormatClass</code> parameter</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>Minimum number of Hadoop Splits to generate.</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each |
| record, directly caching the returned RDD or directly passing it to an aggregation or shuffle |
| operation will create many references to the same object. |
| If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first |
| copy them using a <code>map</code> function.</p></span></dd></dl></div> |
| </li><li name="scala.AnyRef#hashCode" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="hashCode():Int"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#hashCode():Int" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">hashCode</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Int">Int</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd><dt>Annotations</dt><dd> |
| <span class="name">@native</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#initializeLogIfNecessary" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="initializeLogIfNecessary(isInterpreter:Boolean,silent:Boolean):Boolean"></a><a id="initializeLogIfNecessary(Boolean,Boolean):Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#initializeLogIfNecessary(isInterpreter:Boolean,silent:Boolean):Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">initializeLogIfNecessary</span><span class="params">(<span name="isInterpreter">isInterpreter: <span class="extype" name="scala.Boolean">Boolean</span></span>, <span name="silent">silent: <span class="extype" name="scala.Boolean">Boolean</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#initializeLogIfNecessary" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="initializeLogIfNecessary(isInterpreter:Boolean):Unit"></a><a id="initializeLogIfNecessary(Boolean):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#initializeLogIfNecessary(isInterpreter:Boolean):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">initializeLogIfNecessary</span><span class="params">(<span name="isInterpreter">isInterpreter: <span class="extype" name="scala.Boolean">Boolean</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="scala.Any#isInstanceOf" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="isInstanceOf[T0]:Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#isInstanceOf[T0]:Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">isInstanceOf</span><span class="tparams">[<span name="T0">T0</span>]</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>Any</dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#isLocal" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="isLocal:Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#isLocal:Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">isLocal</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| |
| </li><li name="org.apache.spark.SparkContext#isStopped" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="isStopped:Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#isStopped:Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">isStopped</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <p class="shortcomment cmt"></p><div class="fullcomment"><div class="comment cmt"></div><dl class="paramcmts block"><dt>returns</dt><dd class="cmt"><p>true if context is stopped or in the midst of stopping.</p></dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#isTraceEnabled" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="isTraceEnabled():Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#isTraceEnabled():Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">isTraceEnabled</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#jars" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="jars:Seq[String]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#jars:Seq[String]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">jars</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span> |
| </span> |
| |
| |
| </li><li name="org.apache.spark.SparkContext#killExecutor" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="killExecutor(executorId:String):Boolean"></a><a id="killExecutor(String):Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#killExecutor(executorId:String):Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">killExecutor</span><span class="params">(<span name="executorId">executorId: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">:: DeveloperApi :: |
| Request that the cluster manager kill the specified executor.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi :: |
| Request that the cluster manager kill the specified executor. |
| </p></div><dl class="paramcmts block"><dt>returns</dt><dd class="cmt"><p>whether the request is received.</p></dd></dl><dl class="attributes block"> <dt>Annotations</dt><dd> |
| <span class="name">@DeveloperApi</span><span class="args">()</span> |
| |
| </dd><dt>Note</dt><dd><span class="cmt"><p>This is an indication to the cluster manager that the application wishes to adjust |
| its resource usage downwards. If the application wishes to replace the executor it kills |
| through this method with a new one, it should follow up explicitly with a call to |
| {{SparkContext#requestExecutors}}.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#killExecutors" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="killExecutors(executorIds:Seq[String]):Boolean"></a><a id="killExecutors(Seq[String]):Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#killExecutors(executorIds:Seq[String]):Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">killExecutors</span><span class="params">(<span name="executorIds">executorIds: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">:: DeveloperApi :: |
| Request that the cluster manager kill the specified executors.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi :: |
| Request that the cluster manager kill the specified executors.</p><p>This is not supported when dynamic allocation is turned on. |
| </p></div><dl class="paramcmts block"><dt>returns</dt><dd class="cmt"><p>whether the request is received.</p></dd></dl><dl class="attributes block"> <dt>Annotations</dt><dd> |
| <span class="name">@DeveloperApi</span><span class="args">()</span> |
| |
| </dd><dt>Note</dt><dd><span class="cmt"><p>This is an indication to the cluster manager that the application wishes to adjust |
| its resource usage downwards. If the application wishes to replace the executors it kills |
| through this method with new ones, it should follow up explicitly with a call to |
| {{SparkContext#requestExecutors}}.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#killTaskAttempt" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="killTaskAttempt(taskId:Long,interruptThread:Boolean,reason:String):Boolean"></a><a id="killTaskAttempt(Long,Boolean,String):Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#killTaskAttempt(taskId:Long,interruptThread:Boolean,reason:String):Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">killTaskAttempt</span><span class="params">(<span name="taskId">taskId: <span class="extype" name="scala.Long">Long</span></span>, <span name="interruptThread">interruptThread: <span class="extype" name="scala.Boolean">Boolean</span> = <span class="symbol">true</span></span>, <span name="reason">reason: <span class="extype" name="scala.Predef.String">String</span> = <span class="defval" name=""killed via SparkContext.killTaskAttempt"">...</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Kill and reschedule the given task attempt.</p><div class="fullcomment"><div class="comment cmt"><p>Kill and reschedule the given task attempt. Task ids can be obtained from the Spark UI |
| or through SparkListener.onTaskStart. |
| </p></div><dl class="paramcmts block"><dt class="param">taskId</dt><dd class="cmt"><p>the task ID to kill. This id uniquely identifies the task attempt.</p></dd><dt class="param">interruptThread</dt><dd class="cmt"><p>whether to interrupt the thread running the task.</p></dd><dt class="param">reason</dt><dd class="cmt"><p>the reason for killing the task, which should be a short string. If a task |
| is killed multiple times with different reasons, only one reason will be reported.</p></dd><dt>returns</dt><dd class="cmt"><p>Whether the task was successfully killed.</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#listArchives" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="listArchives():Seq[String]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#listArchives():Seq[String]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">listArchives</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">:: Experimental :: |
| Returns a list of archive paths that are added to resources.</p><div class="fullcomment"><div class="comment cmt"><p>:: Experimental :: |
| Returns a list of archive paths that are added to resources. |
| </p></div><dl class="attributes block"> <dt>Annotations</dt><dd> |
| <span class="name">@Experimental</span><span class="args">()</span> |
| |
| </dd><dt>Since</dt><dd><p>3.1.0</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#listFiles" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="listFiles():Seq[String]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#listFiles():Seq[String]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">listFiles</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Returns a list of file paths that are added to resources.</p> |
| </li><li name="org.apache.spark.SparkContext#listJars" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="listJars():Seq[String]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#listJars():Seq[String]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">listJars</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Returns a list of jar files that are added to resources.</p> |
| </li><li name="org.apache.spark.SparkContext#localProperties" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="localProperties:InheritableThreadLocal[java.util.Properties]"></a><a id="localProperties:InheritableThreadLocal[Properties]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#localProperties:InheritableThreadLocal[java.util.Properties]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">val</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">localProperties</span><span class="result">: <span class="extype" name="java.lang.InheritableThreadLocal">InheritableThreadLocal</span>[<span class="extype" name="java.util.Properties">Properties</span>]</span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected[<a href="index.html" class="extype" name="org.apache.spark">spark</a>] </dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#log" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="log:org.slf4j.Logger"></a><a id="log:Logger"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#log:org.slf4j.Logger" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">log</span><span class="result">: <span class="extype" name="org.slf4j.Logger">Logger</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#logDebug" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="logDebug(msg:=>String,throwable:Throwable):Unit"></a><a id="logDebug(⇒String,Throwable):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#logDebug(msg:=>String,throwable:Throwable):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">logDebug</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>, <span name="throwable">throwable: <span class="extype" name="scala.Throwable">Throwable</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#logDebug" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="logDebug(msg:=>String):Unit"></a><a id="logDebug(⇒String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#logDebug(msg:=>String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">logDebug</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#logError" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="logError(msg:=>String,throwable:Throwable):Unit"></a><a id="logError(⇒String,Throwable):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#logError(msg:=>String,throwable:Throwable):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">logError</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>, <span name="throwable">throwable: <span class="extype" name="scala.Throwable">Throwable</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#logError" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="logError(msg:=>String):Unit"></a><a id="logError(⇒String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#logError(msg:=>String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">logError</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#logInfo" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="logInfo(msg:=>String,throwable:Throwable):Unit"></a><a id="logInfo(⇒String,Throwable):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#logInfo(msg:=>String,throwable:Throwable):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">logInfo</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>, <span name="throwable">throwable: <span class="extype" name="scala.Throwable">Throwable</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#logInfo" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="logInfo(msg:=>String):Unit"></a><a id="logInfo(⇒String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#logInfo(msg:=>String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">logInfo</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#logName" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="logName:String"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#logName:String" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">logName</span><span class="result">: <span class="extype" name="java.lang.String">String</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#logTrace" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="logTrace(msg:=>String,throwable:Throwable):Unit"></a><a id="logTrace(⇒String,Throwable):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#logTrace(msg:=>String,throwable:Throwable):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">logTrace</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>, <span name="throwable">throwable: <span class="extype" name="scala.Throwable">Throwable</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#logTrace" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="logTrace(msg:=>String):Unit"></a><a id="logTrace(⇒String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#logTrace(msg:=>String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">logTrace</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#logWarning" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="logWarning(msg:=>String,throwable:Throwable):Unit"></a><a id="logWarning(⇒String,Throwable):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#logWarning(msg:=>String,throwable:Throwable):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">logWarning</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>, <span name="throwable">throwable: <span class="extype" name="scala.Throwable">Throwable</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.internal.Logging#logWarning" visbl="prt" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="logWarning(msg:=>String):Unit"></a><a id="logWarning(⇒String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#logWarning(msg:=>String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">logWarning</span><span class="params">(<span name="msg">msg: ⇒ <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Attributes</dt><dd>protected </dd><dt>Definition Classes</dt><dd>Logging</dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#longAccumulator" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="longAccumulator(name:String):org.apache.spark.util.LongAccumulator"></a><a id="longAccumulator(String):LongAccumulator"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#longAccumulator(name:String):org.apache.spark.util.LongAccumulator" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">longAccumulator</span><span class="params">(<span name="name">name: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <a href="util/LongAccumulator.html" class="extype" name="org.apache.spark.util.LongAccumulator">LongAccumulator</a></span> |
| </span> |
| |
| <p class="shortcomment cmt">Create and register a long accumulator, which starts with 0 and accumulates inputs by <code>add</code>.</p> |
| </li><li name="org.apache.spark.SparkContext#longAccumulator" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="longAccumulator:org.apache.spark.util.LongAccumulator"></a><a id="longAccumulator:LongAccumulator"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#longAccumulator:org.apache.spark.util.LongAccumulator" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">longAccumulator</span><span class="result">: <a href="util/LongAccumulator.html" class="extype" name="org.apache.spark.util.LongAccumulator">LongAccumulator</a></span> |
| </span> |
| |
| <p class="shortcomment cmt">Create and register a long accumulator, which starts with 0 and accumulates inputs by <code>add</code>.</p> |
| </li><li name="org.apache.spark.SparkContext#makeRDD" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="makeRDD[T](seq:Seq[(T,Seq[String])])(implicitevidence$3:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="makeRDD[T](Seq[(T,Seq[String])])(ClassTag[T]):RDD[T]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#makeRDD[T](seq:Seq[(T,Seq[String])])(implicitevidence$3:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">makeRDD</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="seq">seq: <span class="extype" name="scala.Seq">Seq</span>[(<span class="extype" name="org.apache.spark.SparkContext.makeRDD.T">T</span>, <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Predef.String">String</span>])]</span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.makeRDD.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.makeRDD.T">T</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Distribute a local Scala collection to form an RDD, with one or more |
| location preferences (hostnames of Spark nodes) for each object.</p><div class="fullcomment"><div class="comment cmt"><p>Distribute a local Scala collection to form an RDD, with one or more |
| location preferences (hostnames of Spark nodes) for each object. |
| Create a new partition for each collection item.</p></div><dl class="paramcmts block"><dt class="param">seq</dt><dd class="cmt"><p>list of tuples of data and location preferences (hostnames of Spark nodes)</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing data partitioned according to location preferences</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#makeRDD" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="makeRDD[T](seq:Seq[T],numSlices:Int)(implicitevidence$2:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="makeRDD[T](Seq[T],Int)(ClassTag[T]):RDD[T]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#makeRDD[T](seq:Seq[T],numSlices:Int)(implicitevidence$2:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">makeRDD</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="seq">seq: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="org.apache.spark.SparkContext.makeRDD.T">T</span>]</span>, <span name="numSlices">numSlices: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultParallelism:Int">defaultParallelism</a></span></span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.makeRDD.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.makeRDD.T">T</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Distribute a local Scala collection to form an RDD.</p><div class="fullcomment"><div class="comment cmt"><p>Distribute a local Scala collection to form an RDD.</p><p>This method is identical to <code>parallelize</code>.</p></div><dl class="paramcmts block"><dt class="param">seq</dt><dd class="cmt"><p>Scala collection to distribute</p></dd><dt class="param">numSlices</dt><dd class="cmt"><p>number of partitions to divide the collection into</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing distributed collection</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#master" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="master:String"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#master:String" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">master</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span> |
| </span> |
| |
| |
| </li><li name="scala.AnyRef#ne" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="ne(x$1:AnyRef):Boolean"></a><a id="ne(AnyRef):Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#ne(x$1:AnyRef):Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">ne</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.AnyRef">AnyRef</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#newAPIHadoopFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="newAPIHadoopFile[K,V,F<:org.apache.hadoop.mapreduce.InputFormat[K,V]](path:String,fClass:Class[F],kClass:Class[K],vClass:Class[V],conf:org.apache.hadoop.conf.Configuration):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="newAPIHadoopFile[K,V,F<:InputFormat[K,V]](String,Class[F],Class[K],Class[V],Configuration):RDD[(K,V)]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#newAPIHadoopFile[K,V,F<:org.apache.hadoop.mapreduce.InputFormat[K,V]](path:String,fClass:Class[F],kClass:Class[K],vClass:Class[V],conf:org.apache.hadoop.conf.Configuration):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">newAPIHadoopFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>, <span name="F">F <: <span class="extype" name="org.apache.hadoop.mapreduce.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.V">V</span>]</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="fClass">fClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.F">F</span>]</span>, <span name="kClass">kClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.K">K</span>]</span>, <span name="vClass">vClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.V">V</span>]</span>, <span name="conf">conf: <span class="extype" name="org.apache.hadoop.conf.Configuration">Configuration</span> = <span class="symbol"><span class="name"><a href="#hadoopConfiguration:org.apache.hadoop.conf.Configuration">hadoopConfiguration</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.V">V</span>)]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Get an RDD for a given Hadoop file with an arbitrary new API InputFormat |
| and extra configuration options to pass to the input format.</p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a given Hadoop file with an arbitrary new API InputFormat |
| and extra configuration options to pass to the input format. |
| </p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths |
| as a list of inputs</p></dd><dt class="param">fClass</dt><dd class="cmt"><p>storage format of the data to be read</p></dd><dt class="param">kClass</dt><dd class="cmt"><p><code>Class</code> of the key associated with the <code>fClass</code> parameter</p></dd><dt class="param">vClass</dt><dd class="cmt"><p><code>Class</code> of the value associated with the <code>fClass</code> parameter</p></dd><dt class="param">conf</dt><dd class="cmt"><p>Hadoop configuration</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each |
| record, directly caching the returned RDD or directly passing it to an aggregation or shuffle |
| operation will create many references to the same object. |
| If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first |
| copy them using a <code>map</code> function.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#newAPIHadoopFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="newAPIHadoopFile[K,V,F<:org.apache.hadoop.mapreduce.InputFormat[K,V]](path:String)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="newAPIHadoopFile[K,V,F<:InputFormat[K,V]](String)(ClassTag[K],ClassTag[V],ClassTag[F]):RDD[(K,V)]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#newAPIHadoopFile[K,V,F<:org.apache.hadoop.mapreduce.InputFormat[K,V]](path:String)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">newAPIHadoopFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>, <span name="F">F <: <span class="extype" name="org.apache.hadoop.mapreduce.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.V">V</span>]</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="km">km: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.K">K</span>]</span>, <span name="vm">vm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.V">V</span>]</span>, <span name="fm">fm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.F">F</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopFile.V">V</span>)]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Smarter version of <code>newApiHadoopFile</code> that uses class tags to figure out the classes of keys, |
| values and the <code>org.apache.hadoop.mapreduce.InputFormat</code> (new MapReduce API) so that user |
| don't need to pass them directly.</p><div class="fullcomment"><div class="comment cmt"><p>Smarter version of <code>newApiHadoopFile</code> that uses class tags to figure out the classes of keys, |
| values and the <code>org.apache.hadoop.mapreduce.InputFormat</code> (new MapReduce API) so that user |
| don't need to pass them directly. Instead, callers can just write, for example: |
| <code><code><code> |
| val file = sparkContext.hadoopFile[LongWritable, Text, TextInputFormat](path) |
| <code><code><code> |
| </code></code></code></code></code></code></p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths |
| as a list of inputs</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each |
| record, directly caching the returned RDD or directly passing it to an aggregation or shuffle |
| operation will create many references to the same object. |
| If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first |
| copy them using a <code>map</code> function.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#newAPIHadoopRDD" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="newAPIHadoopRDD[K,V,F<:org.apache.hadoop.mapreduce.InputFormat[K,V]](conf:org.apache.hadoop.conf.Configuration,fClass:Class[F],kClass:Class[K],vClass:Class[V]):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="newAPIHadoopRDD[K,V,F<:InputFormat[K,V]](Configuration,Class[F],Class[K],Class[V]):RDD[(K,V)]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#newAPIHadoopRDD[K,V,F<:org.apache.hadoop.mapreduce.InputFormat[K,V]](conf:org.apache.hadoop.conf.Configuration,fClass:Class[F],kClass:Class[K],vClass:Class[V]):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">newAPIHadoopRDD</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>, <span name="F">F <: <span class="extype" name="org.apache.hadoop.mapreduce.InputFormat">InputFormat</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.V">V</span>]</span>]</span><span class="params">(<span name="conf">conf: <span class="extype" name="org.apache.hadoop.conf.Configuration">Configuration</span> = <span class="symbol"><span class="name"><a href="#hadoopConfiguration:org.apache.hadoop.conf.Configuration">hadoopConfiguration</a></span></span></span>, <span name="fClass">fClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.F">F</span>]</span>, <span name="kClass">kClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.K">K</span>]</span>, <span name="vClass">vClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.V">V</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.newAPIHadoopRDD.V">V</span>)]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Get an RDD for a given Hadoop file with an arbitrary new API InputFormat |
| and extra configuration options to pass to the input format.</p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a given Hadoop file with an arbitrary new API InputFormat |
| and extra configuration options to pass to the input format. |
| </p></div><dl class="paramcmts block"><dt class="param">conf</dt><dd class="cmt"><p>Configuration for setting up the dataset. Note: This will be put into a Broadcast. |
| Therefore if you plan to reuse this conf to create multiple RDDs, you need to make |
| sure you won't modify the conf. A safe approach is always creating a new conf for |
| a new RDD.</p></dd><dt class="param">fClass</dt><dd class="cmt"><p>storage format of the data to be read</p></dd><dt class="param">kClass</dt><dd class="cmt"><p><code>Class</code> of the key associated with the <code>fClass</code> parameter</p></dd><dt class="param">vClass</dt><dd class="cmt"><p><code>Class</code> of the value associated with the <code>fClass</code> parameter</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each |
| record, directly caching the returned RDD or directly passing it to an aggregation or shuffle |
| operation will create many references to the same object. |
| If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first |
| copy them using a <code>map</code> function.</p></span></dd></dl></div> |
| </li><li name="scala.AnyRef#notify" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="notify():Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#notify():Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">notify</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd> |
| <span class="name">@native</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="scala.AnyRef#notifyAll" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="notifyAll():Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#notifyAll():Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">notifyAll</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd> |
| <span class="name">@native</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#objectFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="objectFile[T](path:String,minPartitions:Int)(implicitevidence$4:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="objectFile[T](String,Int)(ClassTag[T]):RDD[T]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#objectFile[T](path:String,minPartitions:Int)(implicitevidence$4:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">objectFile</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.objectFile.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.objectFile.T">T</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Load an RDD saved as a SequenceFile containing serialized objects, with NullWritable keys and |
| BytesWritable values that contain a serialized partition.</p><div class="fullcomment"><div class="comment cmt"><p>Load an RDD saved as a SequenceFile containing serialized objects, with NullWritable keys and |
| BytesWritable values that contain a serialized partition. This is still an experimental |
| storage format and may not be supported exactly as is in future Spark releases. It will also |
| be pretty slow if you use the default serializer (Java serialization), |
| though the nice thing about it is that there's very little effort required to save arbitrary |
| objects. |
| </p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths |
| as a list of inputs</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>suggested minimum number of partitions for the resulting RDD</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing deserialized data from the file(s)</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#parallelize" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="parallelize[T](seq:Seq[T],numSlices:Int)(implicitevidence$1:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="parallelize[T](Seq[T],Int)(ClassTag[T]):RDD[T]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#parallelize[T](seq:Seq[T],numSlices:Int)(implicitevidence$1:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">parallelize</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="seq">seq: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="org.apache.spark.SparkContext.parallelize.T">T</span>]</span>, <span name="numSlices">numSlices: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultParallelism:Int">defaultParallelism</a></span></span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.parallelize.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.parallelize.T">T</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Distribute a local Scala collection to form an RDD.</p><div class="fullcomment"><div class="comment cmt"><p>Distribute a local Scala collection to form an RDD. |
| </p></div><dl class="paramcmts block"><dt class="param">seq</dt><dd class="cmt"><p>Scala collection to distribute</p></dd><dt class="param">numSlices</dt><dd class="cmt"><p>number of partitions to divide the collection into</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing distributed collection</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Parallelize acts lazily. If <code>seq</code> is a mutable collection and is altered after the call |
| to parallelize and before the first action on the RDD, the resultant RDD will reflect the |
| modified collection. Pass a copy of the argument to avoid this.</p></span>, <span class="cmt"><p>avoid using <code>parallelize(Seq())</code> to create an empty <code>RDD</code>. Consider <code>emptyRDD</code> for an |
| RDD with no partitions, or <code>parallelize(Seq[T]())</code> for an RDD of <code>T</code> with empty partitions.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#range" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="range(start:Long,end:Long,step:Long,numSlices:Int):org.apache.spark.rdd.RDD[Long]"></a><a id="range(Long,Long,Long,Int):RDD[Long]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#range(start:Long,end:Long,step:Long,numSlices:Int):org.apache.spark.rdd.RDD[Long]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">range</span><span class="params">(<span name="start">start: <span class="extype" name="scala.Long">Long</span></span>, <span name="end">end: <span class="extype" name="scala.Long">Long</span></span>, <span name="step">step: <span class="extype" name="scala.Long">Long</span> = <span class="symbol">1</span></span>, <span name="numSlices">numSlices: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultParallelism:Int">defaultParallelism</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="scala.Long">Long</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Creates a new RDD[Long] containing elements from <code>start</code> to <code>end</code>(exclusive), increased by |
| <code>step</code> every element.</p><div class="fullcomment"><div class="comment cmt"><p>Creates a new RDD[Long] containing elements from <code>start</code> to <code>end</code>(exclusive), increased by |
| <code>step</code> every element. |
| </p></div><dl class="paramcmts block"><dt class="param">start</dt><dd class="cmt"><p>the start value.</p></dd><dt class="param">end</dt><dd class="cmt"><p>the end value.</p></dd><dt class="param">step</dt><dd class="cmt"><p>the incremental step</p></dd><dt class="param">numSlices</dt><dd class="cmt"><p>number of partitions to divide the collection into</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing distributed range</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>if we need to cache this RDD, we should make sure each partition does not exceed limit.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#register" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="register(acc:org.apache.spark.util.AccumulatorV2[_,_],name:String):Unit"></a><a id="register(AccumulatorV2[_,_],String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#register(acc:org.apache.spark.util.AccumulatorV2[_,_],name:String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">register</span><span class="params">(<span name="acc">acc: <a href="util/AccumulatorV2.html" class="extype" name="org.apache.spark.util.AccumulatorV2">AccumulatorV2</a>[_, _]</span>, <span name="name">name: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Register the given accumulator with given name.</p><div class="fullcomment"><div class="comment cmt"><p>Register the given accumulator with given name. |
| </p></div><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Accumulators must be registered before use, or it will throw exception.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#register" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="register(acc:org.apache.spark.util.AccumulatorV2[_,_]):Unit"></a><a id="register(AccumulatorV2[_,_]):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#register(acc:org.apache.spark.util.AccumulatorV2[_,_]):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">register</span><span class="params">(<span name="acc">acc: <a href="util/AccumulatorV2.html" class="extype" name="org.apache.spark.util.AccumulatorV2">AccumulatorV2</a>[_, _]</span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Register the given accumulator.</p><div class="fullcomment"><div class="comment cmt"><p>Register the given accumulator. |
| </p></div><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Accumulators must be registered before use, or it will throw exception.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#removeSparkListener" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="removeSparkListener(listener:org.apache.spark.scheduler.SparkListenerInterface):Unit"></a><a id="removeSparkListener(SparkListenerInterface):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#removeSparkListener(listener:org.apache.spark.scheduler.SparkListenerInterface):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">removeSparkListener</span><span class="params">(<span name="listener">listener: <span class="extype" name="org.apache.spark.scheduler.SparkListenerInterface">SparkListenerInterface</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">:: DeveloperApi :: |
| Deregister the listener from Spark's listener bus.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi :: |
| Deregister the listener from Spark's listener bus. |
| </p></div><dl class="attributes block"> <dt>Annotations</dt><dd> |
| <span class="name">@DeveloperApi</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#requestExecutors" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="requestExecutors(numAdditionalExecutors:Int):Boolean"></a><a id="requestExecutors(Int):Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#requestExecutors(numAdditionalExecutors:Int):Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">requestExecutors</span><span class="params">(<span name="numAdditionalExecutors">numAdditionalExecutors: <span class="extype" name="scala.Int">Int</span></span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">:: DeveloperApi :: |
| Request an additional number of executors from the cluster manager.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi :: |
| Request an additional number of executors from the cluster manager.</p></div><dl class="paramcmts block"><dt>returns</dt><dd class="cmt"><p>whether the request is received.</p></dd></dl><dl class="attributes block"> <dt>Annotations</dt><dd> |
| <span class="name">@DeveloperApi</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#requestTotalExecutors" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="requestTotalExecutors(numExecutors:Int,localityAwareTasks:Int,hostToLocalTaskCount:scala.collection.immutable.Map[String,Int]):Boolean"></a><a id="requestTotalExecutors(Int,Int,Map[String,Int]):Boolean"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#requestTotalExecutors(numExecutors:Int,localityAwareTasks:Int,hostToLocalTaskCount:scala.collection.immutable.Map[String,Int]):Boolean" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">requestTotalExecutors</span><span class="params">(<span name="numExecutors">numExecutors: <span class="extype" name="scala.Int">Int</span></span>, <span name="localityAwareTasks">localityAwareTasks: <span class="extype" name="scala.Int">Int</span></span>, <span name="hostToLocalTaskCount">hostToLocalTaskCount: <span class="extype" name="scala.collection.immutable.Map">Map</span>[<span class="extype" name="scala.Predef.String">String</span>, <span class="extype" name="scala.Int">Int</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Boolean">Boolean</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Update the cluster manager on our scheduling needs.</p><div class="fullcomment"><div class="comment cmt"><p>Update the cluster manager on our scheduling needs. Three bits of information are included |
| to help it make decisions. This applies to the default ResourceProfile.</p></div><dl class="paramcmts block"><dt class="param">numExecutors</dt><dd class="cmt"><p>The total number of executors we'd like to have. The cluster manager |
| shouldn't kill any running executor to reach this number, but, |
| if all existing executors were to die, this is the number of executors |
| we'd want to be allocated.</p></dd><dt class="param">localityAwareTasks</dt><dd class="cmt"><p>The number of tasks in all active stages that have a locality |
| preferences. This includes running, pending, and completed tasks.</p></dd><dt class="param">hostToLocalTaskCount</dt><dd class="cmt"><p>A map of hosts to the number of tasks from all active stages |
| that would like to like to run on that host. |
| This includes running, pending, and completed tasks.</p></dd><dt>returns</dt><dd class="cmt"><p>whether the request is acknowledged by the cluster manager.</p></dd></dl><dl class="attributes block"> <dt>Annotations</dt><dd> |
| <span class="name">@DeveloperApi</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#resources" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="resources:scala.collection.Map[String,org.apache.spark.resource.ResourceInformation]"></a><a id="resources:Map[String,ResourceInformation]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#resources:scala.collection.Map[String,org.apache.spark.resource.ResourceInformation]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">resources</span><span class="result">: <span class="extype" name="scala.collection.Map">Map</span>[<span class="extype" name="scala.Predef.String">String</span>, <a href="resource/ResourceInformation.html" class="extype" name="org.apache.spark.resource.ResourceInformation">ResourceInformation</a>]</span> |
| </span> |
| |
| |
| </li><li name="org.apache.spark.SparkContext#runApproximateJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="runApproximateJob[T,U,R](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=>U,evaluator:org.apache.spark.partial.ApproximateEvaluator[U,R],timeout:Long):org.apache.spark.partial.PartialResult[R]"></a><a id="runApproximateJob[T,U,R](RDD[T],(TaskContext,Iterator[T])⇒U,ApproximateEvaluator[U,R],Long):PartialResult[R]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#runApproximateJob[T,U,R](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=>U,evaluator:org.apache.spark.partial.ApproximateEvaluator[U,R],timeout:Long):org.apache.spark.partial.PartialResult[R]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">runApproximateJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>, <span name="R">R</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runApproximateJob.T">T</span>]</span>, <span name="func">func: (<a href="TaskContext.html" class="extype" name="org.apache.spark.TaskContext">TaskContext</a>, <span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runApproximateJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runApproximateJob.U">U</span></span>, <span name="evaluator">evaluator: <span class="extype" name="org.apache.spark.partial.ApproximateEvaluator">ApproximateEvaluator</span>[<span class="extype" name="org.apache.spark.SparkContext.runApproximateJob.U">U</span>, <span class="extype" name="org.apache.spark.SparkContext.runApproximateJob.R">R</span>]</span>, <span name="timeout">timeout: <span class="extype" name="scala.Long">Long</span></span>)</span><span class="result">: <a href="partial/PartialResult.html" class="extype" name="org.apache.spark.partial.PartialResult">PartialResult</a>[<span class="extype" name="org.apache.spark.SparkContext.runApproximateJob.R">R</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">:: DeveloperApi :: |
| Run a job that can return approximate results.</p><div class="fullcomment"><div class="comment cmt"><p>:: DeveloperApi :: |
| Run a job that can return approximate results. |
| </p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">func</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">evaluator</dt><dd class="cmt"><p><code>ApproximateEvaluator</code> to receive the partial results</p></dd><dt class="param">timeout</dt><dd class="cmt"><p>maximum time to wait for the job, in milliseconds</p></dd><dt>returns</dt><dd class="cmt"><p>partial result (how partial depends on whether the job was finished before or |
| after timeout)</p></dd></dl><dl class="attributes block"> <dt>Annotations</dt><dd> |
| <span class="name">@DeveloperApi</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],processPartition:Iterator[T]=>U,resultHandler:(Int,U)=>Unit)(implicitevidence$16:scala.reflect.ClassTag[U]):Unit"></a><a id="runJob[T,U](RDD[T],(Iterator[T])⇒U,(Int,U)⇒Unit)(ClassTag[U]):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],processPartition:Iterator[T]=>U,resultHandler:(Int,U)=>Unit)(implicitevidence$16:scala.reflect.ClassTag[U]):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="processPartition">processPartition: (<span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>, <span name="resultHandler">resultHandler: (<span class="extype" name="scala.Int">Int</span>, <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>) ⇒ <span class="extype" name="scala.Unit">Unit</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Run a job on all partitions in an RDD and pass the results to a handler function.</p><div class="fullcomment"><div class="comment cmt"><p>Run a job on all partitions in an RDD and pass the results to a handler function. |
| </p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">processPartition</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">resultHandler</dt><dd class="cmt"><p>callback to pass each result to</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],processPartition:(org.apache.spark.TaskContext,Iterator[T])=>U,resultHandler:(Int,U)=>Unit)(implicitevidence$15:scala.reflect.ClassTag[U]):Unit"></a><a id="runJob[T,U](RDD[T],(TaskContext,Iterator[T])⇒U,(Int,U)⇒Unit)(ClassTag[U]):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],processPartition:(org.apache.spark.TaskContext,Iterator[T])=>U,resultHandler:(Int,U)=>Unit)(implicitevidence$15:scala.reflect.ClassTag[U]):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="processPartition">processPartition: (<a href="TaskContext.html" class="extype" name="org.apache.spark.TaskContext">TaskContext</a>, <span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>, <span name="resultHandler">resultHandler: (<span class="extype" name="scala.Int">Int</span>, <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>) ⇒ <span class="extype" name="scala.Unit">Unit</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Run a job on all partitions in an RDD and pass the results to a handler function.</p><div class="fullcomment"><div class="comment cmt"><p>Run a job on all partitions in an RDD and pass the results to a handler function. The function |
| that is run against each partition additionally takes <code>TaskContext</code> argument. |
| </p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">processPartition</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">resultHandler</dt><dd class="cmt"><p>callback to pass each result to</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:Iterator[T]=>U)(implicitevidence$14:scala.reflect.ClassTag[U]):Array[U]"></a><a id="runJob[T,U](RDD[T],(Iterator[T])⇒U)(ClassTag[U]):Array[U]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:Iterator[T]=>U)(implicitevidence$14:scala.reflect.ClassTag[U]):Array[U]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="func">func: (<span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Array">Array</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Run a job on all partitions in an RDD and return the results in an array.</p><div class="fullcomment"><div class="comment cmt"><p>Run a job on all partitions in an RDD and return the results in an array. |
| </p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">func</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt>returns</dt><dd class="cmt"><p>in-memory collection with a result of the job (each collection element will contain |
| a result from one partition)</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=>U)(implicitevidence$13:scala.reflect.ClassTag[U]):Array[U]"></a><a id="runJob[T,U](RDD[T],(TaskContext,Iterator[T])⇒U)(ClassTag[U]):Array[U]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=>U)(implicitevidence$13:scala.reflect.ClassTag[U]):Array[U]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="func">func: (<a href="TaskContext.html" class="extype" name="org.apache.spark.TaskContext">TaskContext</a>, <span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Array">Array</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Run a job on all partitions in an RDD and return the results in an array.</p><div class="fullcomment"><div class="comment cmt"><p>Run a job on all partitions in an RDD and return the results in an array. The function |
| that is run against each partition additionally takes <code>TaskContext</code> argument. |
| </p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">func</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt>returns</dt><dd class="cmt"><p>in-memory collection with a result of the job (each collection element will contain |
| a result from one partition)</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:Iterator[T]=>U,partitions:Seq[Int])(implicitevidence$12:scala.reflect.ClassTag[U]):Array[U]"></a><a id="runJob[T,U](RDD[T],(Iterator[T])⇒U,Seq[Int])(ClassTag[U]):Array[U]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:Iterator[T]=>U,partitions:Seq[Int])(implicitevidence$12:scala.reflect.ClassTag[U]):Array[U]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="func">func: (<span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>, <span name="partitions">partitions: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Int">Int</span>]</span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Array">Array</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Run a function on a given set of partitions in an RDD and return the results as an array.</p><div class="fullcomment"><div class="comment cmt"><p>Run a function on a given set of partitions in an RDD and return the results as an array. |
| </p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">func</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">partitions</dt><dd class="cmt"><p>set of partitions to run on; some jobs may not want to compute on all |
| partitions of the target RDD, e.g. for operations like <code>first()</code></p></dd><dt>returns</dt><dd class="cmt"><p>in-memory collection with a result of the job (each collection element will contain |
| a result from one partition)</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=>U,partitions:Seq[Int])(implicitevidence$11:scala.reflect.ClassTag[U]):Array[U]"></a><a id="runJob[T,U](RDD[T],(TaskContext,Iterator[T])⇒U,Seq[Int])(ClassTag[U]):Array[U]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=>U,partitions:Seq[Int])(implicitevidence$11:scala.reflect.ClassTag[U]):Array[U]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="func">func: (<a href="TaskContext.html" class="extype" name="org.apache.spark.TaskContext">TaskContext</a>, <span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>, <span name="partitions">partitions: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Int">Int</span>]</span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Array">Array</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Run a function on a given set of partitions in an RDD and return the results as an array.</p><div class="fullcomment"><div class="comment cmt"><p>Run a function on a given set of partitions in an RDD and return the results as an array. |
| The function that is run against each partition additionally takes <code>TaskContext</code> argument. |
| </p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">func</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">partitions</dt><dd class="cmt"><p>set of partitions to run on; some jobs may not want to compute on all |
| partitions of the target RDD, e.g. for operations like <code>first()</code></p></dd><dt>returns</dt><dd class="cmt"><p>in-memory collection with a result of the job (each collection element will contain |
| a result from one partition)</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#runJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=>U,partitions:Seq[Int],resultHandler:(Int,U)=>Unit)(implicitevidence$10:scala.reflect.ClassTag[U]):Unit"></a><a id="runJob[T,U](RDD[T],(TaskContext,Iterator[T])⇒U,Seq[Int],(Int,U)⇒Unit)(ClassTag[U]):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#runJob[T,U](rdd:org.apache.spark.rdd.RDD[T],func:(org.apache.spark.TaskContext,Iterator[T])=>U,partitions:Seq[Int],resultHandler:(Int,U)=>Unit)(implicitevidence$10:scala.reflect.ClassTag[U]):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">runJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]</span>, <span name="func">func: (<a href="TaskContext.html" class="extype" name="org.apache.spark.TaskContext">TaskContext</a>, <span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span></span>, <span name="partitions">partitions: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Int">Int</span>]</span>, <span name="resultHandler">resultHandler: (<span class="extype" name="scala.Int">Int</span>, <span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>) ⇒ <span class="extype" name="scala.Unit">Unit</span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.runJob.U">U</span>]</span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Run a function on a given set of partitions in an RDD and pass the results to the given |
| handler function.</p><div class="fullcomment"><div class="comment cmt"><p>Run a function on a given set of partitions in an RDD and pass the results to the given |
| handler function. This is the main entry point for all actions in Spark. |
| </p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">func</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">partitions</dt><dd class="cmt"><p>set of partitions to run on; some jobs may not want to compute on all |
| partitions of the target RDD, e.g. for operations like <code>first()</code></p></dd><dt class="param">resultHandler</dt><dd class="cmt"><p>callback to pass each result to</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#sequenceFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="sequenceFile[K,V](path:String,minPartitions:Int)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitkcf:()=>org.apache.spark.WritableConverter[K],implicitvcf:()=>org.apache.spark.WritableConverter[V]):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="sequenceFile[K,V](String,Int)(ClassTag[K],ClassTag[V],()⇒WritableConverter[K],()⇒WritableConverter[V]):RDD[(K,V)]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#sequenceFile[K,V](path:String,minPartitions:Int)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitkcf:()=>org.apache.spark.WritableConverter[K],implicitvcf:()=>org.apache.spark.WritableConverter[V]):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">sequenceFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="params">(<span class="implicit">implicit </span><span name="km">km: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>]</span>, <span name="vm">vm: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>]</span>, <span name="kcf">kcf: () ⇒ <span class="extype" name="org.apache.spark.WritableConverter">WritableConverter</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>]</span>, <span name="vcf">vcf: () ⇒ <span class="extype" name="org.apache.spark.WritableConverter">WritableConverter</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>)]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Version of sequenceFile() for types implicitly convertible to Writables through a |
| WritableConverter.</p><div class="fullcomment"><div class="comment cmt"><p>Version of sequenceFile() for types implicitly convertible to Writables through a |
| WritableConverter. For example, to access a SequenceFile where the keys are Text and the |
| values are IntWritable, you could simply write</p><pre>sparkContext.sequenceFile[<span class="std">String</span>, <span class="std">Int</span>](path, ...)</pre><p>WritableConverters are provided in a somewhat strange way (by an implicit function) to support |
| both subclasses of Writable and types for which we define a converter (e.g. Int to |
| IntWritable). The most natural thing would've been to have implicit objects for the |
| converters, but then we couldn't have an object for every subclass of Writable (you can't |
| have a parameterized singleton object). We use functions instead to create a new converter |
| for the appropriate type. In addition, we pass the converter a ClassTag of its type to |
| allow it to figure out the Writable class to use in the subclass case. |
| </p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths |
| as a list of inputs</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>suggested minimum number of partitions for the resulting RDD</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each |
| record, directly caching the returned RDD or directly passing it to an aggregation or shuffle |
| operation will create many references to the same object. |
| If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first |
| copy them using a <code>map</code> function.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#sequenceFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="sequenceFile[K,V](path:String,keyClass:Class[K],valueClass:Class[V]):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="sequenceFile[K,V](String,Class[K],Class[V]):RDD[(K,V)]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#sequenceFile[K,V](path:String,keyClass:Class[K],valueClass:Class[V]):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">sequenceFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="keyClass">keyClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>]</span>, <span name="valueClass">valueClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>)]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Get an RDD for a Hadoop SequenceFile with given key and value types.</p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a Hadoop SequenceFile with given key and value types. |
| </p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths |
| as a list of inputs</p></dd><dt class="param">keyClass</dt><dd class="cmt"><p><code>Class</code> of the key associated with <code>SequenceFileInputFormat</code></p></dd><dt class="param">valueClass</dt><dd class="cmt"><p><code>Class</code> of the value associated with <code>SequenceFileInputFormat</code></p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each |
| record, directly caching the returned RDD or directly passing it to an aggregation or shuffle |
| operation will create many references to the same object. |
| If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first |
| copy them using a <code>map</code> function.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#sequenceFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="sequenceFile[K,V](path:String,keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]"></a><a id="sequenceFile[K,V](String,Class[K],Class[V],Int):RDD[(K,V)]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#sequenceFile[K,V](path:String,keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">sequenceFile</span><span class="tparams">[<span name="K">K</span>, <span name="V">V</span>]</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="keyClass">keyClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>]</span>, <span name="valueClass">valueClass: <span class="extype" name="scala.Predef.Class">Class</span>[<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>]</span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="org.apache.spark.SparkContext.sequenceFile.K">K</span>, <span class="extype" name="org.apache.spark.SparkContext.sequenceFile.V">V</span>)]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Get an RDD for a Hadoop SequenceFile with given key and value types.</p><div class="fullcomment"><div class="comment cmt"><p>Get an RDD for a Hadoop SequenceFile with given key and value types. |
| </p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>directory to the input data files, the path can be comma separated paths |
| as a list of inputs</p></dd><dt class="param">keyClass</dt><dd class="cmt"><p><code>Class</code> of the key associated with <code>SequenceFileInputFormat</code></p></dd><dt class="param">valueClass</dt><dd class="cmt"><p><code>Class</code> of the value associated with <code>SequenceFileInputFormat</code></p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>suggested minimum number of partitions for the resulting RDD</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of tuples of key and corresponding value</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Because Hadoop's RecordReader class re-uses the same Writable object for each |
| record, directly caching the returned RDD or directly passing it to an aggregation or shuffle |
| operation will create many references to the same object. |
| If you plan to directly cache, sort, or aggregate Hadoop writable objects, you should first |
| copy them using a <code>map</code> function.</p></span></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#setCallSite" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="setCallSite(shortCallSite:String):Unit"></a><a id="setCallSite(String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#setCallSite(shortCallSite:String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">setCallSite</span><span class="params">(<span name="shortCallSite">shortCallSite: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Set the thread-local property for overriding the call sites |
| of actions and RDDs.</p> |
| </li><li name="org.apache.spark.SparkContext#setCheckpointDir" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="setCheckpointDir(directory:String):Unit"></a><a id="setCheckpointDir(String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#setCheckpointDir(directory:String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">setCheckpointDir</span><span class="params">(<span name="directory">directory: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Set the directory under which RDDs are going to be checkpointed.</p><div class="fullcomment"><div class="comment cmt"><p>Set the directory under which RDDs are going to be checkpointed.</p></div><dl class="paramcmts block"><dt class="param">directory</dt><dd class="cmt"><p>path to the directory where checkpoint files will be stored |
| (must be HDFS path if running in cluster)</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#setJobDescription" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="setJobDescription(value:String):Unit"></a><a id="setJobDescription(String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#setJobDescription(value:String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">setJobDescription</span><span class="params">(<span name="value">value: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Set a human readable description of the current job.</p> |
| </li><li name="org.apache.spark.SparkContext#setJobGroup" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="setJobGroup(groupId:String,description:String,interruptOnCancel:Boolean):Unit"></a><a id="setJobGroup(String,String,Boolean):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#setJobGroup(groupId:String,description:String,interruptOnCancel:Boolean):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">setJobGroup</span><span class="params">(<span name="groupId">groupId: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="description">description: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="interruptOnCancel">interruptOnCancel: <span class="extype" name="scala.Boolean">Boolean</span> = <span class="symbol">false</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Assigns a group ID to all the jobs started by this thread until the group ID is set to a |
| different value or cleared.</p><div class="fullcomment"><div class="comment cmt"><p>Assigns a group ID to all the jobs started by this thread until the group ID is set to a |
| different value or cleared.</p><p>Often, a unit of execution in an application consists of multiple Spark actions or jobs. |
| Application programmers can use this method to group all those jobs together and give a |
| group description. Once set, the Spark web UI will associate such jobs with this group.</p><p>The application can also use <code>org.apache.spark.SparkContext.cancelJobGroup</code> to cancel all |
| running jobs in this group. For example,</p><pre><span class="cmt">// In the main thread:</span> |
| sc.setJobGroup(<span class="lit">"some_job_to_cancel"</span>, <span class="lit">"some job description"</span>) |
| sc.parallelize(<span class="num">1</span> to <span class="num">10000</span>, <span class="num">2</span>).map { i <span class="kw">=></span> Thread.sleep(<span class="num">10</span>); i }.count() |
| |
| <span class="cmt">// In a separate thread:</span> |
| sc.cancelJobGroup(<span class="lit">"some_job_to_cancel"</span>)</pre></div><dl class="paramcmts block"><dt class="param">interruptOnCancel</dt><dd class="cmt"><p>If true, then job cancellation will result in <code>Thread.interrupt()</code> |
| being called on the job's executor threads. This is useful to help ensure that the tasks |
| are actually stopped in a timely manner, but is off by default due to HDFS-1208, where HDFS |
| may respond to Thread.interrupt() by marking nodes as dead.</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#setLocalProperty" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="setLocalProperty(key:String,value:String):Unit"></a><a id="setLocalProperty(String,String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#setLocalProperty(key:String,value:String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">setLocalProperty</span><span class="params">(<span name="key">key: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="value">value: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Set a local property that affects jobs submitted from this thread, such as the Spark fair |
| scheduler pool.</p><div class="fullcomment"><div class="comment cmt"><p>Set a local property that affects jobs submitted from this thread, such as the Spark fair |
| scheduler pool. User-defined properties may also be set here. These properties are propagated |
| through to worker tasks and can be accessed there via |
| <a href="TaskContext.html#getLocalProperty(key:String):String" class="extmbr" name="org.apache.spark.TaskContext#getLocalProperty">org.apache.spark.TaskContext#getLocalProperty</a>.</p><p>These properties are inherited by child threads spawned from this thread. This |
| may have unexpected consequences when working with thread pools. The standard java |
| implementation of thread pools have worker threads spawn other worker threads. |
| As a result, local properties may propagate unpredictably. |
| </p></div></div> |
| </li><li name="org.apache.spark.SparkContext#setLogLevel" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="setLogLevel(logLevel:String):Unit"></a><a id="setLogLevel(String):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#setLogLevel(logLevel:String):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">setLogLevel</span><span class="params">(<span name="logLevel">logLevel: <span class="extype" name="scala.Predef.String">String</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Control our logLevel.</p><div class="fullcomment"><div class="comment cmt"><p>Control our logLevel. This overrides any user-defined log settings.</p></div><dl class="paramcmts block"><dt class="param">logLevel</dt><dd class="cmt"><p>The desired log level as a string. |
| Valid log levels include: ALL, DEBUG, ERROR, FATAL, INFO, OFF, TRACE, WARN</p></dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#sparkUser" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="sparkUser:String"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#sparkUser:String" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">val</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">sparkUser</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span> |
| </span> |
| |
| |
| </li><li name="org.apache.spark.SparkContext#startTime" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="startTime:Long"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#startTime:Long" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">val</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">startTime</span><span class="result">: <span class="extype" name="scala.Long">Long</span></span> |
| </span> |
| |
| |
| </li><li name="org.apache.spark.SparkContext#statusTracker" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="statusTracker:org.apache.spark.SparkStatusTracker"></a><a id="statusTracker:SparkStatusTracker"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#statusTracker:org.apache.spark.SparkStatusTracker" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">statusTracker</span><span class="result">: <a href="SparkStatusTracker.html" class="extype" name="org.apache.spark.SparkStatusTracker">SparkStatusTracker</a></span> |
| </span> |
| |
| |
| </li><li name="org.apache.spark.SparkContext#stop" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="stop():Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#stop():Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">stop</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">Shut down the SparkContext.</p> |
| </li><li name="org.apache.spark.SparkContext#submitJob" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="submitJob[T,U,R](rdd:org.apache.spark.rdd.RDD[T],processPartition:Iterator[T]=>U,partitions:Seq[Int],resultHandler:(Int,U)=>Unit,resultFunc:=>R):org.apache.spark.SimpleFutureAction[R]"></a><a id="submitJob[T,U,R](RDD[T],(Iterator[T])⇒U,Seq[Int],(Int,U)⇒Unit,⇒R):SimpleFutureAction[R]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#submitJob[T,U,R](rdd:org.apache.spark.rdd.RDD[T],processPartition:Iterator[T]=>U,partitions:Seq[Int],resultHandler:(Int,U)=>Unit,resultFunc:=>R):org.apache.spark.SimpleFutureAction[R]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">submitJob</span><span class="tparams">[<span name="T">T</span>, <span name="U">U</span>, <span name="R">R</span>]</span><span class="params">(<span name="rdd">rdd: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.submitJob.T">T</span>]</span>, <span name="processPartition">processPartition: (<span class="extype" name="scala.Iterator">Iterator</span>[<span class="extype" name="org.apache.spark.SparkContext.submitJob.T">T</span>]) ⇒ <span class="extype" name="org.apache.spark.SparkContext.submitJob.U">U</span></span>, <span name="partitions">partitions: <span class="extype" name="scala.Seq">Seq</span>[<span class="extype" name="scala.Int">Int</span>]</span>, <span name="resultHandler">resultHandler: (<span class="extype" name="scala.Int">Int</span>, <span class="extype" name="org.apache.spark.SparkContext.submitJob.U">U</span>) ⇒ <span class="extype" name="scala.Unit">Unit</span></span>, <span name="resultFunc">resultFunc: ⇒ <span class="extype" name="org.apache.spark.SparkContext.submitJob.R">R</span></span>)</span><span class="result">: <a href="SimpleFutureAction.html" class="extype" name="org.apache.spark.SimpleFutureAction">SimpleFutureAction</a>[<span class="extype" name="org.apache.spark.SparkContext.submitJob.R">R</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Submit a job for execution and return a FutureJob holding the result.</p><div class="fullcomment"><div class="comment cmt"><p>Submit a job for execution and return a FutureJob holding the result. |
| </p></div><dl class="paramcmts block"><dt class="param">rdd</dt><dd class="cmt"><p>target RDD to run tasks on</p></dd><dt class="param">processPartition</dt><dd class="cmt"><p>a function to run on each partition of the RDD</p></dd><dt class="param">partitions</dt><dd class="cmt"><p>set of partitions to run on; some jobs may not want to compute on all |
| partitions of the target RDD, e.g. for operations like <code>first()</code></p></dd><dt class="param">resultHandler</dt><dd class="cmt"><p>callback to pass each result to</p></dd><dt class="param">resultFunc</dt><dd class="cmt"><p>function to be executed when the result is ready</p></dd></dl></div> |
| </li><li name="scala.AnyRef#synchronized" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="synchronized[T0](x$1:=>T0):T0"></a><a id="synchronized[T0](⇒T0):T0"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#synchronized[T0](x$1:=>T0):T0" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">synchronized</span><span class="tparams">[<span name="T0">T0</span>]</span><span class="params">(<span name="arg0">arg0: ⇒ <span class="extype" name="java.lang.AnyRef.synchronized.T0">T0</span></span>)</span><span class="result">: <span class="extype" name="java.lang.AnyRef.synchronized.T0">T0</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#textFile" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="textFile(path:String,minPartitions:Int):org.apache.spark.rdd.RDD[String]"></a><a id="textFile(String,Int):RDD[String]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#textFile(path:String,minPartitions:Int):org.apache.spark.rdd.RDD[String]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">textFile</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="scala.Predef.String">String</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Read a text file from HDFS, a local file system (available on all nodes), or any |
| Hadoop-supported file system URI, and return it as an RDD of Strings.</p><div class="fullcomment"><div class="comment cmt"><p>Read a text file from HDFS, a local file system (available on all nodes), or any |
| Hadoop-supported file system URI, and return it as an RDD of Strings. |
| The text files must be encoded as UTF-8. |
| </p></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>path to the text file on a supported file system</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>suggested minimum number of partitions for the resulting RDD</p></dd><dt>returns</dt><dd class="cmt"><p>RDD of lines of the text file</p></dd></dl></div> |
| </li><li name="scala.AnyRef#toString" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="toString():String"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#toString():String" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">toString</span><span class="params">()</span><span class="result">: <span class="extype" name="java.lang.String">String</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef → Any</dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#uiWebUrl" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="uiWebUrl:Option[String]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#uiWebUrl:Option[String]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">uiWebUrl</span><span class="result">: <span class="extype" name="scala.Option">Option</span>[<span class="extype" name="scala.Predef.String">String</span>]</span> |
| </span> |
| |
| |
| </li><li name="org.apache.spark.SparkContext#union" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="union[T](first:org.apache.spark.rdd.RDD[T],rest:org.apache.spark.rdd.RDD[T]*)(implicitevidence$7:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="union[T](RDD[T],RDD[T]*)(ClassTag[T]):RDD[T]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#union[T](first:org.apache.spark.rdd.RDD[T],rest:org.apache.spark.rdd.RDD[T]*)(implicitevidence$7:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">union</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="first">first: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]</span>, <span name="rest">rest: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]*</span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Build the union of a list of RDDs passed as variable-length arguments.</p> |
| </li><li name="org.apache.spark.SparkContext#union" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="union[T](rdds:Seq[org.apache.spark.rdd.RDD[T]])(implicitevidence$6:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]"></a><a id="union[T](Seq[RDD[T]])(ClassTag[T]):RDD[T]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#union[T](rdds:Seq[org.apache.spark.rdd.RDD[T]])(implicitevidence$6:scala.reflect.ClassTag[T]):org.apache.spark.rdd.RDD[T]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">union</span><span class="tparams">[<span name="T">T</span>]</span><span class="params">(<span name="rdds">rdds: <span class="extype" name="scala.Seq">Seq</span>[<a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]]</span>)</span><span class="params">(<span class="implicit">implicit </span><span name="arg0">arg0: <span class="extype" name="scala.reflect.ClassTag">ClassTag</span>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]</span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[<span class="extype" name="org.apache.spark.SparkContext.union.T">T</span>]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Build the union of a list of RDDs.</p> |
| </li><li name="org.apache.spark.SparkContext#version" visbl="pub" class="indented0 " data-isabs="false" fullComment="no" group="Ungrouped"> |
| <a id="version:String"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#version:String" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">version</span><span class="result">: <span class="extype" name="scala.Predef.String">String</span></span> |
| </span> |
| |
| <p class="shortcomment cmt">The version of Spark on which this application is running.</p> |
| </li><li name="scala.AnyRef#wait" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="wait():Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#wait():Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">wait</span><span class="params">()</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd> |
| <span class="name">@throws</span><span class="args">(<span> |
| |
| <span class="defval" name="classOf[java.lang.InterruptedException]">...</span> |
| </span>)</span> |
| |
| </dd></dl></div> |
| </li><li name="scala.AnyRef#wait" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="wait(x$1:Long,x$2:Int):Unit"></a><a id="wait(Long,Int):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#wait(x$1:Long,x$2:Int):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">wait</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.Long">Long</span></span>, <span name="arg1">arg1: <span class="extype" name="scala.Int">Int</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd> |
| <span class="name">@throws</span><span class="args">(<span> |
| |
| <span class="defval" name="classOf[java.lang.InterruptedException]">...</span> |
| </span>)</span> |
| |
| </dd></dl></div> |
| </li><li name="scala.AnyRef#wait" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="wait(x$1:Long):Unit"></a><a id="wait(Long):Unit"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#wait(x$1:Long):Unit" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier">final </span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">wait</span><span class="params">(<span name="arg0">arg0: <span class="extype" name="scala.Long">Long</span></span>)</span><span class="result">: <span class="extype" name="scala.Unit">Unit</span></span> |
| </span> |
| |
| <div class="fullcomment"><dl class="attributes block"> <dt>Definition Classes</dt><dd>AnyRef</dd><dt>Annotations</dt><dd> |
| <span class="name">@throws</span><span class="args">(<span> |
| |
| <span class="defval" name="classOf[java.lang.InterruptedException]">...</span> |
| </span>)</span> |
| |
| <span class="name">@native</span><span class="args">()</span> |
| |
| </dd></dl></div> |
| </li><li name="org.apache.spark.SparkContext#wholeTextFiles" visbl="pub" class="indented0 " data-isabs="false" fullComment="yes" group="Ungrouped"> |
| <a id="wholeTextFiles(path:String,minPartitions:Int):org.apache.spark.rdd.RDD[(String,String)]"></a><a id="wholeTextFiles(String,Int):RDD[(String,String)]"></a> |
| <span class="permalink"> |
| <a href="../../../org/apache/spark/SparkContext.html#wholeTextFiles(path:String,minPartitions:Int):org.apache.spark.rdd.RDD[(String,String)]" title="Permalink"> |
| <i class="material-icons"></i> |
| </a> |
| </span> |
| <span class="modifier_kind"> |
| <span class="modifier"></span> |
| <span class="kind">def</span> |
| </span> |
| <span class="symbol"> |
| <span class="name">wholeTextFiles</span><span class="params">(<span name="path">path: <span class="extype" name="scala.Predef.String">String</span></span>, <span name="minPartitions">minPartitions: <span class="extype" name="scala.Int">Int</span> = <span class="symbol"><span class="name"><a href="#defaultMinPartitions:Int">defaultMinPartitions</a></span></span></span>)</span><span class="result">: <a href="rdd/RDD.html" class="extype" name="org.apache.spark.rdd.RDD">RDD</a>[(<span class="extype" name="scala.Predef.String">String</span>, <span class="extype" name="scala.Predef.String">String</span>)]</span> |
| </span> |
| |
| <p class="shortcomment cmt">Read a directory of text files from HDFS, a local file system (available on all nodes), or any |
| Hadoop-supported file system URI.</p><div class="fullcomment"><div class="comment cmt"><p>Read a directory of text files from HDFS, a local file system (available on all nodes), or any |
| Hadoop-supported file system URI. Each file is read as a single record and returned in a |
| key-value pair, where the key is the path of each file, the value is the content of each file. |
| The text files must be encoded as UTF-8.</p><p> For example, if you have the following files:</p><pre>hdfs:<span class="cmt">//a-hdfs-path/part-00000</span> |
| hdfs:<span class="cmt">//a-hdfs-path/part-00001</span> |
| ... |
| hdfs:<span class="cmt">//a-hdfs-path/part-nnnnn</span></pre><p>Do <code>val rdd = sparkContext.wholeTextFile("hdfs://a-hdfs-path")</code>,</p><p> then <code>rdd</code> contains</p><pre>(a-hdfs-path/part-<span class="num">00000</span>, its content) |
| (a-hdfs-path/part-<span class="num">00001</span>, its content) |
| ... |
| (a-hdfs-path/part-nnnnn, its content)</pre></div><dl class="paramcmts block"><dt class="param">path</dt><dd class="cmt"><p>Directory to the input data files, the path can be comma separated paths as the |
| list of inputs.</p></dd><dt class="param">minPartitions</dt><dd class="cmt"><p>A suggestion value of the minimal splitting number for input data.</p></dd><dt>returns</dt><dd class="cmt"><p>RDD representing tuples of file path and the corresponding file content</p></dd></dl><dl class="attributes block"> <dt>Note</dt><dd><span class="cmt"><p>Small files are preferred, large file is also allowable, but may cause bad performance.</p></span>, <span class="cmt"><p>On some filesystems, <code>.../path/*</code> can be a more efficient way to read all files |
| in a directory rather than <code>.../path/</code> or <code>.../path</code></p></span>, <span class="cmt"><p>Partitioning is determined by data locality. This may result in too few partitions |
| by default.</p></span></dd></dl></div> |
| </li> |
| </ol> |
| </div> |
| |
| |
| |
| |
| </div> |
| |
| <div id="inheritedMembers"> |
| <div class="parent" name="org.apache.spark.internal.Logging"> |
| <h3>Inherited from <span class="extype" name="org.apache.spark.internal.Logging">Logging</span></h3> |
| </div><div class="parent" name="scala.AnyRef"> |
| <h3>Inherited from <span class="extype" name="scala.AnyRef">AnyRef</span></h3> |
| </div><div class="parent" name="scala.Any"> |
| <h3>Inherited from <span class="extype" name="scala.Any">Any</span></h3> |
| </div> |
| |
| </div> |
| |
| <div id="groupedMembers"> |
| <div class="group" name="Ungrouped"> |
| <h3>Ungrouped</h3> |
| |
| </div> |
| </div> |
| |
| </div> |
| |
| <div id="tooltip"></div> |
| |
| <div id="footer"> </div> |
| </body> |
| </div> |
| </div> |
| </div> |
| </body> |
| </html> |