blob: 32d982cbac33f685232d08a2007d08770d8e02fa [file] [log] [blame]
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Save the contents of SparkDataFrame to a data source.</title>
<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
<link rel="stylesheet" type="text/css" href="R.css" />
<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
<script>hljs.initHighlightingOnLoad();</script>
</head><body>
<table width="100%" summary="page for write.df {SparkR}"><tr><td>write.df {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
<h2>Save the contents of SparkDataFrame to a data source.</h2>
<h3>Description</h3>
<p>The data source is specified by the <code>source</code> and a set of options (...).
If <code>source</code> is not specified, the default data source configured by
spark.sql.sources.default will be used.
</p>
<h3>Usage</h3>
<pre>
write.df(df, path = NULL, ...)
saveDF(df, path, source = NULL, mode = "error", ...)
write.df(df, path = NULL, ...)
## S4 method for signature 'SparkDataFrame'
write.df(df, path = NULL, source = NULL,
mode = "error", ...)
## S4 method for signature 'SparkDataFrame,character'
saveDF(df, path, source = NULL,
mode = "error", ...)
</pre>
<h3>Arguments</h3>
<table summary="R argblock">
<tr valign="top"><td><code>df</code></td>
<td>
<p>a SparkDataFrame.</p>
</td></tr>
<tr valign="top"><td><code>path</code></td>
<td>
<p>a name for the table.</p>
</td></tr>
<tr valign="top"><td><code>...</code></td>
<td>
<p>additional argument(s) passed to the method.</p>
</td></tr>
<tr valign="top"><td><code>source</code></td>
<td>
<p>a name for external data source.</p>
</td></tr>
<tr valign="top"><td><code>mode</code></td>
<td>
<p>one of 'append', 'overwrite', 'error', 'errorifexists', 'ignore'
save mode (it is 'error' by default)</p>
</td></tr>
</table>
<h3>Details</h3>
<p>Additionally, mode is used to specify the behavior of the save operation when data already
exists in the data source. There are four modes:
</p>
<ul>
<li><p> 'append': Contents of this SparkDataFrame are expected to be appended to existing data.
</p>
</li>
<li><p> 'overwrite': Existing data is expected to be overwritten by the contents of this
SparkDataFrame.
</p>
</li>
<li><p> 'error' or 'errorifexists': An exception is expected to be thrown.
</p>
</li>
<li><p> 'ignore': The save operation is expected to not save the contents of the SparkDataFrame
and to not change the existing data.
</p>
</li></ul>
<h3>Note</h3>
<p>write.df since 1.4.0
</p>
<p>saveDF since 1.4.0
</p>
<h3>See Also</h3>
<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
<code><a href="coltypes.html">coltypes</a></code>,
<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
<code><a href="dtypes.html">dtypes</a></code>, <code><a href="except.html">except</a></code>,
<code><a href="explain.html">explain</a></code>, <code><a href="filter.html">filter</a></code>,
<code><a href="first.html">first</a></code>, <code><a href="gapplyCollect.html">gapplyCollect</a></code>,
<code><a href="gapply.html">gapply</a></code>, <code><a href="getNumPartitions.html">getNumPartitions</a></code>,
<code><a href="groupBy.html">group_by</a></code>, <code><a href="head.html">head</a></code>,
<code><a href="hint.html">hint</a></code>, <code><a href="histogram.html">histogram</a></code>,
<code><a href="insertInto.html">insertInto</a></code>, <code><a href="intersect.html">intersect</a></code>,
<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
<code><a href="rbind.html">rbind</a></code>, <code><a href="registerTempTable-deprecated.html">registerTempTable</a></code>,
<code><a href="rename.html">rename</a></code>, <code><a href="repartition.html">repartition</a></code>,
<code><a href="rollup.html">rollup</a></code>, <code><a href="sample.html">sample</a></code>,
<code><a href="saveAsTable.html">saveAsTable</a></code>, <code><a href="schema.html">schema</a></code>,
<code><a href="selectExpr.html">selectExpr</a></code>, <code><a href="select.html">select</a></code>,
<code><a href="showDF.html">showDF</a></code>, <code><a href="show.html">show</a></code>,
<code><a href="storageLevel.html">storageLevel</a></code>, <code><a href="str.html">str</a></code>,
<code><a href="subset.html">subset</a></code>, <code><a href="summary.html">summary</a></code>,
<code><a href="take.html">take</a></code>, <code><a href="toJSON.html">toJSON</a></code>,
<code><a href="unionByName.html">unionByName</a></code>, <code><a href="union.html">union</a></code>,
<code><a href="unpersist.html">unpersist</a></code>, <code><a href="withColumn.html">withColumn</a></code>,
<code><a href="withWatermark.html">withWatermark</a></code>, <code><a href="with.html">with</a></code>,
<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
</p>
<h3>Examples</h3>
<pre><code class="r">## Not run:
##D sparkR.session()
##D path &lt;- &quot;path/to/file.json&quot;
##D df &lt;- read.json(path)
##D write.df(df, &quot;myfile&quot;, &quot;parquet&quot;, &quot;overwrite&quot;)
##D saveDF(df, parquetPath2, &quot;parquet&quot;, mode = &quot;append&quot;, mergeSchema = TRUE)
## End(Not run)
</code></pre>
<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.2 <a href="00Index.html">Index</a>]</div>
</body></html>