| |
| <!DOCTYPE html> |
| |
| <html> |
| <head> |
| <meta charset="utf-8" /> |
| <title>Input/Output — PySpark 3.3.4 documentation</title> |
| |
| <link rel="stylesheet" href="../../_static/css/index.73d71520a4ca3b99cfee5594769eaaae.css"> |
| |
| |
| <link rel="stylesheet" |
| href="../../_static/vendor/fontawesome/5.13.0/css/all.min.css"> |
| <link rel="preload" as="font" type="font/woff2" crossorigin |
| href="../../_static/vendor/fontawesome/5.13.0/webfonts/fa-solid-900.woff2"> |
| <link rel="preload" as="font" type="font/woff2" crossorigin |
| href="../../_static/vendor/fontawesome/5.13.0/webfonts/fa-brands-400.woff2"> |
| |
| |
| |
| <link rel="stylesheet" |
| href="../../_static/vendor/open-sans_all/1.44.1/index.css"> |
| <link rel="stylesheet" |
| href="../../_static/vendor/lato_latin-ext/1.44.1/index.css"> |
| |
| |
| <link rel="stylesheet" href="../../_static/basic.css" type="text/css" /> |
| <link rel="stylesheet" href="../../_static/pygments.css" type="text/css" /> |
| <link rel="stylesheet" type="text/css" href="../../_static/copybutton.css" /> |
| <link rel="stylesheet" type="text/css" href="../../_static/css/pyspark.css" /> |
| |
| <link rel="preload" as="script" href="../../_static/js/index.3da636dd464baa7582d2.js"> |
| |
| <script id="documentation_options" data-url_root="../../" src="../../_static/documentation_options.js"></script> |
| <script src="../../_static/jquery.js"></script> |
| <script src="../../_static/underscore.js"></script> |
| <script src="../../_static/doctools.js"></script> |
| <script src="../../_static/language_data.js"></script> |
| <script src="../../_static/clipboard.min.js"></script> |
| <script src="../../_static/copybutton.js"></script> |
| <script crossorigin="anonymous" integrity="sha256-Ae2Vz/4ePdIu6ZyI/5ZGsYnb+m0JlOmKPjt6XZ9JJkA=" src="https://cdnjs.cloudflare.com/ajax/libs/require.js/2.3.4/require.min.js"></script> |
| <script async="async" src="https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.5/latest.js?config=TeX-AMS-MML_HTMLorMML"></script> |
| <script type="text/x-mathjax-config">MathJax.Hub.Config({"tex2jax": {"inlineMath": [["$", "$"], ["\\(", "\\)"]], "processEscapes": true, "ignoreClass": "document", "processClass": "math|output_area"}})</script> |
| <link rel="canonical" href="https://spark.apache.org/docs/latest/api/python/reference/pyspark.sql/io.html" /> |
| <link rel="search" title="Search" href="../../search.html" /> |
| <link rel="next" title="pyspark.sql.DataFrameReader.csv" href="api/pyspark.sql.DataFrameReader.csv.html" /> |
| <link rel="prev" title="pyspark.sql.conf.RuntimeConfig" href="api/pyspark.sql.conf.RuntimeConfig.html" /> |
| <meta name="viewport" content="width=device-width, initial-scale=1" /> |
| <meta name="docsearch:language" content="en" /> |
| </head> |
| <body data-spy="scroll" data-target="#bd-toc-nav" data-offset="80"> |
| |
| <nav class="navbar navbar-light navbar-expand-lg bg-light fixed-top bd-navbar" id="navbar-main"> |
| <div class="container-xl"> |
| |
| <a class="navbar-brand" href="../../index.html"> |
| |
| <img src="../../_static/spark-logo-reverse.png" class="logo" alt="logo" /> |
| |
| </a> |
| <button class="navbar-toggler" type="button" data-toggle="collapse" data-target="#navbar-menu" aria-controls="navbar-menu" aria-expanded="false" aria-label="Toggle navigation"> |
| <span class="navbar-toggler-icon"></span> |
| </button> |
| |
| <div id="navbar-menu" class="col-lg-9 collapse navbar-collapse"> |
| <ul id="navbar-main-elements" class="navbar-nav mr-auto"> |
| |
| |
| <li class="nav-item "> |
| <a class="nav-link" href="../../getting_started/index.html">Getting Started</a> |
| </li> |
| |
| <li class="nav-item "> |
| <a class="nav-link" href="../../user_guide/index.html">User Guide</a> |
| </li> |
| |
| <li class="nav-item active"> |
| <a class="nav-link" href="../index.html">API Reference</a> |
| </li> |
| |
| <li class="nav-item "> |
| <a class="nav-link" href="../../development/index.html">Development</a> |
| </li> |
| |
| <li class="nav-item "> |
| <a class="nav-link" href="../../migration_guide/index.html">Migration Guide</a> |
| </li> |
| |
| |
| </ul> |
| |
| |
| |
| |
| <ul class="navbar-nav"> |
| |
| |
| </ul> |
| </div> |
| </div> |
| </nav> |
| |
| |
| <div class="container-xl"> |
| <div class="row"> |
| |
| <div class="col-12 col-md-3 bd-sidebar"><form class="bd-search d-flex align-items-center" action="../../search.html" method="get"> |
| <i class="icon fas fa-search"></i> |
| <input type="search" class="form-control" name="q" id="search-input" placeholder="Search the docs ..." aria-label="Search the docs ..." autocomplete="off" > |
| </form> |
| <nav class="bd-links" id="bd-docs-nav" aria-label="Main navigation"> |
| |
| <div class="bd-toc-item active"> |
| |
| |
| <ul class="nav bd-sidenav"> |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| <li class="active"> |
| <a href="index.html">Spark SQL</a> |
| <ul> |
| |
| <li class=""> |
| <a href="core_classes.html">Core Classes</a> |
| </li> |
| |
| <li class=""> |
| <a href="spark_session.html">Spark Session</a> |
| </li> |
| |
| <li class=""> |
| <a href="configuration.html">Configuration</a> |
| </li> |
| |
| <li class="active"> |
| <a href="">Input/Output</a> |
| </li> |
| |
| <li class=""> |
| <a href="dataframe.html">DataFrame</a> |
| </li> |
| |
| <li class=""> |
| <a href="column.html">Column</a> |
| </li> |
| |
| <li class=""> |
| <a href="data_types.html">Data Types</a> |
| </li> |
| |
| <li class=""> |
| <a href="row.html">Row</a> |
| </li> |
| |
| <li class=""> |
| <a href="functions.html">Functions</a> |
| </li> |
| |
| <li class=""> |
| <a href="window.html">Window</a> |
| </li> |
| |
| <li class=""> |
| <a href="grouping.html">Grouping</a> |
| </li> |
| |
| <li class=""> |
| <a href="catalog.html">Catalog</a> |
| </li> |
| |
| <li class=""> |
| <a href="observation.html">Observation</a> |
| </li> |
| |
| <li class=""> |
| <a href="avro.html">Avro</a> |
| </li> |
| |
| </ul> |
| </li> |
| |
| |
| |
| <li class=""> |
| <a href="../pyspark.pandas/index.html">Pandas API on Spark</a> |
| </li> |
| |
| |
| |
| <li class=""> |
| <a href="../pyspark.ss/index.html">Structured Streaming</a> |
| </li> |
| |
| |
| |
| <li class=""> |
| <a href="../pyspark.ml.html">MLlib (DataFrame-based)</a> |
| </li> |
| |
| |
| |
| <li class=""> |
| <a href="../pyspark.streaming.html">Spark Streaming</a> |
| </li> |
| |
| |
| |
| <li class=""> |
| <a href="../pyspark.mllib.html">MLlib (RDD-based)</a> |
| </li> |
| |
| |
| |
| <li class=""> |
| <a href="../pyspark.html">Spark Core</a> |
| </li> |
| |
| |
| |
| <li class=""> |
| <a href="../pyspark.resource.html">Resource Management</a> |
| </li> |
| |
| |
| |
| |
| |
| |
| |
| |
| </ul> |
| |
| </nav> |
| </div> |
| |
| |
| |
| <div class="d-none d-xl-block col-xl-2 bd-toc"> |
| |
| |
| <nav id="bd-toc-nav"> |
| <ul class="nav section-nav flex-column"> |
| |
| </ul> |
| </nav> |
| |
| |
| |
| </div> |
| |
| |
| |
| <main class="col-12 col-md-9 col-xl-7 py-md-5 pl-md-5 pr-md-4 bd-content" role="main"> |
| |
| <div> |
| |
| <div class="section" id="input-output"> |
| <h1>Input/Output<a class="headerlink" href="#input-output" title="Permalink to this headline">¶</a></h1> |
| <table class="longtable table autosummary"> |
| <colgroup> |
| <col style="width: 10%" /> |
| <col style="width: 90%" /> |
| </colgroup> |
| <tbody> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.csv.html#pyspark.sql.DataFrameReader.csv" title="pyspark.sql.DataFrameReader.csv"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.csv</span></code></a>(path[, schema, sep, …])</p></td> |
| <td><p>Loads a CSV file and returns the result as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.format.html#pyspark.sql.DataFrameReader.format" title="pyspark.sql.DataFrameReader.format"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.format</span></code></a>(source)</p></td> |
| <td><p>Specifies the input data source format.</p></td> |
| </tr> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.jdbc.html#pyspark.sql.DataFrameReader.jdbc" title="pyspark.sql.DataFrameReader.jdbc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.jdbc</span></code></a>(url, table[, column, …])</p></td> |
| <td><p>Construct a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> representing the database table named <code class="docutils literal notranslate"><span class="pre">table</span></code> accessible via JDBC URL <code class="docutils literal notranslate"><span class="pre">url</span></code> and connection <code class="docutils literal notranslate"><span class="pre">properties</span></code>.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.json.html#pyspark.sql.DataFrameReader.json" title="pyspark.sql.DataFrameReader.json"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.json</span></code></a>(path[, schema, …])</p></td> |
| <td><p>Loads JSON files and returns the results as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td> |
| </tr> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.load.html#pyspark.sql.DataFrameReader.load" title="pyspark.sql.DataFrameReader.load"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.load</span></code></a>([path, format, schema])</p></td> |
| <td><p>Loads data from a data source and returns it as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.option.html#pyspark.sql.DataFrameReader.option" title="pyspark.sql.DataFrameReader.option"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.option</span></code></a>(key, value)</p></td> |
| <td><p>Adds an input option for the underlying data source.</p></td> |
| </tr> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.options.html#pyspark.sql.DataFrameReader.options" title="pyspark.sql.DataFrameReader.options"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.options</span></code></a>(**options)</p></td> |
| <td><p>Adds input options for the underlying data source.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.orc.html#pyspark.sql.DataFrameReader.orc" title="pyspark.sql.DataFrameReader.orc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.orc</span></code></a>(path[, mergeSchema, …])</p></td> |
| <td><p>Loads ORC files, returning the result as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td> |
| </tr> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.parquet.html#pyspark.sql.DataFrameReader.parquet" title="pyspark.sql.DataFrameReader.parquet"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.parquet</span></code></a>(*paths, **options)</p></td> |
| <td><p>Loads Parquet files, returning the result as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.schema.html#pyspark.sql.DataFrameReader.schema" title="pyspark.sql.DataFrameReader.schema"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.schema</span></code></a>(schema)</p></td> |
| <td><p>Specifies the input schema.</p></td> |
| </tr> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.table.html#pyspark.sql.DataFrameReader.table" title="pyspark.sql.DataFrameReader.table"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.table</span></code></a>(tableName)</p></td> |
| <td><p>Returns the specified table as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.text.html#pyspark.sql.DataFrameReader.text" title="pyspark.sql.DataFrameReader.text"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.text</span></code></a>(paths[, wholetext, …])</p></td> |
| <td><p>Loads text files and returns a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> whose schema starts with a string column named “value”, and followed by partitioned columns if there are any.</p></td> |
| </tr> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.bucketBy.html#pyspark.sql.DataFrameWriter.bucketBy" title="pyspark.sql.DataFrameWriter.bucketBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.bucketBy</span></code></a>(numBuckets, col, *cols)</p></td> |
| <td><p>Buckets the output by the given columns.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.csv.html#pyspark.sql.DataFrameWriter.csv" title="pyspark.sql.DataFrameWriter.csv"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.csv</span></code></a>(path[, mode, …])</p></td> |
| <td><p>Saves the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> in CSV format at the specified path.</p></td> |
| </tr> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.format.html#pyspark.sql.DataFrameWriter.format" title="pyspark.sql.DataFrameWriter.format"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.format</span></code></a>(source)</p></td> |
| <td><p>Specifies the underlying output data source.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.insertInto.html#pyspark.sql.DataFrameWriter.insertInto" title="pyspark.sql.DataFrameWriter.insertInto"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.insertInto</span></code></a>(tableName[, …])</p></td> |
| <td><p>Inserts the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> to the specified table.</p></td> |
| </tr> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.jdbc.html#pyspark.sql.DataFrameWriter.jdbc" title="pyspark.sql.DataFrameWriter.jdbc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.jdbc</span></code></a>(url, table[, mode, …])</p></td> |
| <td><p>Saves the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> to an external database table via JDBC.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.json.html#pyspark.sql.DataFrameWriter.json" title="pyspark.sql.DataFrameWriter.json"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.json</span></code></a>(path[, mode, …])</p></td> |
| <td><p>Saves the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> in JSON format (<a class="reference external" href="http://jsonlines.org/">JSON Lines text format or newline-delimited JSON</a>) at the specified path.</p></td> |
| </tr> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.mode.html#pyspark.sql.DataFrameWriter.mode" title="pyspark.sql.DataFrameWriter.mode"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.mode</span></code></a>(saveMode)</p></td> |
| <td><p>Specifies the behavior when data or table already exists.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.option.html#pyspark.sql.DataFrameWriter.option" title="pyspark.sql.DataFrameWriter.option"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.option</span></code></a>(key, value)</p></td> |
| <td><p>Adds an output option for the underlying data source.</p></td> |
| </tr> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.options.html#pyspark.sql.DataFrameWriter.options" title="pyspark.sql.DataFrameWriter.options"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.options</span></code></a>(**options)</p></td> |
| <td><p>Adds output options for the underlying data source.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.orc.html#pyspark.sql.DataFrameWriter.orc" title="pyspark.sql.DataFrameWriter.orc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.orc</span></code></a>(path[, mode, …])</p></td> |
| <td><p>Saves the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> in ORC format at the specified path.</p></td> |
| </tr> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.parquet.html#pyspark.sql.DataFrameWriter.parquet" title="pyspark.sql.DataFrameWriter.parquet"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.parquet</span></code></a>(path[, mode, …])</p></td> |
| <td><p>Saves the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> in Parquet format at the specified path.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.partitionBy.html#pyspark.sql.DataFrameWriter.partitionBy" title="pyspark.sql.DataFrameWriter.partitionBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.partitionBy</span></code></a>(*cols)</p></td> |
| <td><p>Partitions the output by the given columns on the file system.</p></td> |
| </tr> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.save.html#pyspark.sql.DataFrameWriter.save" title="pyspark.sql.DataFrameWriter.save"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.save</span></code></a>([path, format, mode, …])</p></td> |
| <td><p>Saves the contents of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> to a data source.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.saveAsTable.html#pyspark.sql.DataFrameWriter.saveAsTable" title="pyspark.sql.DataFrameWriter.saveAsTable"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.saveAsTable</span></code></a>(name[, format, …])</p></td> |
| <td><p>Saves the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> as the specified table.</p></td> |
| </tr> |
| <tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.sortBy.html#pyspark.sql.DataFrameWriter.sortBy" title="pyspark.sql.DataFrameWriter.sortBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.sortBy</span></code></a>(col, *cols)</p></td> |
| <td><p>Sorts the output in each bucket by the given columns on the file system.</p></td> |
| </tr> |
| <tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.text.html#pyspark.sql.DataFrameWriter.text" title="pyspark.sql.DataFrameWriter.text"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.text</span></code></a>(path[, compression, …])</p></td> |
| <td><p>Saves the content of the DataFrame in a text file at the specified path.</p></td> |
| </tr> |
| </tbody> |
| </table> |
| </div> |
| |
| |
| </div> |
| |
| |
| <div class='prev-next-bottom'> |
| |
| <a class='left-prev' id="prev-link" href="api/pyspark.sql.conf.RuntimeConfig.html" title="previous page">pyspark.sql.conf.RuntimeConfig</a> |
| <a class='right-next' id="next-link" href="api/pyspark.sql.DataFrameReader.csv.html" title="next page">pyspark.sql.DataFrameReader.csv</a> |
| |
| </div> |
| |
| </main> |
| |
| |
| </div> |
| </div> |
| |
| |
| <script src="../../_static/js/index.3da636dd464baa7582d2.js"></script> |
| |
| |
| <footer class="footer mt-5 mt-md-0"> |
| <div class="container"> |
| <p> |
| © Copyright .<br/> |
| Created using <a href="http://sphinx-doc.org/">Sphinx</a> 3.0.4.<br/> |
| </p> |
| </div> |
| </footer> |
| </body> |
| </html> |