|  |  | 
|  | <!DOCTYPE html> | 
|  | <!--[if lt IE 7]>      <html class="no-js lt-ie9 lt-ie8 lt-ie7"> <![endif]--> | 
|  | <!--[if IE 7]>         <html class="no-js lt-ie9 lt-ie8"> <![endif]--> | 
|  | <!--[if IE 8]>         <html class="no-js lt-ie9"> <![endif]--> | 
|  | <!--[if gt IE 8]><!--> <html class="no-js"> <!--<![endif]--> | 
|  | <head> | 
|  | <meta charset="utf-8"> | 
|  | <meta http-equiv="X-UA-Compatible" content="IE=edge,chrome=1"> | 
|  | <meta name="viewport" content="width=device-width, initial-scale=1.0"> | 
|  |  | 
|  | <title>Generic File Source Options - Spark 3.5.0 Documentation</title> | 
|  |  | 
|  |  | 
|  |  | 
|  |  | 
|  |  | 
|  | <link rel="stylesheet" href="css/bootstrap.min.css"> | 
|  | <link rel="preconnect" href="https://fonts.googleapis.com"> | 
|  | <link rel="preconnect" href="https://fonts.gstatic.com" crossorigin> | 
|  | <link href="https://fonts.googleapis.com/css2?family=DM+Sans:ital,wght@0,400;0,500;0,700;1,400;1,500;1,700&Courier+Prime:wght@400;700&display=swap" rel="stylesheet"> | 
|  | <link href="css/custom.css" rel="stylesheet"> | 
|  | <script src="js/vendor/modernizr-2.6.1-respond-1.1.0.min.js"></script> | 
|  |  | 
|  | <link rel="stylesheet" href="css/pygments-default.css"> | 
|  | <link rel="stylesheet" href="css/docsearch.min.css" /> | 
|  | <link rel="stylesheet" href="css/docsearch.css"> | 
|  |  | 
|  | <!-- Matomo --> | 
|  | <script type="text/javascript"> | 
|  | var _paq = window._paq = window._paq || []; | 
|  | /* tracker methods like "setCustomDimension" should be called before "trackPageView" */ | 
|  | _paq.push(["disableCookies"]); | 
|  | _paq.push(['trackPageView']); | 
|  | _paq.push(['enableLinkTracking']); | 
|  | (function() { | 
|  | var u="https://analytics.apache.org/"; | 
|  | _paq.push(['setTrackerUrl', u+'matomo.php']); | 
|  | _paq.push(['setSiteId', '40']); | 
|  | var d=document, g=d.createElement('script'), s=d.getElementsByTagName('script')[0]; | 
|  | g.async=true; g.src=u+'matomo.js'; s.parentNode.insertBefore(g,s); | 
|  | })(); | 
|  | </script> | 
|  | <!-- End Matomo Code --> | 
|  | </head> | 
|  | <body class="global"> | 
|  | <!--[if lt IE 7]> | 
|  | <p class="chromeframe">You are using an outdated browser. <a href="https://browsehappy.com/">Upgrade your browser today</a> or <a href="http://www.google.com/chromeframe/?redirect=true">install Google Chrome Frame</a> to better experience this site.</p> | 
|  | <![endif]--> | 
|  |  | 
|  | <!-- This code is taken from http://twitter.github.com/bootstrap/examples/hero.html --> | 
|  |  | 
|  | <nav class="navbar navbar-expand-lg navbar-dark p-0 px-4 fixed-top" style="background: #1d6890;" id="topbar"> | 
|  | <div class="navbar-brand"><a href="index.html"> | 
|  | <img src="img/spark-logo-rev.svg" width="141" height="72"/></a><span class="version">3.5.0</span> | 
|  | </div> | 
|  | <button class="navbar-toggler" type="button" data-toggle="collapse" | 
|  | data-target="#navbarCollapse" aria-controls="navbarCollapse" | 
|  | aria-expanded="false" aria-label="Toggle navigation"> | 
|  | <span class="navbar-toggler-icon"></span> | 
|  | </button> | 
|  | <div class="collapse navbar-collapse" id="navbarCollapse"> | 
|  | <ul class="navbar-nav me-auto"> | 
|  | <li class="nav-item"><a href="index.html" class="nav-link">Overview</a></li> | 
|  |  | 
|  | <li class="nav-item dropdown"> | 
|  | <a href="#" class="nav-link dropdown-toggle" id="navbarQuickStart" role="button" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false">Programming Guides</a> | 
|  | <div class="dropdown-menu" aria-labelledby="navbarQuickStart"> | 
|  | <a class="dropdown-item" href="quick-start.html">Quick Start</a> | 
|  | <a class="dropdown-item" href="rdd-programming-guide.html">RDDs, Accumulators, Broadcasts Vars</a> | 
|  | <a class="dropdown-item" href="sql-programming-guide.html">SQL, DataFrames, and Datasets</a> | 
|  | <a class="dropdown-item" href="structured-streaming-programming-guide.html">Structured Streaming</a> | 
|  | <a class="dropdown-item" href="streaming-programming-guide.html">Spark Streaming (DStreams)</a> | 
|  | <a class="dropdown-item" href="ml-guide.html">MLlib (Machine Learning)</a> | 
|  | <a class="dropdown-item" href="graphx-programming-guide.html">GraphX (Graph Processing)</a> | 
|  | <a class="dropdown-item" href="sparkr.html">SparkR (R on Spark)</a> | 
|  | <a class="dropdown-item" href="api/python/getting_started/index.html">PySpark (Python on Spark)</a> | 
|  | </div> | 
|  | </li> | 
|  |  | 
|  | <li class="nav-item dropdown"> | 
|  | <a href="#" class="nav-link dropdown-toggle" id="navbarAPIDocs" role="button" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false">API Docs</a> | 
|  | <div class="dropdown-menu" aria-labelledby="navbarAPIDocs"> | 
|  | <a class="dropdown-item" href="api/scala/org/apache/spark/index.html">Scala</a> | 
|  | <a class="dropdown-item" href="api/java/index.html">Java</a> | 
|  | <a class="dropdown-item" href="api/python/index.html">Python</a> | 
|  | <a class="dropdown-item" href="api/R/index.html">R</a> | 
|  | <a class="dropdown-item" href="api/sql/index.html">SQL, Built-in Functions</a> | 
|  | </div> | 
|  | </li> | 
|  |  | 
|  | <li class="nav-item dropdown"> | 
|  | <a href="#" class="nav-link dropdown-toggle" id="navbarDeploying" role="button" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false">Deploying</a> | 
|  | <div class="dropdown-menu" aria-labelledby="navbarDeploying"> | 
|  | <a class="dropdown-item" href="cluster-overview.html">Overview</a> | 
|  | <a class="dropdown-item" href="submitting-applications.html">Submitting Applications</a> | 
|  | <div class="dropdown-divider"></div> | 
|  | <a class="dropdown-item" href="spark-standalone.html">Spark Standalone</a> | 
|  | <a class="dropdown-item" href="running-on-mesos.html">Mesos</a> | 
|  | <a class="dropdown-item" href="running-on-yarn.html">YARN</a> | 
|  | <a class="dropdown-item" href="running-on-kubernetes.html">Kubernetes</a> | 
|  | </div> | 
|  | </li> | 
|  |  | 
|  | <li class="nav-item dropdown"> | 
|  | <a href="#" class="nav-link dropdown-toggle" id="navbarMore" role="button" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false">More</a> | 
|  | <div class="dropdown-menu" aria-labelledby="navbarMore"> | 
|  | <a class="dropdown-item" href="configuration.html">Configuration</a> | 
|  | <a class="dropdown-item" href="monitoring.html">Monitoring</a> | 
|  | <a class="dropdown-item" href="tuning.html">Tuning Guide</a> | 
|  | <a class="dropdown-item" href="job-scheduling.html">Job Scheduling</a> | 
|  | <a class="dropdown-item" href="security.html">Security</a> | 
|  | <a class="dropdown-item" href="hardware-provisioning.html">Hardware Provisioning</a> | 
|  | <a class="dropdown-item" href="migration-guide.html">Migration Guide</a> | 
|  | <div class="dropdown-divider"></div> | 
|  | <a class="dropdown-item" href="building-spark.html">Building Spark</a> | 
|  | <a class="dropdown-item" href="https://spark.apache.org/contributing.html">Contributing to Spark</a> | 
|  | <a class="dropdown-item" href="https://spark.apache.org/third-party-projects.html">Third Party Projects</a> | 
|  | </div> | 
|  | </li> | 
|  |  | 
|  | <li class="nav-item"> | 
|  | <input type="text" id="docsearch-input" placeholder="Search the docs…"> | 
|  | </li> | 
|  | </ul> | 
|  | <!--<span class="navbar-text navbar-right"><span class="version-text">v3.5.0</span></span>--> | 
|  | </div> | 
|  | </nav> | 
|  |  | 
|  |  | 
|  |  | 
|  | <div class="container"> | 
|  |  | 
|  |  | 
|  |  | 
|  | <div class="left-menu-wrapper"> | 
|  | <div class="left-menu"> | 
|  | <h3><a href="sql-programming-guide.html">Spark SQL Guide</a></h3> | 
|  |  | 
|  | <ul> | 
|  |  | 
|  | <li> | 
|  | <a href="sql-getting-started.html"> | 
|  |  | 
|  | Getting Started | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources.html"> | 
|  |  | 
|  | Data Sources | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <ul> | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources-load-save-functions.html"> | 
|  |  | 
|  | Generic Load/Save Functions | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources-generic-options.html"> | 
|  |  | 
|  | Generic File Source Options | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources-parquet.html"> | 
|  |  | 
|  | Parquet Files | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources-orc.html"> | 
|  |  | 
|  | ORC Files | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources-json.html"> | 
|  |  | 
|  | JSON Files | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources-csv.html"> | 
|  |  | 
|  | CSV Files | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources-text.html"> | 
|  |  | 
|  | Text Files | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources-hive-tables.html"> | 
|  |  | 
|  | Hive Tables | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources-jdbc.html"> | 
|  |  | 
|  | JDBC To Other Databases | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources-avro.html"> | 
|  |  | 
|  | Avro Files | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources-protobuf.html"> | 
|  |  | 
|  | Protobuf data | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources-binaryFile.html"> | 
|  |  | 
|  | Whole Binary Files | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-data-sources-troubleshooting.html"> | 
|  |  | 
|  | Troubleshooting | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | </ul> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-performance-tuning.html"> | 
|  |  | 
|  | Performance Tuning | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-distributed-sql-engine.html"> | 
|  |  | 
|  | Distributed SQL Engine | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-pyspark-pandas-with-arrow.html"> | 
|  |  | 
|  | PySpark Usage Guide for Pandas with Apache Arrow | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-migration-guide.html"> | 
|  |  | 
|  | Migration Guide | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-ref.html"> | 
|  |  | 
|  | SQL Reference | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | <li> | 
|  | <a href="sql-error-conditions.html"> | 
|  |  | 
|  | Error Conditions | 
|  |  | 
|  | </a> | 
|  | </li> | 
|  |  | 
|  |  | 
|  |  | 
|  | </ul> | 
|  |  | 
|  | </div> | 
|  | </div> | 
|  |  | 
|  | <input id="nav-trigger" class="nav-trigger" checked type="checkbox"> | 
|  | <label for="nav-trigger"></label> | 
|  | <div class="content-with-sidebar mr-3" id="content"> | 
|  |  | 
|  | <h1 class="title">Generic File Source Options</h1> | 
|  |  | 
|  |  | 
|  | <ul id="markdown-toc"> | 
|  | <li><a href="#ignore-corrupt-files" id="markdown-toc-ignore-corrupt-files">Ignore Corrupt Files</a></li> | 
|  | <li><a href="#ignore-missing-files" id="markdown-toc-ignore-missing-files">Ignore Missing Files</a></li> | 
|  | <li><a href="#path-glob-filter" id="markdown-toc-path-glob-filter">Path Glob Filter</a></li> | 
|  | <li><a href="#recursive-file-lookup" id="markdown-toc-recursive-file-lookup">Recursive File Lookup</a></li> | 
|  | <li><a href="#modification-time-path-filters" id="markdown-toc-modification-time-path-filters">Modification Time Path Filters</a></li> | 
|  | </ul> | 
|  |  | 
|  | <p>These generic options/configurations are effective only when using file-based sources: parquet, orc, avro, json, csv, text.</p> | 
|  |  | 
|  | <p>Please note that the hierarchy of directories used in examples below are:</p> | 
|  |  | 
|  | <figure class="highlight"><pre><code class="language-text" data-lang="text">dir1/ | 
|  | ├── dir2/ | 
|  | │    └── file2.parquet (schema: <file: string>, content: "file2.parquet") | 
|  | └── file1.parquet (schema: <file, string>, content: "file1.parquet") | 
|  | └── file3.json (schema: <file, string>, content: "{'file':'corrupt.json'}")</code></pre></figure> | 
|  |  | 
|  | <h3 id="ignore-corrupt-files">Ignore Corrupt Files</h3> | 
|  |  | 
|  | <p>Spark allows you to use the configuration <code class="language-plaintext highlighter-rouge">spark.sql.files.ignoreCorruptFiles</code> or the data source option <code class="language-plaintext highlighter-rouge">ignoreCorruptFiles</code> to ignore corrupt files while reading data | 
|  | from files. When set to true, the Spark jobs will continue to run when encountering corrupted files and | 
|  | the contents that have been read will still be returned.</p> | 
|  |  | 
|  | <p>To ignore corrupt files while reading data files, you can use:</p> | 
|  |  | 
|  | <div class="codetabs"> | 
|  |  | 
|  | <div data-lang="python"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="c1"># enable ignore corrupt files via the data source option | 
|  | # dir1/file3.json is corrupt from parquet's view | 
|  | </span><span class="n">test_corrupt_df0</span> <span class="o">=</span> <span class="n">spark</span><span class="p">.</span><span class="n">read</span><span class="p">.</span><span class="n">option</span><span class="p">(</span><span class="s">"ignoreCorruptFiles"</span><span class="p">,</span> <span class="s">"true"</span><span class="p">)</span>\ | 
|  | <span class="p">.</span><span class="n">parquet</span><span class="p">(</span><span class="s">"examples/src/main/resources/dir1/"</span><span class="p">,</span> | 
|  | <span class="s">"examples/src/main/resources/dir1/dir2/"</span><span class="p">)</span> | 
|  | <span class="n">test_corrupt_df0</span><span class="p">.</span><span class="n">show</span><span class="p">()</span> | 
|  | <span class="c1"># +-------------+ | 
|  | # |         file| | 
|  | # +-------------+ | 
|  | # |file1.parquet| | 
|  | # |file2.parquet| | 
|  | # +-------------+ | 
|  | </span> | 
|  | <span class="c1"># enable ignore corrupt files via the configuration | 
|  | </span><span class="n">spark</span><span class="p">.</span><span class="n">sql</span><span class="p">(</span><span class="s">"set spark.sql.files.ignoreCorruptFiles=true"</span><span class="p">)</span> | 
|  | <span class="c1"># dir1/file3.json is corrupt from parquet's view | 
|  | </span><span class="n">test_corrupt_df1</span> <span class="o">=</span> <span class="n">spark</span><span class="p">.</span><span class="n">read</span><span class="p">.</span><span class="n">parquet</span><span class="p">(</span><span class="s">"examples/src/main/resources/dir1/"</span><span class="p">,</span> | 
|  | <span class="s">"examples/src/main/resources/dir1/dir2/"</span><span class="p">)</span> | 
|  | <span class="n">test_corrupt_df1</span><span class="p">.</span><span class="n">show</span><span class="p">()</span> | 
|  | <span class="c1"># +-------------+ | 
|  | # |         file| | 
|  | # +-------------+ | 
|  | # |file1.parquet| | 
|  | # |file2.parquet| | 
|  | # +-------------+</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/python/sql/datasource.py" in the Spark repo.</small></div> | 
|  | </div> | 
|  |  | 
|  | <div data-lang="scala"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="c1">// enable ignore corrupt files via the data source option</span> | 
|  | <span class="c1">// dir1/file3.json is corrupt from parquet's view</span> | 
|  | <span class="k">val</span> <span class="nv">testCorruptDF0</span> <span class="k">=</span> <span class="nv">spark</span><span class="o">.</span><span class="py">read</span><span class="o">.</span><span class="py">option</span><span class="o">(</span><span class="s">"ignoreCorruptFiles"</span><span class="o">,</span> <span class="s">"true"</span><span class="o">).</span><span class="py">parquet</span><span class="o">(</span> | 
|  | <span class="s">"examples/src/main/resources/dir1/"</span><span class="o">,</span> | 
|  | <span class="s">"examples/src/main/resources/dir1/dir2/"</span><span class="o">)</span> | 
|  | <span class="nv">testCorruptDF0</span><span class="o">.</span><span class="py">show</span><span class="o">()</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |         file|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |file1.parquet|</span> | 
|  | <span class="c1">// |file2.parquet|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  |  | 
|  | <span class="c1">// enable ignore corrupt files via the configuration</span> | 
|  | <span class="nv">spark</span><span class="o">.</span><span class="py">sql</span><span class="o">(</span><span class="s">"set spark.sql.files.ignoreCorruptFiles=true"</span><span class="o">)</span> | 
|  | <span class="c1">// dir1/file3.json is corrupt from parquet's view</span> | 
|  | <span class="k">val</span> <span class="nv">testCorruptDF1</span> <span class="k">=</span> <span class="nv">spark</span><span class="o">.</span><span class="py">read</span><span class="o">.</span><span class="py">parquet</span><span class="o">(</span> | 
|  | <span class="s">"examples/src/main/resources/dir1/"</span><span class="o">,</span> | 
|  | <span class="s">"examples/src/main/resources/dir1/dir2/"</span><span class="o">)</span> | 
|  | <span class="nv">testCorruptDF1</span><span class="o">.</span><span class="py">show</span><span class="o">()</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |         file|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |file1.parquet|</span> | 
|  | <span class="c1">// |file2.parquet|</span> | 
|  | <span class="c1">// +-------------+</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/scala/org/apache/spark/examples/sql/SQLDataSourceExample.scala" in the Spark repo.</small></div> | 
|  | </div> | 
|  |  | 
|  | <div data-lang="java"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="c1">// enable ignore corrupt files via the data source option</span> | 
|  | <span class="c1">// dir1/file3.json is corrupt from parquet's view</span> | 
|  | <span class="nc">Dataset</span><span class="o"><</span><span class="nc">Row</span><span class="o">></span> <span class="n">testCorruptDF0</span> <span class="o">=</span> <span class="n">spark</span><span class="o">.</span><span class="na">read</span><span class="o">().</span><span class="na">option</span><span class="o">(</span><span class="s">"ignoreCorruptFiles"</span><span class="o">,</span> <span class="s">"true"</span><span class="o">).</span><span class="na">parquet</span><span class="o">(</span> | 
|  | <span class="s">"examples/src/main/resources/dir1/"</span><span class="o">,</span> | 
|  | <span class="s">"examples/src/main/resources/dir1/dir2/"</span><span class="o">);</span> | 
|  | <span class="n">testCorruptDF0</span><span class="o">.</span><span class="na">show</span><span class="o">();</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |         file|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |file1.parquet|</span> | 
|  | <span class="c1">// |file2.parquet|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  |  | 
|  | <span class="c1">// enable ignore corrupt files via the configuration</span> | 
|  | <span class="n">spark</span><span class="o">.</span><span class="na">sql</span><span class="o">(</span><span class="s">"set spark.sql.files.ignoreCorruptFiles=true"</span><span class="o">);</span> | 
|  | <span class="c1">// dir1/file3.json is corrupt from parquet's view</span> | 
|  | <span class="nc">Dataset</span><span class="o"><</span><span class="nc">Row</span><span class="o">></span> <span class="n">testCorruptDF1</span> <span class="o">=</span> <span class="n">spark</span><span class="o">.</span><span class="na">read</span><span class="o">().</span><span class="na">parquet</span><span class="o">(</span> | 
|  | <span class="s">"examples/src/main/resources/dir1/"</span><span class="o">,</span> | 
|  | <span class="s">"examples/src/main/resources/dir1/dir2/"</span><span class="o">);</span> | 
|  | <span class="n">testCorruptDF1</span><span class="o">.</span><span class="na">show</span><span class="o">();</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |         file|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |file1.parquet|</span> | 
|  | <span class="c1">// |file2.parquet|</span> | 
|  | <span class="c1">// +-------------+</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/java/org/apache/spark/examples/sql/JavaSQLDataSourceExample.java" in the Spark repo.</small></div> | 
|  | </div> | 
|  |  | 
|  | <div data-lang="r"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="c1"># enable ignore corrupt files via the data source option</span><span class="w"> | 
|  | </span><span class="c1"># dir1/file3.json is corrupt from parquet's view</span><span class="w"> | 
|  | </span><span class="n">testCorruptDF0</span><span class="w"> </span><span class="o"><-</span><span class="w"> </span><span class="n">read.parquet</span><span class="p">(</span><span class="nf">c</span><span class="p">(</span><span class="s2">"examples/src/main/resources/dir1/"</span><span class="p">,</span><span class="w"> </span><span class="s2">"examples/src/main/resources/dir1/dir2/"</span><span class="p">),</span><span class="w"> </span><span class="n">ignoreCorruptFiles</span><span class="w"> </span><span class="o">=</span><span class="w"> </span><span class="s2">"true"</span><span class="p">)</span><span class="w"> | 
|  | </span><span class="n">head</span><span class="p">(</span><span class="n">testCorruptDF0</span><span class="p">)</span><span class="w"> | 
|  | </span><span class="c1">#            file</span><span class="w"> | 
|  | </span><span class="c1"># 1 file1.parquet</span><span class="w"> | 
|  | </span><span class="c1"># 2 file2.parquet</span><span class="w"> | 
|  |  | 
|  | </span><span class="c1"># enable ignore corrupt files via the configuration</span><span class="w"> | 
|  | </span><span class="n">sql</span><span class="p">(</span><span class="s2">"set spark.sql.files.ignoreCorruptFiles=true"</span><span class="p">)</span><span class="w"> | 
|  | </span><span class="c1"># dir1/file3.json is corrupt from parquet's view</span><span class="w"> | 
|  | </span><span class="n">testCorruptDF1</span><span class="w"> </span><span class="o"><-</span><span class="w"> </span><span class="n">read.parquet</span><span class="p">(</span><span class="nf">c</span><span class="p">(</span><span class="s2">"examples/src/main/resources/dir1/"</span><span class="p">,</span><span class="w"> </span><span class="s2">"examples/src/main/resources/dir1/dir2/"</span><span class="p">))</span><span class="w"> | 
|  | </span><span class="n">head</span><span class="p">(</span><span class="n">testCorruptDF1</span><span class="p">)</span><span class="w"> | 
|  | </span><span class="c1">#            file</span><span class="w"> | 
|  | </span><span class="c1"># 1 file1.parquet</span><span class="w"> | 
|  | </span><span class="c1"># 2 file2.parquet</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/r/RSparkSQLExample.R" in the Spark repo.</small></div> | 
|  | </div> | 
|  | </div> | 
|  |  | 
|  | <h3 id="ignore-missing-files">Ignore Missing Files</h3> | 
|  |  | 
|  | <p>Spark allows you to use the configuration <code class="language-plaintext highlighter-rouge">spark.sql.files.ignoreMissingFiles</code> or the data source option <code class="language-plaintext highlighter-rouge">ignoreMissingFiles</code> to ignore missing files while reading data | 
|  | from files. Here, missing file really means the deleted file under directory after you construct the | 
|  | <code class="language-plaintext highlighter-rouge">DataFrame</code>. When set to true, the Spark jobs will continue to run when encountering missing files and | 
|  | the contents that have been read will still be returned.</p> | 
|  |  | 
|  | <h3 id="path-glob-filter">Path Glob Filter</h3> | 
|  |  | 
|  | <p><code class="language-plaintext highlighter-rouge">pathGlobFilter</code> is used to only include files with file names matching the pattern. The syntax follows | 
|  | <code>org.apache.hadoop.fs.GlobFilter</code>. It does not change the behavior of partition discovery.</p> | 
|  |  | 
|  | <p>To load files with paths matching a given glob pattern while keeping the behavior of partition discovery, | 
|  | you can use:</p> | 
|  |  | 
|  | <div class="codetabs"> | 
|  |  | 
|  | <div data-lang="python"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="n">df</span> <span class="o">=</span> <span class="n">spark</span><span class="p">.</span><span class="n">read</span><span class="p">.</span><span class="n">load</span><span class="p">(</span><span class="s">"examples/src/main/resources/dir1"</span><span class="p">,</span> | 
|  | <span class="nb">format</span><span class="o">=</span><span class="s">"parquet"</span><span class="p">,</span> <span class="n">pathGlobFilter</span><span class="o">=</span><span class="s">"*.parquet"</span><span class="p">)</span> | 
|  | <span class="n">df</span><span class="p">.</span><span class="n">show</span><span class="p">()</span> | 
|  | <span class="c1"># +-------------+ | 
|  | # |         file| | 
|  | # +-------------+ | 
|  | # |file1.parquet| | 
|  | # +-------------+</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/python/sql/datasource.py" in the Spark repo.</small></div> | 
|  | </div> | 
|  |  | 
|  | <div data-lang="scala"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="k">val</span> <span class="nv">testGlobFilterDF</span> <span class="k">=</span> <span class="nv">spark</span><span class="o">.</span><span class="py">read</span><span class="o">.</span><span class="py">format</span><span class="o">(</span><span class="s">"parquet"</span><span class="o">)</span> | 
|  | <span class="o">.</span><span class="py">option</span><span class="o">(</span><span class="s">"pathGlobFilter"</span><span class="o">,</span> <span class="s">"*.parquet"</span><span class="o">)</span> <span class="c1">// json file should be filtered out</span> | 
|  | <span class="o">.</span><span class="py">load</span><span class="o">(</span><span class="s">"examples/src/main/resources/dir1"</span><span class="o">)</span> | 
|  | <span class="nv">testGlobFilterDF</span><span class="o">.</span><span class="py">show</span><span class="o">()</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |         file|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |file1.parquet|</span> | 
|  | <span class="c1">// +-------------+</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/scala/org/apache/spark/examples/sql/SQLDataSourceExample.scala" in the Spark repo.</small></div> | 
|  | </div> | 
|  |  | 
|  | <div data-lang="java"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="nc">Dataset</span><span class="o"><</span><span class="nc">Row</span><span class="o">></span> <span class="n">testGlobFilterDF</span> <span class="o">=</span> <span class="n">spark</span><span class="o">.</span><span class="na">read</span><span class="o">().</span><span class="na">format</span><span class="o">(</span><span class="s">"parquet"</span><span class="o">)</span> | 
|  | <span class="o">.</span><span class="na">option</span><span class="o">(</span><span class="s">"pathGlobFilter"</span><span class="o">,</span> <span class="s">"*.parquet"</span><span class="o">)</span> <span class="c1">// json file should be filtered out</span> | 
|  | <span class="o">.</span><span class="na">load</span><span class="o">(</span><span class="s">"examples/src/main/resources/dir1"</span><span class="o">);</span> | 
|  | <span class="n">testGlobFilterDF</span><span class="o">.</span><span class="na">show</span><span class="o">();</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |         file|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |file1.parquet|</span> | 
|  | <span class="c1">// +-------------+</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/java/org/apache/spark/examples/sql/JavaSQLDataSourceExample.java" in the Spark repo.</small></div> | 
|  | </div> | 
|  |  | 
|  | <div data-lang="r"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="n">df</span><span class="w"> </span><span class="o"><-</span><span class="w"> </span><span class="n">read.df</span><span class="p">(</span><span class="s2">"examples/src/main/resources/dir1"</span><span class="p">,</span><span class="w"> </span><span class="s2">"parquet"</span><span class="p">,</span><span class="w"> </span><span class="n">pathGlobFilter</span><span class="w"> </span><span class="o">=</span><span class="w"> </span><span class="s2">"*.parquet"</span><span class="p">)</span><span class="w"> | 
|  | </span><span class="c1">#            file</span><span class="w"> | 
|  | </span><span class="c1"># 1 file1.parquet</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/r/RSparkSQLExample.R" in the Spark repo.</small></div> | 
|  | </div> | 
|  | </div> | 
|  |  | 
|  | <h3 id="recursive-file-lookup">Recursive File Lookup</h3> | 
|  | <p><code class="language-plaintext highlighter-rouge">recursiveFileLookup</code> is used to recursively load files and it disables partition inferring. Its default value is <code class="language-plaintext highlighter-rouge">false</code>. | 
|  | If data source explicitly specifies the <code class="language-plaintext highlighter-rouge">partitionSpec</code> when <code class="language-plaintext highlighter-rouge">recursiveFileLookup</code> is true, exception will be thrown.</p> | 
|  |  | 
|  | <p>To load all files recursively, you can use:</p> | 
|  |  | 
|  | <div class="codetabs"> | 
|  |  | 
|  | <div data-lang="python"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="n">recursive_loaded_df</span> <span class="o">=</span> <span class="n">spark</span><span class="p">.</span><span class="n">read</span><span class="p">.</span><span class="nb">format</span><span class="p">(</span><span class="s">"parquet"</span><span class="p">)</span>\ | 
|  | <span class="p">.</span><span class="n">option</span><span class="p">(</span><span class="s">"recursiveFileLookup"</span><span class="p">,</span> <span class="s">"true"</span><span class="p">)</span>\ | 
|  | <span class="p">.</span><span class="n">load</span><span class="p">(</span><span class="s">"examples/src/main/resources/dir1"</span><span class="p">)</span> | 
|  | <span class="n">recursive_loaded_df</span><span class="p">.</span><span class="n">show</span><span class="p">()</span> | 
|  | <span class="c1"># +-------------+ | 
|  | # |         file| | 
|  | # +-------------+ | 
|  | # |file1.parquet| | 
|  | # |file2.parquet| | 
|  | # +-------------+</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/python/sql/datasource.py" in the Spark repo.</small></div> | 
|  | </div> | 
|  |  | 
|  | <div data-lang="scala"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="k">val</span> <span class="nv">recursiveLoadedDF</span> <span class="k">=</span> <span class="nv">spark</span><span class="o">.</span><span class="py">read</span><span class="o">.</span><span class="py">format</span><span class="o">(</span><span class="s">"parquet"</span><span class="o">)</span> | 
|  | <span class="o">.</span><span class="py">option</span><span class="o">(</span><span class="s">"recursiveFileLookup"</span><span class="o">,</span> <span class="s">"true"</span><span class="o">)</span> | 
|  | <span class="o">.</span><span class="py">load</span><span class="o">(</span><span class="s">"examples/src/main/resources/dir1"</span><span class="o">)</span> | 
|  | <span class="nv">recursiveLoadedDF</span><span class="o">.</span><span class="py">show</span><span class="o">()</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |         file|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |file1.parquet|</span> | 
|  | <span class="c1">// |file2.parquet|</span> | 
|  | <span class="c1">// +-------------+</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/scala/org/apache/spark/examples/sql/SQLDataSourceExample.scala" in the Spark repo.</small></div> | 
|  | </div> | 
|  |  | 
|  | <div data-lang="java"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="nc">Dataset</span><span class="o"><</span><span class="nc">Row</span><span class="o">></span> <span class="n">recursiveLoadedDF</span> <span class="o">=</span> <span class="n">spark</span><span class="o">.</span><span class="na">read</span><span class="o">().</span><span class="na">format</span><span class="o">(</span><span class="s">"parquet"</span><span class="o">)</span> | 
|  | <span class="o">.</span><span class="na">option</span><span class="o">(</span><span class="s">"recursiveFileLookup"</span><span class="o">,</span> <span class="s">"true"</span><span class="o">)</span> | 
|  | <span class="o">.</span><span class="na">load</span><span class="o">(</span><span class="s">"examples/src/main/resources/dir1"</span><span class="o">);</span> | 
|  | <span class="n">recursiveLoadedDF</span><span class="o">.</span><span class="na">show</span><span class="o">();</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |         file|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |file1.parquet|</span> | 
|  | <span class="c1">// |file2.parquet|</span> | 
|  | <span class="c1">// +-------------+</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/java/org/apache/spark/examples/sql/JavaSQLDataSourceExample.java" in the Spark repo.</small></div> | 
|  | </div> | 
|  |  | 
|  | <div data-lang="r"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="n">recursiveLoadedDF</span><span class="w"> </span><span class="o"><-</span><span class="w"> </span><span class="n">read.df</span><span class="p">(</span><span class="s2">"examples/src/main/resources/dir1"</span><span class="p">,</span><span class="w"> </span><span class="s2">"parquet"</span><span class="p">,</span><span class="w"> </span><span class="n">recursiveFileLookup</span><span class="w"> </span><span class="o">=</span><span class="w"> </span><span class="s2">"true"</span><span class="p">)</span><span class="w"> | 
|  | </span><span class="n">head</span><span class="p">(</span><span class="n">recursiveLoadedDF</span><span class="p">)</span><span class="w"> | 
|  | </span><span class="c1">#            file</span><span class="w"> | 
|  | </span><span class="c1"># 1 file1.parquet</span><span class="w"> | 
|  | </span><span class="c1"># 2 file2.parquet</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/r/RSparkSQLExample.R" in the Spark repo.</small></div> | 
|  | </div> | 
|  | </div> | 
|  |  | 
|  | <h3 id="modification-time-path-filters">Modification Time Path Filters</h3> | 
|  |  | 
|  | <p><code class="language-plaintext highlighter-rouge">modifiedBefore</code> and <code class="language-plaintext highlighter-rouge">modifiedAfter</code> are options that can be | 
|  | applied together or separately in order to achieve greater | 
|  | granularity over which files may load during a Spark batch query. | 
|  | (Note that Structured Streaming file sources don’t support these options.)</p> | 
|  |  | 
|  | <ul> | 
|  | <li><code class="language-plaintext highlighter-rouge">modifiedBefore</code>: an optional timestamp to only include files with | 
|  | modification times occurring before the specified time. The provided timestamp | 
|  | must be in the following format: YYYY-MM-DDTHH:mm:ss (e.g. 2020-06-01T13:00:00)</li> | 
|  | <li><code class="language-plaintext highlighter-rouge">modifiedAfter</code>: an optional timestamp to only include files with | 
|  | modification times occurring after the specified time. The provided timestamp | 
|  | must be in the following format: YYYY-MM-DDTHH:mm:ss (e.g. 2020-06-01T13:00:00)</li> | 
|  | </ul> | 
|  |  | 
|  | <p>When a timezone option is not provided, the timestamps will be interpreted according | 
|  | to the Spark session timezone (<code class="language-plaintext highlighter-rouge">spark.sql.session.timeZone</code>).</p> | 
|  |  | 
|  | <p>To load files with paths matching a given modified time range, you can use:</p> | 
|  |  | 
|  | <div class="codetabs"> | 
|  |  | 
|  | <div data-lang="python"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="c1"># Only load files modified before 07/1/2050 @ 08:30:00 | 
|  | </span><span class="n">df</span> <span class="o">=</span> <span class="n">spark</span><span class="p">.</span><span class="n">read</span><span class="p">.</span><span class="n">load</span><span class="p">(</span><span class="s">"examples/src/main/resources/dir1"</span><span class="p">,</span> | 
|  | <span class="nb">format</span><span class="o">=</span><span class="s">"parquet"</span><span class="p">,</span> <span class="n">modifiedBefore</span><span class="o">=</span><span class="s">"2050-07-01T08:30:00"</span><span class="p">)</span> | 
|  | <span class="n">df</span><span class="p">.</span><span class="n">show</span><span class="p">()</span> | 
|  | <span class="c1"># +-------------+ | 
|  | # |         file| | 
|  | # +-------------+ | 
|  | # |file1.parquet| | 
|  | # +-------------+ | 
|  | # Only load files modified after 06/01/2050 @ 08:30:00 | 
|  | </span><span class="n">df</span> <span class="o">=</span> <span class="n">spark</span><span class="p">.</span><span class="n">read</span><span class="p">.</span><span class="n">load</span><span class="p">(</span><span class="s">"examples/src/main/resources/dir1"</span><span class="p">,</span> | 
|  | <span class="nb">format</span><span class="o">=</span><span class="s">"parquet"</span><span class="p">,</span> <span class="n">modifiedAfter</span><span class="o">=</span><span class="s">"2050-06-01T08:30:00"</span><span class="p">)</span> | 
|  | <span class="n">df</span><span class="p">.</span><span class="n">show</span><span class="p">()</span> | 
|  | <span class="c1"># +-------------+ | 
|  | # |         file| | 
|  | # +-------------+ | 
|  | # +-------------+</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/python/sql/datasource.py" in the Spark repo.</small></div> | 
|  | </div> | 
|  |  | 
|  | <div data-lang="scala"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="k">val</span> <span class="nv">beforeFilterDF</span> <span class="k">=</span> <span class="nv">spark</span><span class="o">.</span><span class="py">read</span><span class="o">.</span><span class="py">format</span><span class="o">(</span><span class="s">"parquet"</span><span class="o">)</span> | 
|  | <span class="c1">// Files modified before 07/01/2020 at 05:30 are allowed</span> | 
|  | <span class="o">.</span><span class="py">option</span><span class="o">(</span><span class="s">"modifiedBefore"</span><span class="o">,</span> <span class="s">"2020-07-01T05:30:00"</span><span class="o">)</span> | 
|  | <span class="o">.</span><span class="py">load</span><span class="o">(</span><span class="s">"examples/src/main/resources/dir1"</span><span class="o">);</span> | 
|  | <span class="nv">beforeFilterDF</span><span class="o">.</span><span class="py">show</span><span class="o">();</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |         file|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |file1.parquet|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="k">val</span> <span class="nv">afterFilterDF</span> <span class="k">=</span> <span class="nv">spark</span><span class="o">.</span><span class="py">read</span><span class="o">.</span><span class="py">format</span><span class="o">(</span><span class="s">"parquet"</span><span class="o">)</span> | 
|  | <span class="c1">// Files modified after 06/01/2020 at 05:30 are allowed</span> | 
|  | <span class="o">.</span><span class="py">option</span><span class="o">(</span><span class="s">"modifiedAfter"</span><span class="o">,</span> <span class="s">"2020-06-01T05:30:00"</span><span class="o">)</span> | 
|  | <span class="o">.</span><span class="py">load</span><span class="o">(</span><span class="s">"examples/src/main/resources/dir1"</span><span class="o">);</span> | 
|  | <span class="nv">afterFilterDF</span><span class="o">.</span><span class="py">show</span><span class="o">();</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |         file|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// +-------------+</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/scala/org/apache/spark/examples/sql/SQLDataSourceExample.scala" in the Spark repo.</small></div> | 
|  | </div> | 
|  |  | 
|  | <div data-lang="java"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="nc">Dataset</span><span class="o"><</span><span class="nc">Row</span><span class="o">></span> <span class="n">beforeFilterDF</span> <span class="o">=</span> <span class="n">spark</span><span class="o">.</span><span class="na">read</span><span class="o">().</span><span class="na">format</span><span class="o">(</span><span class="s">"parquet"</span><span class="o">)</span> | 
|  | <span class="c1">// Only load files modified before 7/1/2020 at 05:30</span> | 
|  | <span class="o">.</span><span class="na">option</span><span class="o">(</span><span class="s">"modifiedBefore"</span><span class="o">,</span> <span class="s">"2020-07-01T05:30:00"</span><span class="o">)</span> | 
|  | <span class="c1">// Only load files modified after 6/1/2020 at 05:30</span> | 
|  | <span class="o">.</span><span class="na">option</span><span class="o">(</span><span class="s">"modifiedAfter"</span><span class="o">,</span> <span class="s">"2020-06-01T05:30:00"</span><span class="o">)</span> | 
|  | <span class="c1">// Interpret both times above relative to CST timezone</span> | 
|  | <span class="o">.</span><span class="na">option</span><span class="o">(</span><span class="s">"timeZone"</span><span class="o">,</span> <span class="s">"CST"</span><span class="o">)</span> | 
|  | <span class="o">.</span><span class="na">load</span><span class="o">(</span><span class="s">"examples/src/main/resources/dir1"</span><span class="o">);</span> | 
|  | <span class="n">beforeFilterDF</span><span class="o">.</span><span class="na">show</span><span class="o">();</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |         file|</span> | 
|  | <span class="c1">// +-------------+</span> | 
|  | <span class="c1">// |file1.parquet|</span> | 
|  | <span class="c1">// +-------------+</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/java/org/apache/spark/examples/sql/JavaSQLDataSourceExample.java" in the Spark repo.</small></div> | 
|  | </div> | 
|  |  | 
|  | <div data-lang="r"> | 
|  | <div class="highlight"><pre class="codehilite"><code><span class="n">beforeDF</span><span class="w"> </span><span class="o"><-</span><span class="w"> </span><span class="n">read.df</span><span class="p">(</span><span class="s2">"examples/src/main/resources/dir1"</span><span class="p">,</span><span class="w"> </span><span class="s2">"parquet"</span><span class="p">,</span><span class="w"> </span><span class="n">modifiedBefore</span><span class="o">=</span><span class="w"> </span><span class="s2">"2020-07-01T05:30:00"</span><span class="p">)</span><span class="w"> | 
|  | </span><span class="c1">#            file</span><span class="w"> | 
|  | </span><span class="c1"># 1 file1.parquet</span><span class="w"> | 
|  | </span><span class="n">afterDF</span><span class="w"> </span><span class="o"><-</span><span class="w"> </span><span class="n">read.df</span><span class="p">(</span><span class="s2">"examples/src/main/resources/dir1"</span><span class="p">,</span><span class="w"> </span><span class="s2">"parquet"</span><span class="p">,</span><span class="w"> </span><span class="n">modifiedAfter</span><span class="w"> </span><span class="o">=</span><span class="w"> </span><span class="s2">"2020-06-01T05:30:00"</span><span class="p">)</span><span class="w"> | 
|  | </span><span class="c1">#            file</span></code></pre></div> | 
|  | <div><small>Find full example code at "examples/src/main/r/RSparkSQLExample.R" in the Spark repo.</small></div> | 
|  | </div> | 
|  | </div> | 
|  |  | 
|  |  | 
|  | </div> | 
|  |  | 
|  | <!-- /container --> | 
|  | </div> | 
|  |  | 
|  | <script src="js/vendor/jquery-3.5.1.min.js"></script> | 
|  | <script src="js/vendor/bootstrap.bundle.min.js"></script> | 
|  |  | 
|  | <script src="js/vendor/anchor.min.js"></script> | 
|  | <script src="js/main.js"></script> | 
|  |  | 
|  | <script type="text/javascript" src="js/vendor/docsearch.min.js"></script> | 
|  | <script type="text/javascript"> | 
|  | // DocSearch is entirely free and automated. DocSearch is built in two parts: | 
|  | // 1. a crawler which we run on our own infrastructure every 24 hours. It follows every link | 
|  | //    in your website and extract content from every page it traverses. It then pushes this | 
|  | //    content to an Algolia index. | 
|  | // 2. a JavaScript snippet to be inserted in your website that will bind this Algolia index | 
|  | //    to your search input and display its results in a dropdown UI. If you want to find more | 
|  | //    details on how works DocSearch, check the docs of DocSearch. | 
|  | docsearch({ | 
|  | apiKey: 'd62f962a82bc9abb53471cb7b89da35e', | 
|  | appId: 'RAI69RXRSK', | 
|  | indexName: 'apache_spark', | 
|  | inputSelector: '#docsearch-input', | 
|  | enhancedSearchInput: true, | 
|  | algoliaOptions: { | 
|  | 'facetFilters': ["version:3.5.0"] | 
|  | }, | 
|  | debug: false // Set debug to true if you want to inspect the dropdown | 
|  | }); | 
|  |  | 
|  | </script> | 
|  |  | 
|  | <!-- MathJax Section --> | 
|  | <script type="text/x-mathjax-config"> | 
|  | MathJax.Hub.Config({ | 
|  | TeX: { equationNumbers: { autoNumber: "AMS" } } | 
|  | }); | 
|  | </script> | 
|  | <script> | 
|  | // Note that we load MathJax this way to work with local file (file://), HTTP and HTTPS. | 
|  | // We could use "//cdn.mathjax...", but that won't support "file://". | 
|  | (function(d, script) { | 
|  | script = d.createElement('script'); | 
|  | script.type = 'text/javascript'; | 
|  | script.async = true; | 
|  | script.onload = function(){ | 
|  | MathJax.Hub.Config({ | 
|  | tex2jax: { | 
|  | inlineMath: [ ["$", "$"], ["\\\\(","\\\\)"] ], | 
|  | displayMath: [ ["$$","$$"], ["\\[", "\\]"] ], | 
|  | processEscapes: true, | 
|  | skipTags: ['script', 'noscript', 'style', 'textarea', 'pre'] | 
|  | } | 
|  | }); | 
|  | }; | 
|  | script.src = ('https:' == document.location.protocol ? 'https://' : 'http://') + | 
|  | 'cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.1/MathJax.js' + | 
|  | '?config=TeX-AMS-MML_HTMLorMML'; | 
|  | d.getElementsByTagName('head')[0].appendChild(script); | 
|  | }(document)); | 
|  | </script> | 
|  | </body> | 
|  | </html> |