blob: fc5fda9860bdaf728b36202c3b7891915ce108bf [file] [log] [blame]
<!DOCTYPE html>
<html>
<head>
<meta charset="utf-8" /><meta name="generator" content="Docutils 0.17.1: http://docutils.sourceforge.net/" />
<title>Spark SQL &#8212; PySpark 3.2.1 documentation</title>
<link href="../_static/css/theme.css" rel="stylesheet">
<link href="../_static/css/index.ff1ffe594081f20da1ef19478df9384b.css" rel="stylesheet">
<link rel="stylesheet"
href="../_static/vendor/fontawesome/5.13.0/css/all.min.css">
<link rel="preload" as="font" type="font/woff2" crossorigin
href="../_static/vendor/fontawesome/5.13.0/webfonts/fa-solid-900.woff2">
<link rel="preload" as="font" type="font/woff2" crossorigin
href="../_static/vendor/fontawesome/5.13.0/webfonts/fa-brands-400.woff2">
<link rel="stylesheet" href="../_static/css/blank.css" type="text/css" />
<link rel="stylesheet" href="../_static/pygments.css" type="text/css" />
<link rel="stylesheet" type="text/css" href="../_static/css/pyspark.css" />
<link rel="preload" as="script" href="../_static/js/index.be7d3bbb2ef33a8344ce.js">
<script id="documentation_options" data-url_root="../" src="../_static/documentation_options.js"></script>
<script src="../_static/jquery.js"></script>
<script src="../_static/underscore.js"></script>
<script src="../_static/doctools.js"></script>
<script src="../_static/language_data.js"></script>
<script src="../_static/copybutton.js"></script>
<script crossorigin="anonymous" integrity="sha256-Ae2Vz/4ePdIu6ZyI/5ZGsYnb+m0JlOmKPjt6XZ9JJkA=" src="https://cdnjs.cloudflare.com/ajax/libs/require.js/2.3.4/require.min.js"></script>
<script async="async" src="https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.5/latest.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/x-mathjax-config">MathJax.Hub.Config({"tex2jax": {"inlineMath": [["$", "$"], ["\\(", "\\)"]], "processEscapes": true, "ignoreClass": "tex2jax_ignore|mathjax_ignore|document", "processClass": "tex2jax_process|mathjax_process|math|output_area"}})</script>
<link rel="canonical" href="https://spark.apache.org/docs/latest/api/python/reference/pyspark.sql/index.html" />
<link rel="search" title="Search" href="../search.html" />
<link rel="next" title="pyspark.sql.SparkSession" href="api/pyspark.sql.SparkSession.html" />
<link rel="prev" title="API Reference" href="index.html" />
<meta name="viewport" content="width=device-width, initial-scale=1" />
<meta name="docsearch:language" content="None">
<!-- Google Analytics -->
</head>
<body data-spy="scroll" data-target="#bd-toc-nav" data-offset="80">
<div class="container-fluid" id="banner"></div>
<nav class="navbar navbar-light navbar-expand-lg bg-light fixed-top bd-navbar" id="navbar-main"><div class="container-xl">
<div id="navbar-start">
<a class="navbar-brand" href="../index.html">
<img src="../_static/spark-logo-reverse.png" class="logo" alt="logo">
</a>
</div>
<button class="navbar-toggler" type="button" data-toggle="collapse" data-target="#navbar-collapsible" aria-controls="navbar-collapsible" aria-expanded="false" aria-label="Toggle navigation">
<span class="navbar-toggler-icon"></span>
</button>
<div id="navbar-collapsible" class="col-lg-9 collapse navbar-collapse">
<div id="navbar-center" class="mr-auto">
<div class="navbar-center-item">
<ul id="navbar-main-elements" class="navbar-nav">
<li class="toctree-l1 nav-item">
<a class="reference internal nav-link" href="../getting_started/index.html">
Getting Started
</a>
</li>
<li class="toctree-l1 nav-item">
<a class="reference internal nav-link" href="../user_guide/index.html">
User Guide
</a>
</li>
<li class="toctree-l1 current active nav-item">
<a class="reference internal nav-link" href="index.html">
API Reference
</a>
</li>
<li class="toctree-l1 nav-item">
<a class="reference internal nav-link" href="../development/index.html">
Development
</a>
</li>
<li class="toctree-l1 nav-item">
<a class="reference internal nav-link" href="../migration_guide/index.html">
Migration Guide
</a>
</li>
</ul>
</div>
</div>
<div id="navbar-end">
<div class="navbar-end-item">
<ul id="navbar-icon-links" class="navbar-nav" aria-label="Icon Links">
</ul>
</div>
</div>
</div>
</div>
</nav>
<div class="container-xl">
<div class="row">
<!-- Only show if we have sidebars configured, else just a small margin -->
<div class="col-12 col-md-3 bd-sidebar"><form class="bd-search d-flex align-items-center" action="../search.html" method="get">
<i class="icon fas fa-search"></i>
<input type="search" class="form-control" name="q" id="search-input" placeholder="Search the docs ..." aria-label="Search the docs ..." autocomplete="off" >
</form><nav class="bd-links" id="bd-docs-nav" aria-label="Main navigation">
<div class="bd-toc-item active">
<ul class="current nav bd-sidenav">
<li class="toctree-l1 current active has-children">
<a class="current reference internal" href="#">
Spark SQL
</a>
<input checked="" class="toctree-checkbox" id="toctree-checkbox-1" name="toctree-checkbox-1" type="checkbox"/>
<label for="toctree-checkbox-1">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.html">
pyspark.sql.SparkSession
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.html">
pyspark.sql.Catalog
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.html">
pyspark.sql.DataFrame
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.html">
pyspark.sql.Column
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Row.html">
pyspark.sql.Row
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.GroupedData.html">
pyspark.sql.GroupedData
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.PandasCogroupedOps.html">
pyspark.sql.PandasCogroupedOps
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameNaFunctions.html">
pyspark.sql.DataFrameNaFunctions
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.html">
pyspark.sql.DataFrameStatFunctions
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Window.html">
pyspark.sql.Window
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.builder.appName.html">
pyspark.sql.SparkSession.builder.appName
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.builder.config.html">
pyspark.sql.SparkSession.builder.config
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.builder.enableHiveSupport.html">
pyspark.sql.SparkSession.builder.enableHiveSupport
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.builder.getOrCreate.html">
pyspark.sql.SparkSession.builder.getOrCreate
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.builder.master.html">
pyspark.sql.SparkSession.builder.master
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.catalog.html">
pyspark.sql.SparkSession.catalog
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.conf.html">
pyspark.sql.SparkSession.conf
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.createDataFrame.html">
pyspark.sql.SparkSession.createDataFrame
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.getActiveSession.html">
pyspark.sql.SparkSession.getActiveSession
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.newSession.html">
pyspark.sql.SparkSession.newSession
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.range.html">
pyspark.sql.SparkSession.range
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.read.html">
pyspark.sql.SparkSession.read
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.readStream.html">
pyspark.sql.SparkSession.readStream
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.sparkContext.html">
pyspark.sql.SparkSession.sparkContext
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.sql.html">
pyspark.sql.SparkSession.sql
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.stop.html">
pyspark.sql.SparkSession.stop
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.streams.html">
pyspark.sql.SparkSession.streams
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.table.html">
pyspark.sql.SparkSession.table
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.udf.html">
pyspark.sql.SparkSession.udf
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.SparkSession.version.html">
pyspark.sql.SparkSession.version
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.conf.RuntimeConfig.html">
pyspark.sql.conf.RuntimeConfig
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameReader.csv.html">
pyspark.sql.DataFrameReader.csv
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameReader.format.html">
pyspark.sql.DataFrameReader.format
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameReader.jdbc.html">
pyspark.sql.DataFrameReader.jdbc
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameReader.json.html">
pyspark.sql.DataFrameReader.json
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameReader.load.html">
pyspark.sql.DataFrameReader.load
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameReader.option.html">
pyspark.sql.DataFrameReader.option
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameReader.options.html">
pyspark.sql.DataFrameReader.options
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameReader.orc.html">
pyspark.sql.DataFrameReader.orc
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameReader.parquet.html">
pyspark.sql.DataFrameReader.parquet
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameReader.schema.html">
pyspark.sql.DataFrameReader.schema
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameReader.table.html">
pyspark.sql.DataFrameReader.table
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.bucketBy.html">
pyspark.sql.DataFrameWriter.bucketBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.csv.html">
pyspark.sql.DataFrameWriter.csv
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.format.html">
pyspark.sql.DataFrameWriter.format
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.insertInto.html">
pyspark.sql.DataFrameWriter.insertInto
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.jdbc.html">
pyspark.sql.DataFrameWriter.jdbc
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.json.html">
pyspark.sql.DataFrameWriter.json
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.mode.html">
pyspark.sql.DataFrameWriter.mode
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.option.html">
pyspark.sql.DataFrameWriter.option
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.options.html">
pyspark.sql.DataFrameWriter.options
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.orc.html">
pyspark.sql.DataFrameWriter.orc
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.parquet.html">
pyspark.sql.DataFrameWriter.parquet
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.partitionBy.html">
pyspark.sql.DataFrameWriter.partitionBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.save.html">
pyspark.sql.DataFrameWriter.save
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.saveAsTable.html">
pyspark.sql.DataFrameWriter.saveAsTable
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.sortBy.html">
pyspark.sql.DataFrameWriter.sortBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameWriter.text.html">
pyspark.sql.DataFrameWriter.text
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.agg.html">
pyspark.sql.DataFrame.agg
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.alias.html">
pyspark.sql.DataFrame.alias
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.approxQuantile.html">
pyspark.sql.DataFrame.approxQuantile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.cache.html">
pyspark.sql.DataFrame.cache
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.checkpoint.html">
pyspark.sql.DataFrame.checkpoint
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.coalesce.html">
pyspark.sql.DataFrame.coalesce
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.colRegex.html">
pyspark.sql.DataFrame.colRegex
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.collect.html">
pyspark.sql.DataFrame.collect
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.columns.html">
pyspark.sql.DataFrame.columns
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.corr.html">
pyspark.sql.DataFrame.corr
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.count.html">
pyspark.sql.DataFrame.count
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.cov.html">
pyspark.sql.DataFrame.cov
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.createGlobalTempView.html">
pyspark.sql.DataFrame.createGlobalTempView
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.createOrReplaceGlobalTempView.html">
pyspark.sql.DataFrame.createOrReplaceGlobalTempView
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.createOrReplaceTempView.html">
pyspark.sql.DataFrame.createOrReplaceTempView
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.createTempView.html">
pyspark.sql.DataFrame.createTempView
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.crossJoin.html">
pyspark.sql.DataFrame.crossJoin
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.crosstab.html">
pyspark.sql.DataFrame.crosstab
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.cube.html">
pyspark.sql.DataFrame.cube
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.describe.html">
pyspark.sql.DataFrame.describe
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.distinct.html">
pyspark.sql.DataFrame.distinct
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.drop.html">
pyspark.sql.DataFrame.drop
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.dropDuplicates.html">
pyspark.sql.DataFrame.dropDuplicates
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.drop_duplicates.html">
pyspark.sql.DataFrame.drop_duplicates
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.dropna.html">
pyspark.sql.DataFrame.dropna
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.dtypes.html">
pyspark.sql.DataFrame.dtypes
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.exceptAll.html">
pyspark.sql.DataFrame.exceptAll
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.explain.html">
pyspark.sql.DataFrame.explain
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.fillna.html">
pyspark.sql.DataFrame.fillna
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.filter.html">
pyspark.sql.DataFrame.filter
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.first.html">
pyspark.sql.DataFrame.first
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.foreach.html">
pyspark.sql.DataFrame.foreach
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.foreachPartition.html">
pyspark.sql.DataFrame.foreachPartition
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.freqItems.html">
pyspark.sql.DataFrame.freqItems
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.groupBy.html">
pyspark.sql.DataFrame.groupBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.head.html">
pyspark.sql.DataFrame.head
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.hint.html">
pyspark.sql.DataFrame.hint
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.inputFiles.html">
pyspark.sql.DataFrame.inputFiles
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.intersect.html">
pyspark.sql.DataFrame.intersect
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.intersectAll.html">
pyspark.sql.DataFrame.intersectAll
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.isLocal.html">
pyspark.sql.DataFrame.isLocal
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.isStreaming.html">
pyspark.sql.DataFrame.isStreaming
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.join.html">
pyspark.sql.DataFrame.join
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.limit.html">
pyspark.sql.DataFrame.limit
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.localCheckpoint.html">
pyspark.sql.DataFrame.localCheckpoint
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.mapInPandas.html">
pyspark.sql.DataFrame.mapInPandas
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.na.html">
pyspark.sql.DataFrame.na
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.orderBy.html">
pyspark.sql.DataFrame.orderBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.persist.html">
pyspark.sql.DataFrame.persist
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.printSchema.html">
pyspark.sql.DataFrame.printSchema
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.randomSplit.html">
pyspark.sql.DataFrame.randomSplit
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.rdd.html">
pyspark.sql.DataFrame.rdd
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.registerTempTable.html">
pyspark.sql.DataFrame.registerTempTable
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.repartition.html">
pyspark.sql.DataFrame.repartition
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.repartitionByRange.html">
pyspark.sql.DataFrame.repartitionByRange
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.replace.html">
pyspark.sql.DataFrame.replace
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.rollup.html">
pyspark.sql.DataFrame.rollup
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.sameSemantics.html">
pyspark.sql.DataFrame.sameSemantics
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.sample.html">
pyspark.sql.DataFrame.sample
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.sampleBy.html">
pyspark.sql.DataFrame.sampleBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.schema.html">
pyspark.sql.DataFrame.schema
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.select.html">
pyspark.sql.DataFrame.select
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.selectExpr.html">
pyspark.sql.DataFrame.selectExpr
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.semanticHash.html">
pyspark.sql.DataFrame.semanticHash
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.show.html">
pyspark.sql.DataFrame.show
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.sort.html">
pyspark.sql.DataFrame.sort
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.sortWithinPartitions.html">
pyspark.sql.DataFrame.sortWithinPartitions
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.stat.html">
pyspark.sql.DataFrame.stat
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.storageLevel.html">
pyspark.sql.DataFrame.storageLevel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.subtract.html">
pyspark.sql.DataFrame.subtract
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.summary.html">
pyspark.sql.DataFrame.summary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.tail.html">
pyspark.sql.DataFrame.tail
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.take.html">
pyspark.sql.DataFrame.take
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.toDF.html">
pyspark.sql.DataFrame.toDF
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.toJSON.html">
pyspark.sql.DataFrame.toJSON
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.toLocalIterator.html">
pyspark.sql.DataFrame.toLocalIterator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.toPandas.html">
pyspark.sql.DataFrame.toPandas
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.transform.html">
pyspark.sql.DataFrame.transform
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.union.html">
pyspark.sql.DataFrame.union
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.unionAll.html">
pyspark.sql.DataFrame.unionAll
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.unionByName.html">
pyspark.sql.DataFrame.unionByName
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.unpersist.html">
pyspark.sql.DataFrame.unpersist
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.where.html">
pyspark.sql.DataFrame.where
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.withColumn.html">
pyspark.sql.DataFrame.withColumn
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.withColumnRenamed.html">
pyspark.sql.DataFrame.withColumnRenamed
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.withWatermark.html">
pyspark.sql.DataFrame.withWatermark
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.write.html">
pyspark.sql.DataFrame.write
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.writeStream.html">
pyspark.sql.DataFrame.writeStream
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.writeTo.html">
pyspark.sql.DataFrame.writeTo
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrame.to_pandas_on_spark.html">
pyspark.sql.DataFrame.to_pandas_on_spark
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameNaFunctions.drop.html">
pyspark.sql.DataFrameNaFunctions.drop
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameNaFunctions.fill.html">
pyspark.sql.DataFrameNaFunctions.fill
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameNaFunctions.replace.html">
pyspark.sql.DataFrameNaFunctions.replace
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.approxQuantile.html">
pyspark.sql.DataFrameStatFunctions.approxQuantile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.corr.html">
pyspark.sql.DataFrameStatFunctions.corr
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.cov.html">
pyspark.sql.DataFrameStatFunctions.cov
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.crosstab.html">
pyspark.sql.DataFrameStatFunctions.crosstab
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.freqItems.html">
pyspark.sql.DataFrameStatFunctions.freqItems
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.sampleBy.html">
pyspark.sql.DataFrameStatFunctions.sampleBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.alias.html">
pyspark.sql.Column.alias
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.asc.html">
pyspark.sql.Column.asc
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.asc_nulls_first.html">
pyspark.sql.Column.asc_nulls_first
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.asc_nulls_last.html">
pyspark.sql.Column.asc_nulls_last
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.astype.html">
pyspark.sql.Column.astype
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.between.html">
pyspark.sql.Column.between
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.bitwiseAND.html">
pyspark.sql.Column.bitwiseAND
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.bitwiseOR.html">
pyspark.sql.Column.bitwiseOR
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.bitwiseXOR.html">
pyspark.sql.Column.bitwiseXOR
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.cast.html">
pyspark.sql.Column.cast
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.contains.html">
pyspark.sql.Column.contains
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.desc.html">
pyspark.sql.Column.desc
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.desc_nulls_first.html">
pyspark.sql.Column.desc_nulls_first
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.desc_nulls_last.html">
pyspark.sql.Column.desc_nulls_last
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.dropFields.html">
pyspark.sql.Column.dropFields
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.endswith.html">
pyspark.sql.Column.endswith
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.eqNullSafe.html">
pyspark.sql.Column.eqNullSafe
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.getField.html">
pyspark.sql.Column.getField
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.getItem.html">
pyspark.sql.Column.getItem
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.isNotNull.html">
pyspark.sql.Column.isNotNull
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.isNull.html">
pyspark.sql.Column.isNull
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.isin.html">
pyspark.sql.Column.isin
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.like.html">
pyspark.sql.Column.like
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.name.html">
pyspark.sql.Column.name
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.otherwise.html">
pyspark.sql.Column.otherwise
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.over.html">
pyspark.sql.Column.over
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.rlike.html">
pyspark.sql.Column.rlike
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.startswith.html">
pyspark.sql.Column.startswith
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.substr.html">
pyspark.sql.Column.substr
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.when.html">
pyspark.sql.Column.when
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Column.withField.html">
pyspark.sql.Column.withField
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.ArrayType.html">
ArrayType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.BinaryType.html">
BinaryType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.BooleanType.html">
BooleanType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.ByteType.html">
ByteType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.DataType.html">
DataType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.DateType.html">
DateType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.DecimalType.html">
DecimalType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.DoubleType.html">
DoubleType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.FloatType.html">
FloatType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.IntegerType.html">
IntegerType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.LongType.html">
LongType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.MapType.html">
MapType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.NullType.html">
NullType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.ShortType.html">
ShortType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.StringType.html">
StringType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.StructField.html">
StructField
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.StructType.html">
StructType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.types.TimestampType.html">
TimestampType
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Row.asDict.html">
pyspark.sql.Row.asDict
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.abs.html">
pyspark.sql.functions.abs
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.acos.html">
pyspark.sql.functions.acos
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.acosh.html">
pyspark.sql.functions.acosh
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.add_months.html">
pyspark.sql.functions.add_months
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.aggregate.html">
pyspark.sql.functions.aggregate
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.approxCountDistinct.html">
pyspark.sql.functions.approxCountDistinct
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.approx_count_distinct.html">
pyspark.sql.functions.approx_count_distinct
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.array.html">
pyspark.sql.functions.array
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.array_contains.html">
pyspark.sql.functions.array_contains
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.array_distinct.html">
pyspark.sql.functions.array_distinct
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.array_except.html">
pyspark.sql.functions.array_except
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.array_intersect.html">
pyspark.sql.functions.array_intersect
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.array_join.html">
pyspark.sql.functions.array_join
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.array_max.html">
pyspark.sql.functions.array_max
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.array_min.html">
pyspark.sql.functions.array_min
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.array_position.html">
pyspark.sql.functions.array_position
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.array_remove.html">
pyspark.sql.functions.array_remove
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.array_repeat.html">
pyspark.sql.functions.array_repeat
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.array_sort.html">
pyspark.sql.functions.array_sort
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.array_union.html">
pyspark.sql.functions.array_union
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.arrays_overlap.html">
pyspark.sql.functions.arrays_overlap
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.arrays_zip.html">
pyspark.sql.functions.arrays_zip
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.asc.html">
pyspark.sql.functions.asc
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.asc_nulls_first.html">
pyspark.sql.functions.asc_nulls_first
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.asc_nulls_last.html">
pyspark.sql.functions.asc_nulls_last
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.ascii.html">
pyspark.sql.functions.ascii
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.asin.html">
pyspark.sql.functions.asin
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.asinh.html">
pyspark.sql.functions.asinh
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.assert_true.html">
pyspark.sql.functions.assert_true
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.atan.html">
pyspark.sql.functions.atan
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.atanh.html">
pyspark.sql.functions.atanh
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.atan2.html">
pyspark.sql.functions.atan2
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.avg.html">
pyspark.sql.functions.avg
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.base64.html">
pyspark.sql.functions.base64
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.bin.html">
pyspark.sql.functions.bin
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.bitwise_not.html">
pyspark.sql.functions.bitwise_not
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.bitwiseNOT.html">
pyspark.sql.functions.bitwiseNOT
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.broadcast.html">
pyspark.sql.functions.broadcast
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.bround.html">
pyspark.sql.functions.bround
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.bucket.html">
pyspark.sql.functions.bucket
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.cbrt.html">
pyspark.sql.functions.cbrt
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.ceil.html">
pyspark.sql.functions.ceil
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.coalesce.html">
pyspark.sql.functions.coalesce
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.col.html">
pyspark.sql.functions.col
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.collect_list.html">
pyspark.sql.functions.collect_list
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.collect_set.html">
pyspark.sql.functions.collect_set
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.column.html">
pyspark.sql.functions.column
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.concat.html">
pyspark.sql.functions.concat
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.concat_ws.html">
pyspark.sql.functions.concat_ws
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.conv.html">
pyspark.sql.functions.conv
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.corr.html">
pyspark.sql.functions.corr
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.cos.html">
pyspark.sql.functions.cos
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.cosh.html">
pyspark.sql.functions.cosh
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.count.html">
pyspark.sql.functions.count
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.count_distinct.html">
pyspark.sql.functions.count_distinct
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.countDistinct.html">
pyspark.sql.functions.countDistinct
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.covar_pop.html">
pyspark.sql.functions.covar_pop
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.covar_samp.html">
pyspark.sql.functions.covar_samp
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.crc32.html">
pyspark.sql.functions.crc32
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.create_map.html">
pyspark.sql.functions.create_map
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.cume_dist.html">
pyspark.sql.functions.cume_dist
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.current_date.html">
pyspark.sql.functions.current_date
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.current_timestamp.html">
pyspark.sql.functions.current_timestamp
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.date_add.html">
pyspark.sql.functions.date_add
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.date_format.html">
pyspark.sql.functions.date_format
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.date_sub.html">
pyspark.sql.functions.date_sub
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.date_trunc.html">
pyspark.sql.functions.date_trunc
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.datediff.html">
pyspark.sql.functions.datediff
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.dayofmonth.html">
pyspark.sql.functions.dayofmonth
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.dayofweek.html">
pyspark.sql.functions.dayofweek
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.dayofyear.html">
pyspark.sql.functions.dayofyear
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.days.html">
pyspark.sql.functions.days
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.decode.html">
pyspark.sql.functions.decode
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.degrees.html">
pyspark.sql.functions.degrees
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.dense_rank.html">
pyspark.sql.functions.dense_rank
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.desc.html">
pyspark.sql.functions.desc
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.desc_nulls_first.html">
pyspark.sql.functions.desc_nulls_first
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.desc_nulls_last.html">
pyspark.sql.functions.desc_nulls_last
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.element_at.html">
pyspark.sql.functions.element_at
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.encode.html">
pyspark.sql.functions.encode
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.exists.html">
pyspark.sql.functions.exists
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.exp.html">
pyspark.sql.functions.exp
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.explode.html">
pyspark.sql.functions.explode
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.explode_outer.html">
pyspark.sql.functions.explode_outer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.expm1.html">
pyspark.sql.functions.expm1
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.expr.html">
pyspark.sql.functions.expr
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.factorial.html">
pyspark.sql.functions.factorial
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.filter.html">
pyspark.sql.functions.filter
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.first.html">
pyspark.sql.functions.first
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.flatten.html">
pyspark.sql.functions.flatten
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.floor.html">
pyspark.sql.functions.floor
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.forall.html">
pyspark.sql.functions.forall
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.format_number.html">
pyspark.sql.functions.format_number
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.format_string.html">
pyspark.sql.functions.format_string
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.from_csv.html">
pyspark.sql.functions.from_csv
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.from_json.html">
pyspark.sql.functions.from_json
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.from_unixtime.html">
pyspark.sql.functions.from_unixtime
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.from_utc_timestamp.html">
pyspark.sql.functions.from_utc_timestamp
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.get_json_object.html">
pyspark.sql.functions.get_json_object
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.greatest.html">
pyspark.sql.functions.greatest
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.grouping.html">
pyspark.sql.functions.grouping
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.grouping_id.html">
pyspark.sql.functions.grouping_id
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.hash.html">
pyspark.sql.functions.hash
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.hex.html">
pyspark.sql.functions.hex
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.hour.html">
pyspark.sql.functions.hour
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.hours.html">
pyspark.sql.functions.hours
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.hypot.html">
pyspark.sql.functions.hypot
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.initcap.html">
pyspark.sql.functions.initcap
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.input_file_name.html">
pyspark.sql.functions.input_file_name
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.instr.html">
pyspark.sql.functions.instr
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.isnan.html">
pyspark.sql.functions.isnan
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.isnull.html">
pyspark.sql.functions.isnull
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.json_tuple.html">
pyspark.sql.functions.json_tuple
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.kurtosis.html">
pyspark.sql.functions.kurtosis
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.lag.html">
pyspark.sql.functions.lag
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.last.html">
pyspark.sql.functions.last
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.last_day.html">
pyspark.sql.functions.last_day
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.lead.html">
pyspark.sql.functions.lead
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.least.html">
pyspark.sql.functions.least
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.length.html">
pyspark.sql.functions.length
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.levenshtein.html">
pyspark.sql.functions.levenshtein
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.lit.html">
pyspark.sql.functions.lit
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.locate.html">
pyspark.sql.functions.locate
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.log.html">
pyspark.sql.functions.log
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.log10.html">
pyspark.sql.functions.log10
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.log1p.html">
pyspark.sql.functions.log1p
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.log2.html">
pyspark.sql.functions.log2
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.lower.html">
pyspark.sql.functions.lower
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.lpad.html">
pyspark.sql.functions.lpad
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.ltrim.html">
pyspark.sql.functions.ltrim
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.map_concat.html">
pyspark.sql.functions.map_concat
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.map_entries.html">
pyspark.sql.functions.map_entries
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.map_filter.html">
pyspark.sql.functions.map_filter
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.map_from_arrays.html">
pyspark.sql.functions.map_from_arrays
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.map_from_entries.html">
pyspark.sql.functions.map_from_entries
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.map_keys.html">
pyspark.sql.functions.map_keys
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.map_values.html">
pyspark.sql.functions.map_values
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.map_zip_with.html">
pyspark.sql.functions.map_zip_with
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.max.html">
pyspark.sql.functions.max
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.md5.html">
pyspark.sql.functions.md5
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.mean.html">
pyspark.sql.functions.mean
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.min.html">
pyspark.sql.functions.min
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.minute.html">
pyspark.sql.functions.minute
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.monotonically_increasing_id.html">
pyspark.sql.functions.monotonically_increasing_id
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.month.html">
pyspark.sql.functions.month
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.months.html">
pyspark.sql.functions.months
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.months_between.html">
pyspark.sql.functions.months_between
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.nanvl.html">
pyspark.sql.functions.nanvl
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.next_day.html">
pyspark.sql.functions.next_day
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.nth_value.html">
pyspark.sql.functions.nth_value
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.ntile.html">
pyspark.sql.functions.ntile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.overlay.html">
pyspark.sql.functions.overlay
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.pandas_udf.html">
pyspark.sql.functions.pandas_udf
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.percent_rank.html">
pyspark.sql.functions.percent_rank
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.percentile_approx.html">
pyspark.sql.functions.percentile_approx
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.posexplode.html">
pyspark.sql.functions.posexplode
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.posexplode_outer.html">
pyspark.sql.functions.posexplode_outer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.pow.html">
pyspark.sql.functions.pow
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.product.html">
pyspark.sql.functions.product
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.quarter.html">
pyspark.sql.functions.quarter
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.radians.html">
pyspark.sql.functions.radians
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.raise_error.html">
pyspark.sql.functions.raise_error
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.rand.html">
pyspark.sql.functions.rand
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.randn.html">
pyspark.sql.functions.randn
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.rank.html">
pyspark.sql.functions.rank
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.regexp_extract.html">
pyspark.sql.functions.regexp_extract
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.regexp_replace.html">
pyspark.sql.functions.regexp_replace
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.repeat.html">
pyspark.sql.functions.repeat
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.reverse.html">
pyspark.sql.functions.reverse
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.rint.html">
pyspark.sql.functions.rint
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.round.html">
pyspark.sql.functions.round
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.row_number.html">
pyspark.sql.functions.row_number
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.rpad.html">
pyspark.sql.functions.rpad
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.rtrim.html">
pyspark.sql.functions.rtrim
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.schema_of_csv.html">
pyspark.sql.functions.schema_of_csv
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.schema_of_json.html">
pyspark.sql.functions.schema_of_json
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.second.html">
pyspark.sql.functions.second
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.sentences.html">
pyspark.sql.functions.sentences
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.sequence.html">
pyspark.sql.functions.sequence
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.session_window.html">
pyspark.sql.functions.session_window
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.sha1.html">
pyspark.sql.functions.sha1
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.sha2.html">
pyspark.sql.functions.sha2
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.shiftleft.html">
pyspark.sql.functions.shiftleft
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.shiftright.html">
pyspark.sql.functions.shiftright
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.shiftrightunsigned.html">
pyspark.sql.functions.shiftrightunsigned
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.shuffle.html">
pyspark.sql.functions.shuffle
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.signum.html">
pyspark.sql.functions.signum
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.sin.html">
pyspark.sql.functions.sin
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.sinh.html">
pyspark.sql.functions.sinh
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.size.html">
pyspark.sql.functions.size
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.skewness.html">
pyspark.sql.functions.skewness
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.slice.html">
pyspark.sql.functions.slice
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.sort_array.html">
pyspark.sql.functions.sort_array
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.soundex.html">
pyspark.sql.functions.soundex
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.spark_partition_id.html">
pyspark.sql.functions.spark_partition_id
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.split.html">
pyspark.sql.functions.split
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.sqrt.html">
pyspark.sql.functions.sqrt
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.stddev.html">
pyspark.sql.functions.stddev
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.stddev_pop.html">
pyspark.sql.functions.stddev_pop
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.stddev_samp.html">
pyspark.sql.functions.stddev_samp
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.struct.html">
pyspark.sql.functions.struct
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.substring.html">
pyspark.sql.functions.substring
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.substring_index.html">
pyspark.sql.functions.substring_index
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.sum.html">
pyspark.sql.functions.sum
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.sum_distinct.html">
pyspark.sql.functions.sum_distinct
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.sumDistinct.html">
pyspark.sql.functions.sumDistinct
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.tan.html">
pyspark.sql.functions.tan
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.tanh.html">
pyspark.sql.functions.tanh
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.timestamp_seconds.html">
pyspark.sql.functions.timestamp_seconds
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.toDegrees.html">
pyspark.sql.functions.toDegrees
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.toRadians.html">
pyspark.sql.functions.toRadians
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.to_csv.html">
pyspark.sql.functions.to_csv
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.to_date.html">
pyspark.sql.functions.to_date
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.to_json.html">
pyspark.sql.functions.to_json
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.to_timestamp.html">
pyspark.sql.functions.to_timestamp
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.to_utc_timestamp.html">
pyspark.sql.functions.to_utc_timestamp
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.transform.html">
pyspark.sql.functions.transform
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.transform_keys.html">
pyspark.sql.functions.transform_keys
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.transform_values.html">
pyspark.sql.functions.transform_values
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.translate.html">
pyspark.sql.functions.translate
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.trim.html">
pyspark.sql.functions.trim
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.trunc.html">
pyspark.sql.functions.trunc
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.udf.html">
pyspark.sql.functions.udf
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.unbase64.html">
pyspark.sql.functions.unbase64
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.unhex.html">
pyspark.sql.functions.unhex
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.unix_timestamp.html">
pyspark.sql.functions.unix_timestamp
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.upper.html">
pyspark.sql.functions.upper
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.var_pop.html">
pyspark.sql.functions.var_pop
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.var_samp.html">
pyspark.sql.functions.var_samp
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.variance.html">
pyspark.sql.functions.variance
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.weekofyear.html">
pyspark.sql.functions.weekofyear
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.when.html">
pyspark.sql.functions.when
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.window.html">
pyspark.sql.functions.window
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.xxhash64.html">
pyspark.sql.functions.xxhash64
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.year.html">
pyspark.sql.functions.year
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.years.html">
pyspark.sql.functions.years
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.functions.zip_with.html">
pyspark.sql.functions.zip_with
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.avro.functions.from_avro.html">
pyspark.sql.avro.functions.from_avro
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.avro.functions.to_avro.html">
pyspark.sql.avro.functions.to_avro
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Window.currentRow.html">
pyspark.sql.Window.currentRow
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Window.orderBy.html">
pyspark.sql.Window.orderBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Window.partitionBy.html">
pyspark.sql.Window.partitionBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Window.rangeBetween.html">
pyspark.sql.Window.rangeBetween
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Window.rowsBetween.html">
pyspark.sql.Window.rowsBetween
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Window.unboundedFollowing.html">
pyspark.sql.Window.unboundedFollowing
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Window.unboundedPreceding.html">
pyspark.sql.Window.unboundedPreceding
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.WindowSpec.orderBy.html">
pyspark.sql.WindowSpec.orderBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.WindowSpec.partitionBy.html">
pyspark.sql.WindowSpec.partitionBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.WindowSpec.rangeBetween.html">
pyspark.sql.WindowSpec.rangeBetween
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.WindowSpec.rowsBetween.html">
pyspark.sql.WindowSpec.rowsBetween
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.GroupedData.agg.html">
pyspark.sql.GroupedData.agg
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.GroupedData.apply.html">
pyspark.sql.GroupedData.apply
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.GroupedData.applyInPandas.html">
pyspark.sql.GroupedData.applyInPandas
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.GroupedData.avg.html">
pyspark.sql.GroupedData.avg
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.GroupedData.cogroup.html">
pyspark.sql.GroupedData.cogroup
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.GroupedData.count.html">
pyspark.sql.GroupedData.count
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.GroupedData.max.html">
pyspark.sql.GroupedData.max
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.GroupedData.mean.html">
pyspark.sql.GroupedData.mean
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.GroupedData.min.html">
pyspark.sql.GroupedData.min
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.GroupedData.pivot.html">
pyspark.sql.GroupedData.pivot
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.GroupedData.sum.html">
pyspark.sql.GroupedData.sum
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.PandasCogroupedOps.applyInPandas.html">
pyspark.sql.PandasCogroupedOps.applyInPandas
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.cacheTable.html">
pyspark.sql.Catalog.cacheTable
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.clearCache.html">
pyspark.sql.Catalog.clearCache
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.createExternalTable.html">
pyspark.sql.Catalog.createExternalTable
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.createTable.html">
pyspark.sql.Catalog.createTable
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.currentDatabase.html">
pyspark.sql.Catalog.currentDatabase
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.dropGlobalTempView.html">
pyspark.sql.Catalog.dropGlobalTempView
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.dropTempView.html">
pyspark.sql.Catalog.dropTempView
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.isCached.html">
pyspark.sql.Catalog.isCached
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.listColumns.html">
pyspark.sql.Catalog.listColumns
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.listDatabases.html">
pyspark.sql.Catalog.listDatabases
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.listFunctions.html">
pyspark.sql.Catalog.listFunctions
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.listTables.html">
pyspark.sql.Catalog.listTables
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.recoverPartitions.html">
pyspark.sql.Catalog.recoverPartitions
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.refreshByPath.html">
pyspark.sql.Catalog.refreshByPath
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.refreshTable.html">
pyspark.sql.Catalog.refreshTable
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.registerFunction.html">
pyspark.sql.Catalog.registerFunction
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.setCurrentDatabase.html">
pyspark.sql.Catalog.setCurrentDatabase
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.Catalog.uncacheTable.html">
pyspark.sql.Catalog.uncacheTable
</a>
</li>
</ul>
</li>
<li class="toctree-l1 has-children">
<a class="reference internal" href="pyspark.pandas/index.html">
Pandas API on Spark
</a>
<input class="toctree-checkbox" id="toctree-checkbox-2" name="toctree-checkbox-2" type="checkbox"/>
<label for="toctree-checkbox-2">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l2 has-children">
<a class="reference internal" href="pyspark.pandas/io.html">
Input/Output
</a>
<input class="toctree-checkbox" id="toctree-checkbox-3" name="toctree-checkbox-3" type="checkbox"/>
<label for="toctree-checkbox-3">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.range.html">
pyspark.pandas.range
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.read_table.html">
pyspark.pandas.read_table
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_table.html">
pyspark.pandas.DataFrame.to_table
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.read_delta.html">
pyspark.pandas.read_delta
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_delta.html">
pyspark.pandas.DataFrame.to_delta
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.read_parquet.html">
pyspark.pandas.read_parquet
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_parquet.html">
pyspark.pandas.DataFrame.to_parquet
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.read_orc.html">
pyspark.pandas.read_orc
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_orc.html">
pyspark.pandas.DataFrame.to_orc
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.read_spark_io.html">
pyspark.pandas.read_spark_io
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_spark_io.html">
pyspark.pandas.DataFrame.to_spark_io
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.read_csv.html">
pyspark.pandas.read_csv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_csv.html">
pyspark.pandas.DataFrame.to_csv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.read_clipboard.html">
pyspark.pandas.read_clipboard
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_clipboard.html">
pyspark.pandas.DataFrame.to_clipboard
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.read_excel.html">
pyspark.pandas.read_excel
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_excel.html">
pyspark.pandas.DataFrame.to_excel
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.read_json.html">
pyspark.pandas.read_json
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_json.html">
pyspark.pandas.DataFrame.to_json
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.read_html.html">
pyspark.pandas.read_html
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_html.html">
pyspark.pandas.DataFrame.to_html
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.read_sql_table.html">
pyspark.pandas.read_sql_table
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.read_sql_query.html">
pyspark.pandas.read_sql_query
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.read_sql.html">
pyspark.pandas.read_sql
</a>
</li>
</ul>
</li>
<li class="toctree-l2 has-children">
<a class="reference internal" href="pyspark.pandas/general_functions.html">
General functions
</a>
<input class="toctree-checkbox" id="toctree-checkbox-4" name="toctree-checkbox-4" type="checkbox"/>
<label for="toctree-checkbox-4">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.reset_option.html">
pyspark.pandas.reset_option
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.get_option.html">
pyspark.pandas.get_option
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.set_option.html">
pyspark.pandas.set_option
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.option_context.html">
pyspark.pandas.option_context
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.melt.html">
pyspark.pandas.melt
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.merge.html">
pyspark.pandas.merge
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.get_dummies.html">
pyspark.pandas.get_dummies
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.concat.html">
pyspark.pandas.concat
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.sql.html">
pyspark.pandas.sql
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.broadcast.html">
pyspark.pandas.broadcast
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.to_numeric.html">
pyspark.pandas.to_numeric
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.isna.html">
pyspark.pandas.isna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.isnull.html">
pyspark.pandas.isnull
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.notna.html">
pyspark.pandas.notna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.notnull.html">
pyspark.pandas.notnull
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.to_datetime.html">
pyspark.pandas.to_datetime
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.date_range.html">
pyspark.pandas.date_range
</a>
</li>
</ul>
</li>
<li class="toctree-l2 has-children">
<a class="reference internal" href="pyspark.pandas/series.html">
Series
</a>
<input class="toctree-checkbox" id="toctree-checkbox-5" name="toctree-checkbox-5" type="checkbox"/>
<label for="toctree-checkbox-5">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.html">
pyspark.pandas.Series
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.index.html">
pyspark.pandas.Series.index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dtype.html">
pyspark.pandas.Series.dtype
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dtypes.html">
pyspark.pandas.Series.dtypes
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.ndim.html">
pyspark.pandas.Series.ndim
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.name.html">
pyspark.pandas.Series.name
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.shape.html">
pyspark.pandas.Series.shape
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.axes.html">
pyspark.pandas.Series.axes
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.size.html">
pyspark.pandas.Series.size
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.empty.html">
pyspark.pandas.Series.empty
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.T.html">
pyspark.pandas.Series.T
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.hasnans.html">
pyspark.pandas.Series.hasnans
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.values.html">
pyspark.pandas.Series.values
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.astype.html">
pyspark.pandas.Series.astype
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.copy.html">
pyspark.pandas.Series.copy
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.bool.html">
pyspark.pandas.Series.bool
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.at.html">
pyspark.pandas.Series.at
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.iat.html">
pyspark.pandas.Series.iat
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.loc.html">
pyspark.pandas.Series.loc
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.iloc.html">
pyspark.pandas.Series.iloc
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.keys.html">
pyspark.pandas.Series.keys
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.pop.html">
pyspark.pandas.Series.pop
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.items.html">
pyspark.pandas.Series.items
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.iteritems.html">
pyspark.pandas.Series.iteritems
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.item.html">
pyspark.pandas.Series.item
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.xs.html">
pyspark.pandas.Series.xs
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.get.html">
pyspark.pandas.Series.get
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.add.html">
pyspark.pandas.Series.add
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.div.html">
pyspark.pandas.Series.div
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.mul.html">
pyspark.pandas.Series.mul
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.radd.html">
pyspark.pandas.Series.radd
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.rdiv.html">
pyspark.pandas.Series.rdiv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.rmul.html">
pyspark.pandas.Series.rmul
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.rsub.html">
pyspark.pandas.Series.rsub
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.rtruediv.html">
pyspark.pandas.Series.rtruediv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.sub.html">
pyspark.pandas.Series.sub
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.truediv.html">
pyspark.pandas.Series.truediv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.pow.html">
pyspark.pandas.Series.pow
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.rpow.html">
pyspark.pandas.Series.rpow
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.mod.html">
pyspark.pandas.Series.mod
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.rmod.html">
pyspark.pandas.Series.rmod
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.floordiv.html">
pyspark.pandas.Series.floordiv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.rfloordiv.html">
pyspark.pandas.Series.rfloordiv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.divmod.html">
pyspark.pandas.Series.divmod
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.rdivmod.html">
pyspark.pandas.Series.rdivmod
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.combine_first.html">
pyspark.pandas.Series.combine_first
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.lt.html">
pyspark.pandas.Series.lt
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.gt.html">
pyspark.pandas.Series.gt
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.le.html">
pyspark.pandas.Series.le
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.ge.html">
pyspark.pandas.Series.ge
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.ne.html">
pyspark.pandas.Series.ne
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.eq.html">
pyspark.pandas.Series.eq
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.product.html">
pyspark.pandas.Series.product
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dot.html">
pyspark.pandas.Series.dot
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.apply.html">
pyspark.pandas.Series.apply
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.agg.html">
pyspark.pandas.Series.agg
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.aggregate.html">
pyspark.pandas.Series.aggregate
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.transform.html">
pyspark.pandas.Series.transform
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.map.html">
pyspark.pandas.Series.map
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.groupby.html">
pyspark.pandas.Series.groupby
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.rolling.html">
pyspark.pandas.Series.rolling
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.expanding.html">
pyspark.pandas.Series.expanding
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.pipe.html">
pyspark.pandas.Series.pipe
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.abs.html">
pyspark.pandas.Series.abs
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.all.html">
pyspark.pandas.Series.all
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.any.html">
pyspark.pandas.Series.any
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.between.html">
pyspark.pandas.Series.between
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.clip.html">
pyspark.pandas.Series.clip
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.corr.html">
pyspark.pandas.Series.corr
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.count.html">
pyspark.pandas.Series.count
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cummax.html">
pyspark.pandas.Series.cummax
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cummin.html">
pyspark.pandas.Series.cummin
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cumsum.html">
pyspark.pandas.Series.cumsum
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cumprod.html">
pyspark.pandas.Series.cumprod
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.describe.html">
pyspark.pandas.Series.describe
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.filter.html">
pyspark.pandas.Series.filter
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.kurt.html">
pyspark.pandas.Series.kurt
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.mad.html">
pyspark.pandas.Series.mad
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.max.html">
pyspark.pandas.Series.max
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.mean.html">
pyspark.pandas.Series.mean
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.min.html">
pyspark.pandas.Series.min
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.mode.html">
pyspark.pandas.Series.mode
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.nlargest.html">
pyspark.pandas.Series.nlargest
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.nsmallest.html">
pyspark.pandas.Series.nsmallest
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.pct_change.html">
pyspark.pandas.Series.pct_change
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.prod.html">
pyspark.pandas.Series.prod
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.nunique.html">
pyspark.pandas.Series.nunique
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.is_unique.html">
pyspark.pandas.Series.is_unique
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.quantile.html">
pyspark.pandas.Series.quantile
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.rank.html">
pyspark.pandas.Series.rank
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.sem.html">
pyspark.pandas.Series.sem
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.skew.html">
pyspark.pandas.Series.skew
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.std.html">
pyspark.pandas.Series.std
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.sum.html">
pyspark.pandas.Series.sum
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.median.html">
pyspark.pandas.Series.median
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.var.html">
pyspark.pandas.Series.var
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.kurtosis.html">
pyspark.pandas.Series.kurtosis
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.unique.html">
pyspark.pandas.Series.unique
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.value_counts.html">
pyspark.pandas.Series.value_counts
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.round.html">
pyspark.pandas.Series.round
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.diff.html">
pyspark.pandas.Series.diff
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.is_monotonic.html">
pyspark.pandas.Series.is_monotonic
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.is_monotonic_increasing.html">
pyspark.pandas.Series.is_monotonic_increasing
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.is_monotonic_decreasing.html">
pyspark.pandas.Series.is_monotonic_decreasing
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.align.html">
pyspark.pandas.Series.align
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.drop.html">
pyspark.pandas.Series.drop
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.droplevel.html">
pyspark.pandas.Series.droplevel
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.drop_duplicates.html">
pyspark.pandas.Series.drop_duplicates
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.equals.html">
pyspark.pandas.Series.equals
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.add_prefix.html">
pyspark.pandas.Series.add_prefix
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.add_suffix.html">
pyspark.pandas.Series.add_suffix
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.first.html">
pyspark.pandas.Series.first
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.head.html">
pyspark.pandas.Series.head
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.idxmax.html">
pyspark.pandas.Series.idxmax
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.idxmin.html">
pyspark.pandas.Series.idxmin
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.isin.html">
pyspark.pandas.Series.isin
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.last.html">
pyspark.pandas.Series.last
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.rename.html">
pyspark.pandas.Series.rename
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.rename_axis.html">
pyspark.pandas.Series.rename_axis
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.reindex.html">
pyspark.pandas.Series.reindex
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.reindex_like.html">
pyspark.pandas.Series.reindex_like
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.reset_index.html">
pyspark.pandas.Series.reset_index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.sample.html">
pyspark.pandas.Series.sample
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.swaplevel.html">
pyspark.pandas.Series.swaplevel
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.swapaxes.html">
pyspark.pandas.Series.swapaxes
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.take.html">
pyspark.pandas.Series.take
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.tail.html">
pyspark.pandas.Series.tail
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.where.html">
pyspark.pandas.Series.where
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.mask.html">
pyspark.pandas.Series.mask
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.truncate.html">
pyspark.pandas.Series.truncate
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.backfill.html">
pyspark.pandas.Series.backfill
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.bfill.html">
pyspark.pandas.Series.bfill
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.isna.html">
pyspark.pandas.Series.isna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.isnull.html">
pyspark.pandas.Series.isnull
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.notna.html">
pyspark.pandas.Series.notna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.notnull.html">
pyspark.pandas.Series.notnull
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.pad.html">
pyspark.pandas.Series.pad
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dropna.html">
pyspark.pandas.Series.dropna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.fillna.html">
pyspark.pandas.Series.fillna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.argsort.html">
pyspark.pandas.Series.argsort
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.argmin.html">
pyspark.pandas.Series.argmin
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.argmax.html">
pyspark.pandas.Series.argmax
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.sort_index.html">
pyspark.pandas.Series.sort_index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.sort_values.html">
pyspark.pandas.Series.sort_values
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.unstack.html">
pyspark.pandas.Series.unstack
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.explode.html">
pyspark.pandas.Series.explode
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.repeat.html">
pyspark.pandas.Series.repeat
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.squeeze.html">
pyspark.pandas.Series.squeeze
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.factorize.html">
pyspark.pandas.Series.factorize
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.append.html">
pyspark.pandas.Series.append
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.compare.html">
pyspark.pandas.Series.compare
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.replace.html">
pyspark.pandas.Series.replace
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.update.html">
pyspark.pandas.Series.update
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.asof.html">
pyspark.pandas.Series.asof
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.shift.html">
pyspark.pandas.Series.shift
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.first_valid_index.html">
pyspark.pandas.Series.first_valid_index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.last_valid_index.html">
pyspark.pandas.Series.last_valid_index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.at_time.html">
pyspark.pandas.Series.at_time
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.between_time.html">
pyspark.pandas.Series.between_time
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.spark.column.html">
pyspark.pandas.Series.spark.column
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.spark.transform.html">
pyspark.pandas.Series.spark.transform
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.spark.apply.html">
pyspark.pandas.Series.spark.apply
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.date.html">
pyspark.pandas.Series.dt.date
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.year.html">
pyspark.pandas.Series.dt.year
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.month.html">
pyspark.pandas.Series.dt.month
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.day.html">
pyspark.pandas.Series.dt.day
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.hour.html">
pyspark.pandas.Series.dt.hour
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.minute.html">
pyspark.pandas.Series.dt.minute
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.second.html">
pyspark.pandas.Series.dt.second
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.microsecond.html">
pyspark.pandas.Series.dt.microsecond
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.week.html">
pyspark.pandas.Series.dt.week
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.weekofyear.html">
pyspark.pandas.Series.dt.weekofyear
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.dayofweek.html">
pyspark.pandas.Series.dt.dayofweek
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.weekday.html">
pyspark.pandas.Series.dt.weekday
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.dayofyear.html">
pyspark.pandas.Series.dt.dayofyear
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.quarter.html">
pyspark.pandas.Series.dt.quarter
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.is_month_start.html">
pyspark.pandas.Series.dt.is_month_start
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.is_month_end.html">
pyspark.pandas.Series.dt.is_month_end
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.is_quarter_start.html">
pyspark.pandas.Series.dt.is_quarter_start
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.is_quarter_end.html">
pyspark.pandas.Series.dt.is_quarter_end
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.is_year_start.html">
pyspark.pandas.Series.dt.is_year_start
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.is_year_end.html">
pyspark.pandas.Series.dt.is_year_end
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.is_leap_year.html">
pyspark.pandas.Series.dt.is_leap_year
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.daysinmonth.html">
pyspark.pandas.Series.dt.daysinmonth
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.days_in_month.html">
pyspark.pandas.Series.dt.days_in_month
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.normalize.html">
pyspark.pandas.Series.dt.normalize
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.strftime.html">
pyspark.pandas.Series.dt.strftime
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.round.html">
pyspark.pandas.Series.dt.round
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.floor.html">
pyspark.pandas.Series.dt.floor
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.ceil.html">
pyspark.pandas.Series.dt.ceil
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.month_name.html">
pyspark.pandas.Series.dt.month_name
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.dt.day_name.html">
pyspark.pandas.Series.dt.day_name
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.capitalize.html">
pyspark.pandas.Series.str.capitalize
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.cat.html">
pyspark.pandas.Series.str.cat
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.center.html">
pyspark.pandas.Series.str.center
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.contains.html">
pyspark.pandas.Series.str.contains
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.count.html">
pyspark.pandas.Series.str.count
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.decode.html">
pyspark.pandas.Series.str.decode
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.encode.html">
pyspark.pandas.Series.str.encode
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.endswith.html">
pyspark.pandas.Series.str.endswith
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.extract.html">
pyspark.pandas.Series.str.extract
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.extractall.html">
pyspark.pandas.Series.str.extractall
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.find.html">
pyspark.pandas.Series.str.find
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.findall.html">
pyspark.pandas.Series.str.findall
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.get.html">
pyspark.pandas.Series.str.get
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.get_dummies.html">
pyspark.pandas.Series.str.get_dummies
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.index.html">
pyspark.pandas.Series.str.index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.isalnum.html">
pyspark.pandas.Series.str.isalnum
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.isalpha.html">
pyspark.pandas.Series.str.isalpha
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.isdigit.html">
pyspark.pandas.Series.str.isdigit
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.isspace.html">
pyspark.pandas.Series.str.isspace
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.islower.html">
pyspark.pandas.Series.str.islower
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.isupper.html">
pyspark.pandas.Series.str.isupper
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.istitle.html">
pyspark.pandas.Series.str.istitle
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.isnumeric.html">
pyspark.pandas.Series.str.isnumeric
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.isdecimal.html">
pyspark.pandas.Series.str.isdecimal
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.join.html">
pyspark.pandas.Series.str.join
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.len.html">
pyspark.pandas.Series.str.len
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.ljust.html">
pyspark.pandas.Series.str.ljust
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.lower.html">
pyspark.pandas.Series.str.lower
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.lstrip.html">
pyspark.pandas.Series.str.lstrip
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.match.html">
pyspark.pandas.Series.str.match
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.normalize.html">
pyspark.pandas.Series.str.normalize
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.pad.html">
pyspark.pandas.Series.str.pad
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.partition.html">
pyspark.pandas.Series.str.partition
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.repeat.html">
pyspark.pandas.Series.str.repeat
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.replace.html">
pyspark.pandas.Series.str.replace
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.rfind.html">
pyspark.pandas.Series.str.rfind
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.rindex.html">
pyspark.pandas.Series.str.rindex
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.rjust.html">
pyspark.pandas.Series.str.rjust
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.rpartition.html">
pyspark.pandas.Series.str.rpartition
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.rsplit.html">
pyspark.pandas.Series.str.rsplit
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.rstrip.html">
pyspark.pandas.Series.str.rstrip
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.slice.html">
pyspark.pandas.Series.str.slice
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.slice_replace.html">
pyspark.pandas.Series.str.slice_replace
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.split.html">
pyspark.pandas.Series.str.split
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.startswith.html">
pyspark.pandas.Series.str.startswith
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.strip.html">
pyspark.pandas.Series.str.strip
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.swapcase.html">
pyspark.pandas.Series.str.swapcase
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.title.html">
pyspark.pandas.Series.str.title
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.translate.html">
pyspark.pandas.Series.str.translate
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.upper.html">
pyspark.pandas.Series.str.upper
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.wrap.html">
pyspark.pandas.Series.str.wrap
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.str.zfill.html">
pyspark.pandas.Series.str.zfill
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cat.categories.html">
pyspark.pandas.Series.cat.categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cat.ordered.html">
pyspark.pandas.Series.cat.ordered
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cat.codes.html">
pyspark.pandas.Series.cat.codes
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cat.rename_categories.html">
pyspark.pandas.Series.cat.rename_categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cat.reorder_categories.html">
pyspark.pandas.Series.cat.reorder_categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cat.add_categories.html">
pyspark.pandas.Series.cat.add_categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cat.remove_categories.html">
pyspark.pandas.Series.cat.remove_categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cat.remove_unused_categories.html">
pyspark.pandas.Series.cat.remove_unused_categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cat.set_categories.html">
pyspark.pandas.Series.cat.set_categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cat.as_ordered.html">
pyspark.pandas.Series.cat.as_ordered
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.cat.as_unordered.html">
pyspark.pandas.Series.cat.as_unordered
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.plot.html">
pyspark.pandas.Series.plot
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.plot.area.html">
pyspark.pandas.Series.plot.area
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.plot.bar.html">
pyspark.pandas.Series.plot.bar
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.plot.barh.html">
pyspark.pandas.Series.plot.barh
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.plot.box.html">
pyspark.pandas.Series.plot.box
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.plot.density.html">
pyspark.pandas.Series.plot.density
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.plot.hist.html">
pyspark.pandas.Series.plot.hist
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.plot.line.html">
pyspark.pandas.Series.plot.line
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.plot.pie.html">
pyspark.pandas.Series.plot.pie
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.plot.kde.html">
pyspark.pandas.Series.plot.kde
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.hist.html">
pyspark.pandas.Series.hist
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.to_pandas.html">
pyspark.pandas.Series.to_pandas
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.to_numpy.html">
pyspark.pandas.Series.to_numpy
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.to_list.html">
pyspark.pandas.Series.to_list
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.to_string.html">
pyspark.pandas.Series.to_string
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.to_dict.html">
pyspark.pandas.Series.to_dict
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.to_clipboard.html">
pyspark.pandas.Series.to_clipboard
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.to_latex.html">
pyspark.pandas.Series.to_latex
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.to_markdown.html">
pyspark.pandas.Series.to_markdown
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.to_json.html">
pyspark.pandas.Series.to_json
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.to_csv.html">
pyspark.pandas.Series.to_csv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.to_excel.html">
pyspark.pandas.Series.to_excel
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.to_frame.html">
pyspark.pandas.Series.to_frame
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Series.pandas_on_spark.transform_batch.html">
pyspark.pandas.Series.pandas_on_spark.transform_batch
</a>
</li>
</ul>
</li>
<li class="toctree-l2 has-children">
<a class="reference internal" href="pyspark.pandas/frame.html">
DataFrame
</a>
<input class="toctree-checkbox" id="toctree-checkbox-6" name="toctree-checkbox-6" type="checkbox"/>
<label for="toctree-checkbox-6">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.html">
pyspark.pandas.DataFrame
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.index.html">
pyspark.pandas.DataFrame.index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.columns.html">
pyspark.pandas.DataFrame.columns
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.empty.html">
pyspark.pandas.DataFrame.empty
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.dtypes.html">
pyspark.pandas.DataFrame.dtypes
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.shape.html">
pyspark.pandas.DataFrame.shape
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.axes.html">
pyspark.pandas.DataFrame.axes
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.ndim.html">
pyspark.pandas.DataFrame.ndim
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.size.html">
pyspark.pandas.DataFrame.size
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.select_dtypes.html">
pyspark.pandas.DataFrame.select_dtypes
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.values.html">
pyspark.pandas.DataFrame.values
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.copy.html">
pyspark.pandas.DataFrame.copy
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.isna.html">
pyspark.pandas.DataFrame.isna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.astype.html">
pyspark.pandas.DataFrame.astype
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.isnull.html">
pyspark.pandas.DataFrame.isnull
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.notna.html">
pyspark.pandas.DataFrame.notna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.notnull.html">
pyspark.pandas.DataFrame.notnull
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.pad.html">
pyspark.pandas.DataFrame.pad
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.bool.html">
pyspark.pandas.DataFrame.bool
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.at.html">
pyspark.pandas.DataFrame.at
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.iat.html">
pyspark.pandas.DataFrame.iat
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.head.html">
pyspark.pandas.DataFrame.head
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.idxmax.html">
pyspark.pandas.DataFrame.idxmax
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.idxmin.html">
pyspark.pandas.DataFrame.idxmin
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.loc.html">
pyspark.pandas.DataFrame.loc
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.iloc.html">
pyspark.pandas.DataFrame.iloc
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.items.html">
pyspark.pandas.DataFrame.items
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.iteritems.html">
pyspark.pandas.DataFrame.iteritems
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.iterrows.html">
pyspark.pandas.DataFrame.iterrows
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.itertuples.html">
pyspark.pandas.DataFrame.itertuples
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.keys.html">
pyspark.pandas.DataFrame.keys
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.pop.html">
pyspark.pandas.DataFrame.pop
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.tail.html">
pyspark.pandas.DataFrame.tail
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.xs.html">
pyspark.pandas.DataFrame.xs
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.get.html">
pyspark.pandas.DataFrame.get
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.where.html">
pyspark.pandas.DataFrame.where
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.mask.html">
pyspark.pandas.DataFrame.mask
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.query.html">
pyspark.pandas.DataFrame.query
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.add.html">
pyspark.pandas.DataFrame.add
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.radd.html">
pyspark.pandas.DataFrame.radd
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.div.html">
pyspark.pandas.DataFrame.div
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.rdiv.html">
pyspark.pandas.DataFrame.rdiv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.truediv.html">
pyspark.pandas.DataFrame.truediv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.rtruediv.html">
pyspark.pandas.DataFrame.rtruediv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.mul.html">
pyspark.pandas.DataFrame.mul
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.rmul.html">
pyspark.pandas.DataFrame.rmul
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.sub.html">
pyspark.pandas.DataFrame.sub
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.rsub.html">
pyspark.pandas.DataFrame.rsub
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.pow.html">
pyspark.pandas.DataFrame.pow
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.rpow.html">
pyspark.pandas.DataFrame.rpow
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.mod.html">
pyspark.pandas.DataFrame.mod
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.rmod.html">
pyspark.pandas.DataFrame.rmod
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.floordiv.html">
pyspark.pandas.DataFrame.floordiv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.rfloordiv.html">
pyspark.pandas.DataFrame.rfloordiv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.lt.html">
pyspark.pandas.DataFrame.lt
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.gt.html">
pyspark.pandas.DataFrame.gt
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.le.html">
pyspark.pandas.DataFrame.le
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.ge.html">
pyspark.pandas.DataFrame.ge
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.ne.html">
pyspark.pandas.DataFrame.ne
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.eq.html">
pyspark.pandas.DataFrame.eq
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.dot.html">
pyspark.pandas.DataFrame.dot
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.apply.html">
pyspark.pandas.DataFrame.apply
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.applymap.html">
pyspark.pandas.DataFrame.applymap
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.pipe.html">
pyspark.pandas.DataFrame.pipe
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.agg.html">
pyspark.pandas.DataFrame.agg
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.aggregate.html">
pyspark.pandas.DataFrame.aggregate
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.groupby.html">
pyspark.pandas.DataFrame.groupby
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.rolling.html">
pyspark.pandas.DataFrame.rolling
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.expanding.html">
pyspark.pandas.DataFrame.expanding
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.transform.html">
pyspark.pandas.DataFrame.transform
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.abs.html">
pyspark.pandas.DataFrame.abs
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.all.html">
pyspark.pandas.DataFrame.all
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.any.html">
pyspark.pandas.DataFrame.any
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.clip.html">
pyspark.pandas.DataFrame.clip
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.corr.html">
pyspark.pandas.DataFrame.corr
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.count.html">
pyspark.pandas.DataFrame.count
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.describe.html">
pyspark.pandas.DataFrame.describe
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.kurt.html">
pyspark.pandas.DataFrame.kurt
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.kurtosis.html">
pyspark.pandas.DataFrame.kurtosis
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.mad.html">
pyspark.pandas.DataFrame.mad
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.max.html">
pyspark.pandas.DataFrame.max
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.mean.html">
pyspark.pandas.DataFrame.mean
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.min.html">
pyspark.pandas.DataFrame.min
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.median.html">
pyspark.pandas.DataFrame.median
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.pct_change.html">
pyspark.pandas.DataFrame.pct_change
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.prod.html">
pyspark.pandas.DataFrame.prod
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.product.html">
pyspark.pandas.DataFrame.product
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.quantile.html">
pyspark.pandas.DataFrame.quantile
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.nunique.html">
pyspark.pandas.DataFrame.nunique
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.sem.html">
pyspark.pandas.DataFrame.sem
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.skew.html">
pyspark.pandas.DataFrame.skew
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.sum.html">
pyspark.pandas.DataFrame.sum
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.std.html">
pyspark.pandas.DataFrame.std
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.var.html">
pyspark.pandas.DataFrame.var
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.cummin.html">
pyspark.pandas.DataFrame.cummin
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.cummax.html">
pyspark.pandas.DataFrame.cummax
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.cumsum.html">
pyspark.pandas.DataFrame.cumsum
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.cumprod.html">
pyspark.pandas.DataFrame.cumprod
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.round.html">
pyspark.pandas.DataFrame.round
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.diff.html">
pyspark.pandas.DataFrame.diff
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.eval.html">
pyspark.pandas.DataFrame.eval
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.add_prefix.html">
pyspark.pandas.DataFrame.add_prefix
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.add_suffix.html">
pyspark.pandas.DataFrame.add_suffix
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.align.html">
pyspark.pandas.DataFrame.align
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.at_time.html">
pyspark.pandas.DataFrame.at_time
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.between_time.html">
pyspark.pandas.DataFrame.between_time
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.drop.html">
pyspark.pandas.DataFrame.drop
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.droplevel.html">
pyspark.pandas.DataFrame.droplevel
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.drop_duplicates.html">
pyspark.pandas.DataFrame.drop_duplicates
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.duplicated.html">
pyspark.pandas.DataFrame.duplicated
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.equals.html">
pyspark.pandas.DataFrame.equals
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.filter.html">
pyspark.pandas.DataFrame.filter
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.first.html">
pyspark.pandas.DataFrame.first
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.head.html">
pyspark.pandas.DataFrame.head
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.last.html">
pyspark.pandas.DataFrame.last
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.rename.html">
pyspark.pandas.DataFrame.rename
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.rename_axis.html">
pyspark.pandas.DataFrame.rename_axis
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.reset_index.html">
pyspark.pandas.DataFrame.reset_index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.set_index.html">
pyspark.pandas.DataFrame.set_index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.swapaxes.html">
pyspark.pandas.DataFrame.swapaxes
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.swaplevel.html">
pyspark.pandas.DataFrame.swaplevel
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.take.html">
pyspark.pandas.DataFrame.take
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.isin.html">
pyspark.pandas.DataFrame.isin
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.sample.html">
pyspark.pandas.DataFrame.sample
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.truncate.html">
pyspark.pandas.DataFrame.truncate
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.backfill.html">
pyspark.pandas.DataFrame.backfill
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.dropna.html">
pyspark.pandas.DataFrame.dropna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.fillna.html">
pyspark.pandas.DataFrame.fillna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.replace.html">
pyspark.pandas.DataFrame.replace
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.bfill.html">
pyspark.pandas.DataFrame.bfill
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.ffill.html">
pyspark.pandas.DataFrame.ffill
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.pivot_table.html">
pyspark.pandas.DataFrame.pivot_table
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.pivot.html">
pyspark.pandas.DataFrame.pivot
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.sort_index.html">
pyspark.pandas.DataFrame.sort_index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.sort_values.html">
pyspark.pandas.DataFrame.sort_values
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.nlargest.html">
pyspark.pandas.DataFrame.nlargest
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.nsmallest.html">
pyspark.pandas.DataFrame.nsmallest
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.stack.html">
pyspark.pandas.DataFrame.stack
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.unstack.html">
pyspark.pandas.DataFrame.unstack
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.melt.html">
pyspark.pandas.DataFrame.melt
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.explode.html">
pyspark.pandas.DataFrame.explode
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.squeeze.html">
pyspark.pandas.DataFrame.squeeze
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.T.html">
pyspark.pandas.DataFrame.T
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.transpose.html">
pyspark.pandas.DataFrame.transpose
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.reindex.html">
pyspark.pandas.DataFrame.reindex
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.reindex_like.html">
pyspark.pandas.DataFrame.reindex_like
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.rank.html">
pyspark.pandas.DataFrame.rank
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.append.html">
pyspark.pandas.DataFrame.append
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.assign.html">
pyspark.pandas.DataFrame.assign
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.merge.html">
pyspark.pandas.DataFrame.merge
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.join.html">
pyspark.pandas.DataFrame.join
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.update.html">
pyspark.pandas.DataFrame.update
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.insert.html">
pyspark.pandas.DataFrame.insert
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.shift.html">
pyspark.pandas.DataFrame.shift
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.first_valid_index.html">
pyspark.pandas.DataFrame.first_valid_index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.last_valid_index.html">
pyspark.pandas.DataFrame.last_valid_index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.from_records.html">
pyspark.pandas.DataFrame.from_records
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.info.html">
pyspark.pandas.DataFrame.info
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_table.html">
pyspark.pandas.DataFrame.to_table
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_delta.html">
pyspark.pandas.DataFrame.to_delta
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_parquet.html">
pyspark.pandas.DataFrame.to_parquet
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_spark_io.html">
pyspark.pandas.DataFrame.to_spark_io
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_csv.html">
pyspark.pandas.DataFrame.to_csv
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_pandas.html">
pyspark.pandas.DataFrame.to_pandas
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_html.html">
pyspark.pandas.DataFrame.to_html
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_numpy.html">
pyspark.pandas.DataFrame.to_numpy
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_spark.html">
pyspark.pandas.DataFrame.to_spark
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_string.html">
pyspark.pandas.DataFrame.to_string
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_json.html">
pyspark.pandas.DataFrame.to_json
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_dict.html">
pyspark.pandas.DataFrame.to_dict
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_excel.html">
pyspark.pandas.DataFrame.to_excel
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_clipboard.html">
pyspark.pandas.DataFrame.to_clipboard
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_markdown.html">
pyspark.pandas.DataFrame.to_markdown
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_records.html">
pyspark.pandas.DataFrame.to_records
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.to_latex.html">
pyspark.pandas.DataFrame.to_latex
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.style.html">
pyspark.pandas.DataFrame.style
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.spark.frame.html">
pyspark.pandas.DataFrame.spark.frame
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.spark.cache.html">
pyspark.pandas.DataFrame.spark.cache
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.spark.persist.html">
pyspark.pandas.DataFrame.spark.persist
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.spark.hint.html">
pyspark.pandas.DataFrame.spark.hint
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.spark.to_table.html">
pyspark.pandas.DataFrame.spark.to_table
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.spark.to_spark_io.html">
pyspark.pandas.DataFrame.spark.to_spark_io
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.spark.apply.html">
pyspark.pandas.DataFrame.spark.apply
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.spark.repartition.html">
pyspark.pandas.DataFrame.spark.repartition
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.spark.coalesce.html">
pyspark.pandas.DataFrame.spark.coalesce
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.plot.html">
pyspark.pandas.DataFrame.plot
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.plot.area.html">
pyspark.pandas.DataFrame.plot.area
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.plot.barh.html">
pyspark.pandas.DataFrame.plot.barh
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.plot.bar.html">
pyspark.pandas.DataFrame.plot.bar
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.plot.hist.html">
pyspark.pandas.DataFrame.plot.hist
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.plot.line.html">
pyspark.pandas.DataFrame.plot.line
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.plot.pie.html">
pyspark.pandas.DataFrame.plot.pie
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.plot.scatter.html">
pyspark.pandas.DataFrame.plot.scatter
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.plot.density.html">
pyspark.pandas.DataFrame.plot.density
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.hist.html">
pyspark.pandas.DataFrame.hist
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.kde.html">
pyspark.pandas.DataFrame.kde
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.pandas_on_spark.apply_batch.html">
pyspark.pandas.DataFrame.pandas_on_spark.apply_batch
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DataFrame.pandas_on_spark.transform_batch.html">
pyspark.pandas.DataFrame.pandas_on_spark.transform_batch
</a>
</li>
</ul>
</li>
<li class="toctree-l2 has-children">
<a class="reference internal" href="pyspark.pandas/indexing.html">
Index objects
</a>
<input class="toctree-checkbox" id="toctree-checkbox-7" name="toctree-checkbox-7" type="checkbox"/>
<label for="toctree-checkbox-7">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.html">
pyspark.pandas.Index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.is_monotonic.html">
pyspark.pandas.Index.is_monotonic
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.is_monotonic_increasing.html">
pyspark.pandas.Index.is_monotonic_increasing
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.is_monotonic_decreasing.html">
pyspark.pandas.Index.is_monotonic_decreasing
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.is_unique.html">
pyspark.pandas.Index.is_unique
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.has_duplicates.html">
pyspark.pandas.Index.has_duplicates
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.hasnans.html">
pyspark.pandas.Index.hasnans
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.dtype.html">
pyspark.pandas.Index.dtype
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.inferred_type.html">
pyspark.pandas.Index.inferred_type
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.is_all_dates.html">
pyspark.pandas.Index.is_all_dates
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.shape.html">
pyspark.pandas.Index.shape
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.name.html">
pyspark.pandas.Index.name
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.names.html">
pyspark.pandas.Index.names
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.ndim.html">
pyspark.pandas.Index.ndim
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.size.html">
pyspark.pandas.Index.size
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.nlevels.html">
pyspark.pandas.Index.nlevels
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.empty.html">
pyspark.pandas.Index.empty
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.T.html">
pyspark.pandas.Index.T
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.values.html">
pyspark.pandas.Index.values
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.all.html">
pyspark.pandas.Index.all
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.any.html">
pyspark.pandas.Index.any
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.argmin.html">
pyspark.pandas.Index.argmin
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.argmax.html">
pyspark.pandas.Index.argmax
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.copy.html">
pyspark.pandas.Index.copy
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.delete.html">
pyspark.pandas.Index.delete
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.equals.html">
pyspark.pandas.Index.equals
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.factorize.html">
pyspark.pandas.Index.factorize
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.identical.html">
pyspark.pandas.Index.identical
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.insert.html">
pyspark.pandas.Index.insert
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.is_boolean.html">
pyspark.pandas.Index.is_boolean
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.is_categorical.html">
pyspark.pandas.Index.is_categorical
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.is_floating.html">
pyspark.pandas.Index.is_floating
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.is_integer.html">
pyspark.pandas.Index.is_integer
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.is_interval.html">
pyspark.pandas.Index.is_interval
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.is_numeric.html">
pyspark.pandas.Index.is_numeric
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.is_object.html">
pyspark.pandas.Index.is_object
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.drop.html">
pyspark.pandas.Index.drop
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.drop_duplicates.html">
pyspark.pandas.Index.drop_duplicates
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.min.html">
pyspark.pandas.Index.min
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.max.html">
pyspark.pandas.Index.max
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.map.html">
pyspark.pandas.Index.map
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.rename.html">
pyspark.pandas.Index.rename
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.repeat.html">
pyspark.pandas.Index.repeat
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.take.html">
pyspark.pandas.Index.take
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.unique.html">
pyspark.pandas.Index.unique
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.nunique.html">
pyspark.pandas.Index.nunique
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.value_counts.html">
pyspark.pandas.Index.value_counts
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.set_names.html">
pyspark.pandas.Index.set_names
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.droplevel.html">
pyspark.pandas.Index.droplevel
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.fillna.html">
pyspark.pandas.Index.fillna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.dropna.html">
pyspark.pandas.Index.dropna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.isna.html">
pyspark.pandas.Index.isna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.notna.html">
pyspark.pandas.Index.notna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.astype.html">
pyspark.pandas.Index.astype
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.item.html">
pyspark.pandas.Index.item
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.to_list.html">
pyspark.pandas.Index.to_list
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.to_series.html">
pyspark.pandas.Index.to_series
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.to_frame.html">
pyspark.pandas.Index.to_frame
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.view.html">
pyspark.pandas.Index.view
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.to_numpy.html">
pyspark.pandas.Index.to_numpy
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.spark.column.html">
pyspark.pandas.Index.spark.column
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.spark.transform.html">
pyspark.pandas.Index.spark.transform
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.sort_values.html">
pyspark.pandas.Index.sort_values
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.shift.html">
pyspark.pandas.Index.shift
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.append.html">
pyspark.pandas.Index.append
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.intersection.html">
pyspark.pandas.Index.intersection
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.union.html">
pyspark.pandas.Index.union
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.difference.html">
pyspark.pandas.Index.difference
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.symmetric_difference.html">
pyspark.pandas.Index.symmetric_difference
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.asof.html">
pyspark.pandas.Index.asof
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Index.isin.html">
pyspark.pandas.Index.isin
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Int64Index.html">
pyspark.pandas.Int64Index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.Float64Index.html">
pyspark.pandas.Float64Index
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.CategoricalIndex.html">
pyspark.pandas.CategoricalIndex
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.CategoricalIndex.codes.html">
pyspark.pandas.CategoricalIndex.codes
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.CategoricalIndex.categories.html">
pyspark.pandas.CategoricalIndex.categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.CategoricalIndex.ordered.html">
pyspark.pandas.CategoricalIndex.ordered
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.CategoricalIndex.rename_categories.html">
pyspark.pandas.CategoricalIndex.rename_categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.CategoricalIndex.reorder_categories.html">
pyspark.pandas.CategoricalIndex.reorder_categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.CategoricalIndex.add_categories.html">
pyspark.pandas.CategoricalIndex.add_categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.CategoricalIndex.remove_categories.html">
pyspark.pandas.CategoricalIndex.remove_categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.CategoricalIndex.remove_unused_categories.html">
pyspark.pandas.CategoricalIndex.remove_unused_categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.CategoricalIndex.set_categories.html">
pyspark.pandas.CategoricalIndex.set_categories
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.CategoricalIndex.as_ordered.html">
pyspark.pandas.CategoricalIndex.as_ordered
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.CategoricalIndex.as_unordered.html">
pyspark.pandas.CategoricalIndex.as_unordered
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.CategoricalIndex.map.html">
pyspark.pandas.CategoricalIndex.map
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.html">
pyspark.pandas.MultiIndex
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.from_arrays.html">
pyspark.pandas.MultiIndex.from_arrays
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.from_tuples.html">
pyspark.pandas.MultiIndex.from_tuples
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.from_product.html">
pyspark.pandas.MultiIndex.from_product
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.from_frame.html">
pyspark.pandas.MultiIndex.from_frame
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.has_duplicates.html">
pyspark.pandas.MultiIndex.has_duplicates
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.hasnans.html">
pyspark.pandas.MultiIndex.hasnans
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.inferred_type.html">
pyspark.pandas.MultiIndex.inferred_type
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.is_all_dates.html">
pyspark.pandas.MultiIndex.is_all_dates
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.shape.html">
pyspark.pandas.MultiIndex.shape
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.names.html">
pyspark.pandas.MultiIndex.names
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.ndim.html">
pyspark.pandas.MultiIndex.ndim
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.empty.html">
pyspark.pandas.MultiIndex.empty
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.T.html">
pyspark.pandas.MultiIndex.T
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.size.html">
pyspark.pandas.MultiIndex.size
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.nlevels.html">
pyspark.pandas.MultiIndex.nlevels
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.levshape.html">
pyspark.pandas.MultiIndex.levshape
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.values.html">
pyspark.pandas.MultiIndex.values
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.swaplevel.html">
pyspark.pandas.MultiIndex.swaplevel
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.droplevel.html">
pyspark.pandas.MultiIndex.droplevel
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.fillna.html">
pyspark.pandas.MultiIndex.fillna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.dropna.html">
pyspark.pandas.MultiIndex.dropna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.equals.html">
pyspark.pandas.MultiIndex.equals
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.identical.html">
pyspark.pandas.MultiIndex.identical
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.insert.html">
pyspark.pandas.MultiIndex.insert
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.drop.html">
pyspark.pandas.MultiIndex.drop
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.copy.html">
pyspark.pandas.MultiIndex.copy
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.delete.html">
pyspark.pandas.MultiIndex.delete
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.rename.html">
pyspark.pandas.MultiIndex.rename
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.repeat.html">
pyspark.pandas.MultiIndex.repeat
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.take.html">
pyspark.pandas.MultiIndex.take
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.unique.html">
pyspark.pandas.MultiIndex.unique
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.min.html">
pyspark.pandas.MultiIndex.min
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.max.html">
pyspark.pandas.MultiIndex.max
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.value_counts.html">
pyspark.pandas.MultiIndex.value_counts
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.append.html">
pyspark.pandas.MultiIndex.append
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.intersection.html">
pyspark.pandas.MultiIndex.intersection
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.union.html">
pyspark.pandas.MultiIndex.union
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.difference.html">
pyspark.pandas.MultiIndex.difference
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.symmetric_difference.html">
pyspark.pandas.MultiIndex.symmetric_difference
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.astype.html">
pyspark.pandas.MultiIndex.astype
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.item.html">
pyspark.pandas.MultiIndex.item
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.to_list.html">
pyspark.pandas.MultiIndex.to_list
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.to_series.html">
pyspark.pandas.MultiIndex.to_series
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.to_frame.html">
pyspark.pandas.MultiIndex.to_frame
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.view.html">
pyspark.pandas.MultiIndex.view
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.to_numpy.html">
pyspark.pandas.MultiIndex.to_numpy
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.spark.data_type.html">
pyspark.pandas.MultiIndex.spark.data_type
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.spark.column.html">
pyspark.pandas.MultiIndex.spark.column
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.spark.transform.html">
pyspark.pandas.MultiIndex.spark.transform
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.MultiIndex.sort_values.html">
pyspark.pandas.MultiIndex.sort_values
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.html">
pyspark.pandas.DatetimeIndex
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.year.html">
pyspark.pandas.DatetimeIndex.year
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.month.html">
pyspark.pandas.DatetimeIndex.month
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.day.html">
pyspark.pandas.DatetimeIndex.day
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.hour.html">
pyspark.pandas.DatetimeIndex.hour
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.minute.html">
pyspark.pandas.DatetimeIndex.minute
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.second.html">
pyspark.pandas.DatetimeIndex.second
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.microsecond.html">
pyspark.pandas.DatetimeIndex.microsecond
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.week.html">
pyspark.pandas.DatetimeIndex.week
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.weekofyear.html">
pyspark.pandas.DatetimeIndex.weekofyear
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.dayofweek.html">
pyspark.pandas.DatetimeIndex.dayofweek
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.day_of_week.html">
pyspark.pandas.DatetimeIndex.day_of_week
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.weekday.html">
pyspark.pandas.DatetimeIndex.weekday
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.dayofyear.html">
pyspark.pandas.DatetimeIndex.dayofyear
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.day_of_year.html">
pyspark.pandas.DatetimeIndex.day_of_year
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.quarter.html">
pyspark.pandas.DatetimeIndex.quarter
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.is_month_start.html">
pyspark.pandas.DatetimeIndex.is_month_start
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.is_month_end.html">
pyspark.pandas.DatetimeIndex.is_month_end
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.is_quarter_start.html">
pyspark.pandas.DatetimeIndex.is_quarter_start
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.is_quarter_end.html">
pyspark.pandas.DatetimeIndex.is_quarter_end
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.is_year_start.html">
pyspark.pandas.DatetimeIndex.is_year_start
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.is_year_end.html">
pyspark.pandas.DatetimeIndex.is_year_end
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.is_leap_year.html">
pyspark.pandas.DatetimeIndex.is_leap_year
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.daysinmonth.html">
pyspark.pandas.DatetimeIndex.daysinmonth
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.days_in_month.html">
pyspark.pandas.DatetimeIndex.days_in_month
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.indexer_between_time.html">
pyspark.pandas.DatetimeIndex.indexer_between_time
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.indexer_at_time.html">
pyspark.pandas.DatetimeIndex.indexer_at_time
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.normalize.html">
pyspark.pandas.DatetimeIndex.normalize
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.strftime.html">
pyspark.pandas.DatetimeIndex.strftime
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.round.html">
pyspark.pandas.DatetimeIndex.round
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.floor.html">
pyspark.pandas.DatetimeIndex.floor
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.ceil.html">
pyspark.pandas.DatetimeIndex.ceil
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.month_name.html">
pyspark.pandas.DatetimeIndex.month_name
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.DatetimeIndex.day_name.html">
pyspark.pandas.DatetimeIndex.day_name
</a>
</li>
</ul>
</li>
<li class="toctree-l2 has-children">
<a class="reference internal" href="pyspark.pandas/window.html">
Window
</a>
<input class="toctree-checkbox" id="toctree-checkbox-8" name="toctree-checkbox-8" type="checkbox"/>
<label for="toctree-checkbox-8">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.window.Rolling.count.html">
pyspark.pandas.window.Rolling.count
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.window.Rolling.sum.html">
pyspark.pandas.window.Rolling.sum
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.window.Rolling.min.html">
pyspark.pandas.window.Rolling.min
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.window.Rolling.max.html">
pyspark.pandas.window.Rolling.max
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.window.Rolling.mean.html">
pyspark.pandas.window.Rolling.mean
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.window.Expanding.count.html">
pyspark.pandas.window.Expanding.count
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.window.Expanding.sum.html">
pyspark.pandas.window.Expanding.sum
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.window.Expanding.min.html">
pyspark.pandas.window.Expanding.min
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.window.Expanding.max.html">
pyspark.pandas.window.Expanding.max
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.window.Expanding.mean.html">
pyspark.pandas.window.Expanding.mean
</a>
</li>
</ul>
</li>
<li class="toctree-l2 has-children">
<a class="reference internal" href="pyspark.pandas/groupby.html">
GroupBy
</a>
<input class="toctree-checkbox" id="toctree-checkbox-9" name="toctree-checkbox-9" type="checkbox"/>
<label for="toctree-checkbox-9">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.get_group.html">
pyspark.pandas.groupby.GroupBy.get_group
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.apply.html">
pyspark.pandas.groupby.GroupBy.apply
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.transform.html">
pyspark.pandas.groupby.GroupBy.transform
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.DataFrameGroupBy.agg.html">
pyspark.pandas.groupby.DataFrameGroupBy.agg
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.DataFrameGroupBy.aggregate.html">
pyspark.pandas.groupby.DataFrameGroupBy.aggregate
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.all.html">
pyspark.pandas.groupby.GroupBy.all
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.any.html">
pyspark.pandas.groupby.GroupBy.any
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.count.html">
pyspark.pandas.groupby.GroupBy.count
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.cumcount.html">
pyspark.pandas.groupby.GroupBy.cumcount
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.cummax.html">
pyspark.pandas.groupby.GroupBy.cummax
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.cummin.html">
pyspark.pandas.groupby.GroupBy.cummin
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.cumprod.html">
pyspark.pandas.groupby.GroupBy.cumprod
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.cumsum.html">
pyspark.pandas.groupby.GroupBy.cumsum
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.filter.html">
pyspark.pandas.groupby.GroupBy.filter
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.first.html">
pyspark.pandas.groupby.GroupBy.first
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.last.html">
pyspark.pandas.groupby.GroupBy.last
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.max.html">
pyspark.pandas.groupby.GroupBy.max
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.mean.html">
pyspark.pandas.groupby.GroupBy.mean
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.median.html">
pyspark.pandas.groupby.GroupBy.median
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.min.html">
pyspark.pandas.groupby.GroupBy.min
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.rank.html">
pyspark.pandas.groupby.GroupBy.rank
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.std.html">
pyspark.pandas.groupby.GroupBy.std
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.sum.html">
pyspark.pandas.groupby.GroupBy.sum
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.var.html">
pyspark.pandas.groupby.GroupBy.var
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.nunique.html">
pyspark.pandas.groupby.GroupBy.nunique
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.size.html">
pyspark.pandas.groupby.GroupBy.size
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.diff.html">
pyspark.pandas.groupby.GroupBy.diff
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.idxmax.html">
pyspark.pandas.groupby.GroupBy.idxmax
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.idxmin.html">
pyspark.pandas.groupby.GroupBy.idxmin
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.fillna.html">
pyspark.pandas.groupby.GroupBy.fillna
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.bfill.html">
pyspark.pandas.groupby.GroupBy.bfill
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.ffill.html">
pyspark.pandas.groupby.GroupBy.ffill
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.head.html">
pyspark.pandas.groupby.GroupBy.head
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.backfill.html">
pyspark.pandas.groupby.GroupBy.backfill
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.shift.html">
pyspark.pandas.groupby.GroupBy.shift
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.GroupBy.tail.html">
pyspark.pandas.groupby.GroupBy.tail
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.DataFrameGroupBy.describe.html">
pyspark.pandas.groupby.DataFrameGroupBy.describe
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.SeriesGroupBy.nsmallest.html">
pyspark.pandas.groupby.SeriesGroupBy.nsmallest
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.SeriesGroupBy.nlargest.html">
pyspark.pandas.groupby.SeriesGroupBy.nlargest
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.SeriesGroupBy.value_counts.html">
pyspark.pandas.groupby.SeriesGroupBy.value_counts
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.groupby.SeriesGroupBy.unique.html">
pyspark.pandas.groupby.SeriesGroupBy.unique
</a>
</li>
</ul>
</li>
<li class="toctree-l2 has-children">
<a class="reference internal" href="pyspark.pandas/ml.html">
Machine Learning utilities
</a>
<input class="toctree-checkbox" id="toctree-checkbox-10" name="toctree-checkbox-10" type="checkbox"/>
<label for="toctree-checkbox-10">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.mlflow.PythonModelWrapper.html">
pyspark.pandas.mlflow.PythonModelWrapper
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.mlflow.load_model.html">
pyspark.pandas.mlflow.load_model
</a>
</li>
</ul>
</li>
<li class="toctree-l2 has-children">
<a class="reference internal" href="pyspark.pandas/extensions.html">
Extensions
</a>
<input class="toctree-checkbox" id="toctree-checkbox-11" name="toctree-checkbox-11" type="checkbox"/>
<label for="toctree-checkbox-11">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.extensions.register_dataframe_accessor.html">
pyspark.pandas.extensions.register_dataframe_accessor
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.extensions.register_series_accessor.html">
pyspark.pandas.extensions.register_series_accessor
</a>
</li>
<li class="toctree-l3">
<a class="reference internal" href="pyspark.pandas/api/pyspark.pandas.extensions.register_index_accessor.html">
pyspark.pandas.extensions.register_index_accessor
</a>
</li>
</ul>
</li>
</ul>
</li>
<li class="toctree-l1 has-children">
<a class="reference internal" href="pyspark.ss.html">
Structured Streaming
</a>
<input class="toctree-checkbox" id="toctree-checkbox-12" name="toctree-checkbox-12" type="checkbox"/>
<label for="toctree-checkbox-12">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamReader.html">
pyspark.sql.streaming.DataStreamReader
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamWriter.html">
pyspark.sql.streaming.DataStreamWriter
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.ForeachBatchFunction.html">
pyspark.sql.streaming.ForeachBatchFunction
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQuery.html">
pyspark.sql.streaming.StreamingQuery
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQueryException.html">
pyspark.sql.streaming.StreamingQueryException
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQueryManager.html">
pyspark.sql.streaming.StreamingQueryManager
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamReader.csv.html">
pyspark.sql.streaming.DataStreamReader.csv
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamReader.format.html">
pyspark.sql.streaming.DataStreamReader.format
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamReader.json.html">
pyspark.sql.streaming.DataStreamReader.json
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamReader.load.html">
pyspark.sql.streaming.DataStreamReader.load
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamReader.option.html">
pyspark.sql.streaming.DataStreamReader.option
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamReader.options.html">
pyspark.sql.streaming.DataStreamReader.options
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamReader.orc.html">
pyspark.sql.streaming.DataStreamReader.orc
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamReader.parquet.html">
pyspark.sql.streaming.DataStreamReader.parquet
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamReader.schema.html">
pyspark.sql.streaming.DataStreamReader.schema
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamReader.text.html">
pyspark.sql.streaming.DataStreamReader.text
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamWriter.foreach.html">
pyspark.sql.streaming.DataStreamWriter.foreach
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamWriter.foreachBatch.html">
pyspark.sql.streaming.DataStreamWriter.foreachBatch
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamWriter.format.html">
pyspark.sql.streaming.DataStreamWriter.format
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamWriter.option.html">
pyspark.sql.streaming.DataStreamWriter.option
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamWriter.options.html">
pyspark.sql.streaming.DataStreamWriter.options
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamWriter.outputMode.html">
pyspark.sql.streaming.DataStreamWriter.outputMode
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamWriter.partitionBy.html">
pyspark.sql.streaming.DataStreamWriter.partitionBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamWriter.queryName.html">
pyspark.sql.streaming.DataStreamWriter.queryName
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamWriter.start.html">
pyspark.sql.streaming.DataStreamWriter.start
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.DataStreamWriter.trigger.html">
pyspark.sql.streaming.DataStreamWriter.trigger
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQuery.awaitTermination.html">
pyspark.sql.streaming.StreamingQuery.awaitTermination
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQuery.exception.html">
pyspark.sql.streaming.StreamingQuery.exception
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQuery.explain.html">
pyspark.sql.streaming.StreamingQuery.explain
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQuery.id.html">
pyspark.sql.streaming.StreamingQuery.id
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQuery.isActive.html">
pyspark.sql.streaming.StreamingQuery.isActive
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQuery.lastProgress.html">
pyspark.sql.streaming.StreamingQuery.lastProgress
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQuery.name.html">
pyspark.sql.streaming.StreamingQuery.name
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQuery.processAllAvailable.html">
pyspark.sql.streaming.StreamingQuery.processAllAvailable
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQuery.recentProgress.html">
pyspark.sql.streaming.StreamingQuery.recentProgress
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQuery.runId.html">
pyspark.sql.streaming.StreamingQuery.runId
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQuery.status.html">
pyspark.sql.streaming.StreamingQuery.status
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQuery.stop.html">
pyspark.sql.streaming.StreamingQuery.stop
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQueryManager.active.html">
pyspark.sql.streaming.StreamingQueryManager.active
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQueryManager.awaitAnyTermination.html">
pyspark.sql.streaming.StreamingQueryManager.awaitAnyTermination
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQueryManager.get.html">
pyspark.sql.streaming.StreamingQueryManager.get
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.sql.streaming.StreamingQueryManager.resetTerminated.html">
pyspark.sql.streaming.StreamingQueryManager.resetTerminated
</a>
</li>
</ul>
</li>
<li class="toctree-l1 has-children">
<a class="reference internal" href="pyspark.ml.html">
MLlib (DataFrame-based)
</a>
<input class="toctree-checkbox" id="toctree-checkbox-13" name="toctree-checkbox-13" type="checkbox"/>
<label for="toctree-checkbox-13">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.Transformer.html">
Transformer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.UnaryTransformer.html">
UnaryTransformer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.Estimator.html">
Estimator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.Model.html">
Model
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.Predictor.html">
Predictor
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.PredictionModel.html">
PredictionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.Pipeline.html">
Pipeline
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.PipelineModel.html">
PipelineModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.param.Param.html">
Param
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.param.Params.html">
Params
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.param.TypeConverters.html">
TypeConverters
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.Binarizer.html">
Binarizer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.BucketedRandomProjectionLSH.html">
BucketedRandomProjectionLSH
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.BucketedRandomProjectionLSHModel.html">
BucketedRandomProjectionLSHModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.Bucketizer.html">
Bucketizer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.ChiSqSelector.html">
ChiSqSelector
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.ChiSqSelectorModel.html">
ChiSqSelectorModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.CountVectorizer.html">
CountVectorizer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.CountVectorizerModel.html">
CountVectorizerModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.DCT.html">
DCT
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.ElementwiseProduct.html">
ElementwiseProduct
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.FeatureHasher.html">
FeatureHasher
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.HashingTF.html">
HashingTF
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.IDF.html">
IDF
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.IDFModel.html">
IDFModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.Imputer.html">
Imputer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.ImputerModel.html">
ImputerModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.IndexToString.html">
IndexToString
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.Interaction.html">
Interaction
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.MaxAbsScaler.html">
MaxAbsScaler
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.MaxAbsScalerModel.html">
MaxAbsScalerModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.MinHashLSH.html">
MinHashLSH
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.MinHashLSHModel.html">
MinHashLSHModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.MinMaxScaler.html">
MinMaxScaler
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.MinMaxScalerModel.html">
MinMaxScalerModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.NGram.html">
NGram
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.Normalizer.html">
Normalizer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.OneHotEncoder.html">
OneHotEncoder
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.OneHotEncoderModel.html">
OneHotEncoderModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.PCA.html">
PCA
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.PCAModel.html">
PCAModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.PolynomialExpansion.html">
PolynomialExpansion
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.QuantileDiscretizer.html">
QuantileDiscretizer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.RobustScaler.html">
RobustScaler
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.RobustScalerModel.html">
RobustScalerModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.RegexTokenizer.html">
RegexTokenizer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.RFormula.html">
RFormula
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.RFormulaModel.html">
RFormulaModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.SQLTransformer.html">
SQLTransformer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.StandardScaler.html">
StandardScaler
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.StandardScalerModel.html">
StandardScalerModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.StopWordsRemover.html">
StopWordsRemover
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.StringIndexer.html">
StringIndexer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.StringIndexerModel.html">
StringIndexerModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.Tokenizer.html">
Tokenizer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.UnivariateFeatureSelector.html">
UnivariateFeatureSelector
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.UnivariateFeatureSelectorModel.html">
UnivariateFeatureSelectorModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.VarianceThresholdSelector.html">
VarianceThresholdSelector
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.VarianceThresholdSelectorModel.html">
VarianceThresholdSelectorModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.VectorAssembler.html">
VectorAssembler
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.VectorIndexer.html">
VectorIndexer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.VectorIndexerModel.html">
VectorIndexerModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.VectorSizeHint.html">
VectorSizeHint
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.VectorSlicer.html">
VectorSlicer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.Word2Vec.html">
Word2Vec
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.feature.Word2VecModel.html">
Word2VecModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.LinearSVC.html">
LinearSVC
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.LinearSVCModel.html">
LinearSVCModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.LinearSVCSummary.html">
LinearSVCSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.LinearSVCTrainingSummary.html">
LinearSVCTrainingSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.LogisticRegression.html">
LogisticRegression
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.LogisticRegressionModel.html">
LogisticRegressionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.LogisticRegressionSummary.html">
LogisticRegressionSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.LogisticRegressionTrainingSummary.html">
LogisticRegressionTrainingSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.BinaryLogisticRegressionSummary.html">
BinaryLogisticRegressionSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.BinaryLogisticRegressionTrainingSummary.html">
BinaryLogisticRegressionTrainingSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.DecisionTreeClassifier.html">
DecisionTreeClassifier
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.DecisionTreeClassificationModel.html">
DecisionTreeClassificationModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.GBTClassifier.html">
GBTClassifier
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.GBTClassificationModel.html">
GBTClassificationModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.RandomForestClassifier.html">
RandomForestClassifier
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.RandomForestClassificationModel.html">
RandomForestClassificationModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.RandomForestClassificationSummary.html">
RandomForestClassificationSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.RandomForestClassificationTrainingSummary.html">
RandomForestClassificationTrainingSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.BinaryRandomForestClassificationSummary.html">
BinaryRandomForestClassificationSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.BinaryRandomForestClassificationTrainingSummary.html">
BinaryRandomForestClassificationTrainingSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.NaiveBayes.html">
NaiveBayes
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.NaiveBayesModel.html">
NaiveBayesModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.MultilayerPerceptronClassifier.html">
MultilayerPerceptronClassifier
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.MultilayerPerceptronClassificationModel.html">
MultilayerPerceptronClassificationModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.MultilayerPerceptronClassificationSummary.html">
MultilayerPerceptronClassificationSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.MultilayerPerceptronClassificationTrainingSummary.html">
MultilayerPerceptronClassificationTrainingSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.OneVsRest.html">
OneVsRest
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.OneVsRestModel.html">
OneVsRestModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.FMClassifier.html">
FMClassifier
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.FMClassificationModel.html">
FMClassificationModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.FMClassificationSummary.html">
FMClassificationSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.classification.FMClassificationTrainingSummary.html">
FMClassificationTrainingSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.BisectingKMeans.html">
BisectingKMeans
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.BisectingKMeansModel.html">
BisectingKMeansModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.BisectingKMeansSummary.html">
BisectingKMeansSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.KMeans.html">
KMeans
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.KMeansModel.html">
KMeansModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.KMeansSummary.html">
KMeansSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.GaussianMixture.html">
GaussianMixture
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.GaussianMixtureModel.html">
GaussianMixtureModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.GaussianMixtureSummary.html">
GaussianMixtureSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.LDA.html">
LDA
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.LDAModel.html">
LDAModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.LocalLDAModel.html">
LocalLDAModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.DistributedLDAModel.html">
DistributedLDAModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.clustering.PowerIterationClustering.html">
PowerIterationClustering
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.functions.array_to_vector.html">
pyspark.ml.functions.array_to_vector
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.functions.vector_to_array.html">
pyspark.ml.functions.vector_to_array
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.linalg.Vector.html">
Vector
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.linalg.DenseVector.html">
DenseVector
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.linalg.SparseVector.html">
SparseVector
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.linalg.Vectors.html">
Vectors
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.linalg.Matrix.html">
Matrix
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.linalg.DenseMatrix.html">
DenseMatrix
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.linalg.SparseMatrix.html">
SparseMatrix
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.linalg.Matrices.html">
Matrices
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.recommendation.ALS.html">
ALS
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.recommendation.ALSModel.html">
ALSModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.AFTSurvivalRegression.html">
AFTSurvivalRegression
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.AFTSurvivalRegressionModel.html">
AFTSurvivalRegressionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.DecisionTreeRegressor.html">
DecisionTreeRegressor
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.DecisionTreeRegressionModel.html">
DecisionTreeRegressionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.GBTRegressor.html">
GBTRegressor
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.GBTRegressionModel.html">
GBTRegressionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.GeneralizedLinearRegression.html">
GeneralizedLinearRegression
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.GeneralizedLinearRegressionModel.html">
GeneralizedLinearRegressionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.GeneralizedLinearRegressionSummary.html">
GeneralizedLinearRegressionSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.GeneralizedLinearRegressionTrainingSummary.html">
GeneralizedLinearRegressionTrainingSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.IsotonicRegression.html">
IsotonicRegression
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.IsotonicRegressionModel.html">
IsotonicRegressionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.LinearRegression.html">
LinearRegression
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.LinearRegressionModel.html">
LinearRegressionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.LinearRegressionSummary.html">
LinearRegressionSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.LinearRegressionTrainingSummary.html">
LinearRegressionTrainingSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.RandomForestRegressor.html">
RandomForestRegressor
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.RandomForestRegressionModel.html">
RandomForestRegressionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.FMRegressor.html">
FMRegressor
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.regression.FMRegressionModel.html">
FMRegressionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.stat.ChiSquareTest.html">
ChiSquareTest
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.stat.Correlation.html">
Correlation
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.stat.KolmogorovSmirnovTest.html">
KolmogorovSmirnovTest
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.stat.MultivariateGaussian.html">
MultivariateGaussian
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.stat.Summarizer.html">
Summarizer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.stat.SummaryBuilder.html">
SummaryBuilder
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.tuning.ParamGridBuilder.html">
ParamGridBuilder
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.tuning.CrossValidator.html">
CrossValidator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.tuning.CrossValidatorModel.html">
CrossValidatorModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.tuning.TrainValidationSplit.html">
TrainValidationSplit
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.tuning.TrainValidationSplitModel.html">
TrainValidationSplitModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.evaluation.Evaluator.html">
Evaluator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.evaluation.BinaryClassificationEvaluator.html">
BinaryClassificationEvaluator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.evaluation.RegressionEvaluator.html">
RegressionEvaluator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.evaluation.MulticlassClassificationEvaluator.html">
MulticlassClassificationEvaluator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.evaluation.MultilabelClassificationEvaluator.html">
MultilabelClassificationEvaluator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.evaluation.ClusteringEvaluator.html">
ClusteringEvaluator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.evaluation.RankingEvaluator.html">
RankingEvaluator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.fpm.FPGrowth.html">
FPGrowth
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.fpm.FPGrowthModel.html">
FPGrowthModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.fpm.PrefixSpan.html">
PrefixSpan
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.image.ImageSchema.html">
ImageSchema
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.image._ImageSchema.html">
_ImageSchema
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.util.BaseReadWrite.html">
BaseReadWrite
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.util.DefaultParamsReadable.html">
DefaultParamsReadable
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.util.DefaultParamsReader.html">
DefaultParamsReader
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.util.DefaultParamsWritable.html">
DefaultParamsWritable
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.util.DefaultParamsWriter.html">
DefaultParamsWriter
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.util.GeneralMLWriter.html">
GeneralMLWriter
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.util.HasTrainingSummary.html">
HasTrainingSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.util.Identifiable.html">
Identifiable
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.util.MLReadable.html">
MLReadable
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.util.MLReader.html">
MLReader
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.util.MLWritable.html">
MLWritable
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.ml.util.MLWriter.html">
MLWriter
</a>
</li>
</ul>
</li>
<li class="toctree-l1 has-children">
<a class="reference internal" href="pyspark.streaming.html">
Spark Streaming
</a>
<input class="toctree-checkbox" id="toctree-checkbox-14" name="toctree-checkbox-14" type="checkbox"/>
<label for="toctree-checkbox-14">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.html">
pyspark.streaming.StreamingContext
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.html">
pyspark.streaming.DStream
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.addStreamingListener.html">
pyspark.streaming.StreamingContext.addStreamingListener
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.awaitTermination.html">
pyspark.streaming.StreamingContext.awaitTermination
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.awaitTerminationOrTimeout.html">
pyspark.streaming.StreamingContext.awaitTerminationOrTimeout
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.checkpoint.html">
pyspark.streaming.StreamingContext.checkpoint
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.getActive.html">
pyspark.streaming.StreamingContext.getActive
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.getActiveOrCreate.html">
pyspark.streaming.StreamingContext.getActiveOrCreate
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.getOrCreate.html">
pyspark.streaming.StreamingContext.getOrCreate
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.remember.html">
pyspark.streaming.StreamingContext.remember
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.sparkContext.html">
pyspark.streaming.StreamingContext.sparkContext
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.start.html">
pyspark.streaming.StreamingContext.start
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.stop.html">
pyspark.streaming.StreamingContext.stop
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.transform.html">
pyspark.streaming.StreamingContext.transform
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.union.html">
pyspark.streaming.StreamingContext.union
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.binaryRecordsStream.html">
pyspark.streaming.StreamingContext.binaryRecordsStream
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.queueStream.html">
pyspark.streaming.StreamingContext.queueStream
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.socketTextStream.html">
pyspark.streaming.StreamingContext.socketTextStream
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.StreamingContext.textFileStream.html">
pyspark.streaming.StreamingContext.textFileStream
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.pprint.html">
pyspark.streaming.DStream.pprint
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.saveAsTextFiles.html">
pyspark.streaming.DStream.saveAsTextFiles
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.cache.html">
pyspark.streaming.DStream.cache
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.checkpoint.html">
pyspark.streaming.DStream.checkpoint
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.cogroup.html">
pyspark.streaming.DStream.cogroup
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.combineByKey.html">
pyspark.streaming.DStream.combineByKey
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.context.html">
pyspark.streaming.DStream.context
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.count.html">
pyspark.streaming.DStream.count
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.countByValue.html">
pyspark.streaming.DStream.countByValue
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.countByValueAndWindow.html">
pyspark.streaming.DStream.countByValueAndWindow
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.countByWindow.html">
pyspark.streaming.DStream.countByWindow
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.filter.html">
pyspark.streaming.DStream.filter
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.flatMap.html">
pyspark.streaming.DStream.flatMap
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.flatMapValues.html">
pyspark.streaming.DStream.flatMapValues
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.foreachRDD.html">
pyspark.streaming.DStream.foreachRDD
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.fullOuterJoin.html">
pyspark.streaming.DStream.fullOuterJoin
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.glom.html">
pyspark.streaming.DStream.glom
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.groupByKey.html">
pyspark.streaming.DStream.groupByKey
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.groupByKeyAndWindow.html">
pyspark.streaming.DStream.groupByKeyAndWindow
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.join.html">
pyspark.streaming.DStream.join
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.leftOuterJoin.html">
pyspark.streaming.DStream.leftOuterJoin
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.map.html">
pyspark.streaming.DStream.map
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.mapPartitions.html">
pyspark.streaming.DStream.mapPartitions
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.mapPartitionsWithIndex.html">
pyspark.streaming.DStream.mapPartitionsWithIndex
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.mapValues.html">
pyspark.streaming.DStream.mapValues
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.partitionBy.html">
pyspark.streaming.DStream.partitionBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.persist.html">
pyspark.streaming.DStream.persist
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.reduce.html">
pyspark.streaming.DStream.reduce
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.reduceByKey.html">
pyspark.streaming.DStream.reduceByKey
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.reduceByKeyAndWindow.html">
pyspark.streaming.DStream.reduceByKeyAndWindow
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.reduceByWindow.html">
pyspark.streaming.DStream.reduceByWindow
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.repartition.html">
pyspark.streaming.DStream.repartition
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.rightOuterJoin.html">
pyspark.streaming.DStream.rightOuterJoin
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.slice.html">
pyspark.streaming.DStream.slice
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.transform.html">
pyspark.streaming.DStream.transform
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.transformWith.html">
pyspark.streaming.DStream.transformWith
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.union.html">
pyspark.streaming.DStream.union
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.updateStateByKey.html">
pyspark.streaming.DStream.updateStateByKey
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.DStream.window.html">
pyspark.streaming.DStream.window
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.kinesis.KinesisUtils.createStream.html">
pyspark.streaming.kinesis.KinesisUtils.createStream
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.kinesis.InitialPositionInStream.LATEST.html">
pyspark.streaming.kinesis.InitialPositionInStream.LATEST
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.streaming.kinesis.InitialPositionInStream.TRIM_HORIZON.html">
pyspark.streaming.kinesis.InitialPositionInStream.TRIM_HORIZON
</a>
</li>
</ul>
</li>
<li class="toctree-l1 has-children">
<a class="reference internal" href="pyspark.mllib.html">
MLlib (RDD-based)
</a>
<input class="toctree-checkbox" id="toctree-checkbox-15" name="toctree-checkbox-15" type="checkbox"/>
<label for="toctree-checkbox-15">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.classification.LogisticRegressionModel.html">
LogisticRegressionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.classification.LogisticRegressionWithSGD.html">
LogisticRegressionWithSGD
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.classification.LogisticRegressionWithLBFGS.html">
LogisticRegressionWithLBFGS
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.classification.SVMModel.html">
SVMModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.classification.SVMWithSGD.html">
SVMWithSGD
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.classification.NaiveBayesModel.html">
NaiveBayesModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.classification.NaiveBayes.html">
NaiveBayes
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.classification.StreamingLogisticRegressionWithSGD.html">
StreamingLogisticRegressionWithSGD
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.clustering.BisectingKMeansModel.html">
BisectingKMeansModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.clustering.BisectingKMeans.html">
BisectingKMeans
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.clustering.KMeansModel.html">
KMeansModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.clustering.KMeans.html">
KMeans
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.clustering.GaussianMixtureModel.html">
GaussianMixtureModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.clustering.GaussianMixture.html">
GaussianMixture
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.clustering.PowerIterationClusteringModel.html">
PowerIterationClusteringModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.clustering.PowerIterationClustering.html">
PowerIterationClustering
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.clustering.StreamingKMeans.html">
StreamingKMeans
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.clustering.StreamingKMeansModel.html">
StreamingKMeansModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.clustering.LDA.html">
LDA
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.clustering.LDAModel.html">
LDAModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.evaluation.BinaryClassificationMetrics.html">
BinaryClassificationMetrics
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.evaluation.RegressionMetrics.html">
RegressionMetrics
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.evaluation.MulticlassMetrics.html">
MulticlassMetrics
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.evaluation.RankingMetrics.html">
RankingMetrics
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.feature.Normalizer.html">
Normalizer
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.feature.StandardScalerModel.html">
StandardScalerModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.feature.StandardScaler.html">
StandardScaler
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.feature.HashingTF.html">
HashingTF
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.feature.IDFModel.html">
IDFModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.feature.IDF.html">
IDF
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.feature.Word2Vec.html">
Word2Vec
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.feature.Word2VecModel.html">
Word2VecModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.feature.ChiSqSelector.html">
ChiSqSelector
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.feature.ChiSqSelectorModel.html">
ChiSqSelectorModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.feature.ElementwiseProduct.html">
ElementwiseProduct
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.fpm.FPGrowth.html">
FPGrowth
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.fpm.FPGrowthModel.html">
FPGrowthModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.fpm.PrefixSpan.html">
PrefixSpan
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.fpm.PrefixSpanModel.html">
PrefixSpanModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.Vector.html">
Vector
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.DenseVector.html">
DenseVector
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.SparseVector.html">
SparseVector
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.Vectors.html">
Vectors
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.Matrix.html">
Matrix
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.DenseMatrix.html">
DenseMatrix
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.SparseMatrix.html">
SparseMatrix
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.Matrices.html">
Matrices
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.QRDecomposition.html">
QRDecomposition
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.distributed.BlockMatrix.html">
BlockMatrix
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.distributed.CoordinateMatrix.html">
CoordinateMatrix
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.distributed.DistributedMatrix.html">
DistributedMatrix
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.distributed.IndexedRow.html">
IndexedRow
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.distributed.IndexedRowMatrix.html">
IndexedRowMatrix
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.distributed.MatrixEntry.html">
MatrixEntry
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.distributed.RowMatrix.html">
RowMatrix
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.linalg.distributed.SingularValueDecomposition.html">
SingularValueDecomposition
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.random.RandomRDDs.html">
RandomRDDs
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.recommendation.MatrixFactorizationModel.html">
MatrixFactorizationModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.recommendation.ALS.html">
ALS
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.recommendation.Rating.html">
Rating
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.regression.LabeledPoint.html">
LabeledPoint
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.regression.LinearModel.html">
LinearModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.regression.LinearRegressionModel.html">
LinearRegressionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.regression.LinearRegressionWithSGD.html">
LinearRegressionWithSGD
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.regression.RidgeRegressionModel.html">
RidgeRegressionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.regression.RidgeRegressionWithSGD.html">
RidgeRegressionWithSGD
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.regression.LassoModel.html">
LassoModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.regression.LassoWithSGD.html">
LassoWithSGD
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.regression.IsotonicRegressionModel.html">
IsotonicRegressionModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.regression.IsotonicRegression.html">
IsotonicRegression
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.regression.StreamingLinearAlgorithm.html">
StreamingLinearAlgorithm
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.regression.StreamingLinearRegressionWithSGD.html">
StreamingLinearRegressionWithSGD
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.stat.Statistics.html">
Statistics
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.stat.MultivariateStatisticalSummary.html">
MultivariateStatisticalSummary
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.stat.ChiSqTestResult.html">
ChiSqTestResult
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.stat.MultivariateGaussian.html">
MultivariateGaussian
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.stat.KernelDensity.html">
KernelDensity
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.stat.ChiSqTestResult.html">
ChiSqTestResult
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.stat.KolmogorovSmirnovTestResult.html">
KolmogorovSmirnovTestResult
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.tree.DecisionTreeModel.html">
DecisionTreeModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.tree.DecisionTree.html">
DecisionTree
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.tree.RandomForestModel.html">
RandomForestModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.tree.RandomForest.html">
RandomForest
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.tree.GradientBoostedTreesModel.html">
GradientBoostedTreesModel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.tree.GradientBoostedTrees.html">
GradientBoostedTrees
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.util.JavaLoader.html">
JavaLoader
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.util.JavaSaveable.html">
JavaSaveable
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.util.LinearDataGenerator.html">
LinearDataGenerator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.util.Loader.html">
Loader
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.util.MLUtils.html">
MLUtils
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.mllib.util.Saveable.html">
Saveable
</a>
</li>
</ul>
</li>
<li class="toctree-l1 has-children">
<a class="reference internal" href="pyspark.html">
Spark Core
</a>
<input class="toctree-checkbox" id="toctree-checkbox-16" name="toctree-checkbox-16" type="checkbox"/>
<label for="toctree-checkbox-16">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.html">
pyspark.SparkContext
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.html">
pyspark.RDD
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.Broadcast.html">
pyspark.Broadcast
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.Accumulator.html">
pyspark.Accumulator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.AccumulatorParam.html">
pyspark.AccumulatorParam
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkConf.html">
pyspark.SparkConf
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkFiles.html">
pyspark.SparkFiles
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.StorageLevel.html">
pyspark.StorageLevel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.TaskContext.html">
pyspark.TaskContext
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDDBarrier.html">
pyspark.RDDBarrier
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.BarrierTaskContext.html">
pyspark.BarrierTaskContext
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.BarrierTaskInfo.html">
pyspark.BarrierTaskInfo
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.InheritableThread.html">
pyspark.InheritableThread
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.util.VersionUtils.html">
pyspark.util.VersionUtils
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.PACKAGE_EXTENSIONS.html">
pyspark.SparkContext.PACKAGE_EXTENSIONS
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.accumulator.html">
pyspark.SparkContext.accumulator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.addFile.html">
pyspark.SparkContext.addFile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.addPyFile.html">
pyspark.SparkContext.addPyFile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.applicationId.html">
pyspark.SparkContext.applicationId
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.binaryFiles.html">
pyspark.SparkContext.binaryFiles
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.binaryRecords.html">
pyspark.SparkContext.binaryRecords
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.broadcast.html">
pyspark.SparkContext.broadcast
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.cancelAllJobs.html">
pyspark.SparkContext.cancelAllJobs
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.cancelJobGroup.html">
pyspark.SparkContext.cancelJobGroup
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.defaultMinPartitions.html">
pyspark.SparkContext.defaultMinPartitions
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.defaultParallelism.html">
pyspark.SparkContext.defaultParallelism
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.dump_profiles.html">
pyspark.SparkContext.dump_profiles
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.emptyRDD.html">
pyspark.SparkContext.emptyRDD
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.getCheckpointDir.html">
pyspark.SparkContext.getCheckpointDir
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.getConf.html">
pyspark.SparkContext.getConf
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.getLocalProperty.html">
pyspark.SparkContext.getLocalProperty
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.getOrCreate.html">
pyspark.SparkContext.getOrCreate
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.hadoopFile.html">
pyspark.SparkContext.hadoopFile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.hadoopRDD.html">
pyspark.SparkContext.hadoopRDD
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.newAPIHadoopFile.html">
pyspark.SparkContext.newAPIHadoopFile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.newAPIHadoopRDD.html">
pyspark.SparkContext.newAPIHadoopRDD
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.parallelize.html">
pyspark.SparkContext.parallelize
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.pickleFile.html">
pyspark.SparkContext.pickleFile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.range.html">
pyspark.SparkContext.range
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.resources.html">
pyspark.SparkContext.resources
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.runJob.html">
pyspark.SparkContext.runJob
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.sequenceFile.html">
pyspark.SparkContext.sequenceFile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.setCheckpointDir.html">
pyspark.SparkContext.setCheckpointDir
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.setJobDescription.html">
pyspark.SparkContext.setJobDescription
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.setJobGroup.html">
pyspark.SparkContext.setJobGroup
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.setLocalProperty.html">
pyspark.SparkContext.setLocalProperty
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.setLogLevel.html">
pyspark.SparkContext.setLogLevel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.setSystemProperty.html">
pyspark.SparkContext.setSystemProperty
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.show_profiles.html">
pyspark.SparkContext.show_profiles
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.sparkUser.html">
pyspark.SparkContext.sparkUser
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.startTime.html">
pyspark.SparkContext.startTime
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.statusTracker.html">
pyspark.SparkContext.statusTracker
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.stop.html">
pyspark.SparkContext.stop
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.textFile.html">
pyspark.SparkContext.textFile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.uiWebUrl.html">
pyspark.SparkContext.uiWebUrl
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.union.html">
pyspark.SparkContext.union
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.version.html">
pyspark.SparkContext.version
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkContext.wholeTextFiles.html">
pyspark.SparkContext.wholeTextFiles
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.aggregate.html">
pyspark.RDD.aggregate
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.aggregateByKey.html">
pyspark.RDD.aggregateByKey
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.barrier.html">
pyspark.RDD.barrier
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.cache.html">
pyspark.RDD.cache
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.cartesian.html">
pyspark.RDD.cartesian
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.checkpoint.html">
pyspark.RDD.checkpoint
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.coalesce.html">
pyspark.RDD.coalesce
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.cogroup.html">
pyspark.RDD.cogroup
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.collect.html">
pyspark.RDD.collect
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.collectAsMap.html">
pyspark.RDD.collectAsMap
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.collectWithJobGroup.html">
pyspark.RDD.collectWithJobGroup
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.combineByKey.html">
pyspark.RDD.combineByKey
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.context.html">
pyspark.RDD.context
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.count.html">
pyspark.RDD.count
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.countApprox.html">
pyspark.RDD.countApprox
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.countApproxDistinct.html">
pyspark.RDD.countApproxDistinct
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.countByKey.html">
pyspark.RDD.countByKey
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.countByValue.html">
pyspark.RDD.countByValue
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.distinct.html">
pyspark.RDD.distinct
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.filter.html">
pyspark.RDD.filter
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.first.html">
pyspark.RDD.first
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.flatMap.html">
pyspark.RDD.flatMap
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.flatMapValues.html">
pyspark.RDD.flatMapValues
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.fold.html">
pyspark.RDD.fold
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.foldByKey.html">
pyspark.RDD.foldByKey
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.foreach.html">
pyspark.RDD.foreach
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.foreachPartition.html">
pyspark.RDD.foreachPartition
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.fullOuterJoin.html">
pyspark.RDD.fullOuterJoin
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.getCheckpointFile.html">
pyspark.RDD.getCheckpointFile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.getNumPartitions.html">
pyspark.RDD.getNumPartitions
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.getResourceProfile.html">
pyspark.RDD.getResourceProfile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.getStorageLevel.html">
pyspark.RDD.getStorageLevel
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.glom.html">
pyspark.RDD.glom
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.groupBy.html">
pyspark.RDD.groupBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.groupByKey.html">
pyspark.RDD.groupByKey
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.groupWith.html">
pyspark.RDD.groupWith
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.histogram.html">
pyspark.RDD.histogram
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.id.html">
pyspark.RDD.id
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.intersection.html">
pyspark.RDD.intersection
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.isCheckpointed.html">
pyspark.RDD.isCheckpointed
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.isEmpty.html">
pyspark.RDD.isEmpty
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.isLocallyCheckpointed.html">
pyspark.RDD.isLocallyCheckpointed
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.join.html">
pyspark.RDD.join
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.keyBy.html">
pyspark.RDD.keyBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.keys.html">
pyspark.RDD.keys
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.leftOuterJoin.html">
pyspark.RDD.leftOuterJoin
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.localCheckpoint.html">
pyspark.RDD.localCheckpoint
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.lookup.html">
pyspark.RDD.lookup
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.map.html">
pyspark.RDD.map
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.mapPartitions.html">
pyspark.RDD.mapPartitions
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.mapPartitionsWithIndex.html">
pyspark.RDD.mapPartitionsWithIndex
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.mapPartitionsWithSplit.html">
pyspark.RDD.mapPartitionsWithSplit
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.mapValues.html">
pyspark.RDD.mapValues
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.max.html">
pyspark.RDD.max
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.mean.html">
pyspark.RDD.mean
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.meanApprox.html">
pyspark.RDD.meanApprox
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.min.html">
pyspark.RDD.min
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.name.html">
pyspark.RDD.name
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.partitionBy.html">
pyspark.RDD.partitionBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.persist.html">
pyspark.RDD.persist
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.pipe.html">
pyspark.RDD.pipe
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.randomSplit.html">
pyspark.RDD.randomSplit
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.reduce.html">
pyspark.RDD.reduce
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.reduceByKey.html">
pyspark.RDD.reduceByKey
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.reduceByKeyLocally.html">
pyspark.RDD.reduceByKeyLocally
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.repartition.html">
pyspark.RDD.repartition
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.repartitionAndSortWithinPartitions.html">
pyspark.RDD.repartitionAndSortWithinPartitions
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.rightOuterJoin.html">
pyspark.RDD.rightOuterJoin
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.sample.html">
pyspark.RDD.sample
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.sampleByKey.html">
pyspark.RDD.sampleByKey
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.sampleStdev.html">
pyspark.RDD.sampleStdev
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.sampleVariance.html">
pyspark.RDD.sampleVariance
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.saveAsHadoopDataset.html">
pyspark.RDD.saveAsHadoopDataset
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.saveAsHadoopFile.html">
pyspark.RDD.saveAsHadoopFile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.saveAsNewAPIHadoopDataset.html">
pyspark.RDD.saveAsNewAPIHadoopDataset
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.saveAsNewAPIHadoopFile.html">
pyspark.RDD.saveAsNewAPIHadoopFile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.saveAsPickleFile.html">
pyspark.RDD.saveAsPickleFile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.saveAsSequenceFile.html">
pyspark.RDD.saveAsSequenceFile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.saveAsTextFile.html">
pyspark.RDD.saveAsTextFile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.setName.html">
pyspark.RDD.setName
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.sortBy.html">
pyspark.RDD.sortBy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.sortByKey.html">
pyspark.RDD.sortByKey
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.stats.html">
pyspark.RDD.stats
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.stdev.html">
pyspark.RDD.stdev
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.subtract.html">
pyspark.RDD.subtract
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.subtractByKey.html">
pyspark.RDD.subtractByKey
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.sum.html">
pyspark.RDD.sum
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.sumApprox.html">
pyspark.RDD.sumApprox
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.take.html">
pyspark.RDD.take
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.takeOrdered.html">
pyspark.RDD.takeOrdered
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.takeSample.html">
pyspark.RDD.takeSample
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.toDebugString.html">
pyspark.RDD.toDebugString
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.toLocalIterator.html">
pyspark.RDD.toLocalIterator
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.top.html">
pyspark.RDD.top
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.treeAggregate.html">
pyspark.RDD.treeAggregate
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.treeReduce.html">
pyspark.RDD.treeReduce
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.union.html">
pyspark.RDD.union
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.unpersist.html">
pyspark.RDD.unpersist
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.values.html">
pyspark.RDD.values
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.variance.html">
pyspark.RDD.variance
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.withResources.html">
pyspark.RDD.withResources
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.zip.html">
pyspark.RDD.zip
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.zipWithIndex.html">
pyspark.RDD.zipWithIndex
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDD.zipWithUniqueId.html">
pyspark.RDD.zipWithUniqueId
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.Broadcast.destroy.html">
pyspark.Broadcast.destroy
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.Broadcast.dump.html">
pyspark.Broadcast.dump
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.Broadcast.load.html">
pyspark.Broadcast.load
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.Broadcast.load_from_path.html">
pyspark.Broadcast.load_from_path
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.Broadcast.unpersist.html">
pyspark.Broadcast.unpersist
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.Broadcast.value.html">
pyspark.Broadcast.value
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.Accumulator.add.html">
pyspark.Accumulator.add
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.Accumulator.value.html">
pyspark.Accumulator.value
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.AccumulatorParam.addInPlace.html">
pyspark.AccumulatorParam.addInPlace
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.AccumulatorParam.zero.html">
pyspark.AccumulatorParam.zero
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.inheritable_thread_target.html">
pyspark.inheritable_thread_target
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkConf.contains.html">
pyspark.SparkConf.contains
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkConf.get.html">
pyspark.SparkConf.get
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkConf.getAll.html">
pyspark.SparkConf.getAll
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkConf.set.html">
pyspark.SparkConf.set
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkConf.setAll.html">
pyspark.SparkConf.setAll
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkConf.setAppName.html">
pyspark.SparkConf.setAppName
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkConf.setExecutorEnv.html">
pyspark.SparkConf.setExecutorEnv
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkConf.setIfMissing.html">
pyspark.SparkConf.setIfMissing
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkConf.setMaster.html">
pyspark.SparkConf.setMaster
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkConf.setSparkHome.html">
pyspark.SparkConf.setSparkHome
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkConf.toDebugString.html">
pyspark.SparkConf.toDebugString
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkFiles.get.html">
pyspark.SparkFiles.get
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.SparkFiles.getRootDirectory.html">
pyspark.SparkFiles.getRootDirectory
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.StorageLevel.DISK_ONLY.html">
pyspark.StorageLevel.DISK_ONLY
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.StorageLevel.DISK_ONLY_2.html">
pyspark.StorageLevel.DISK_ONLY_2
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.StorageLevel.DISK_ONLY_3.html">
pyspark.StorageLevel.DISK_ONLY_3
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.StorageLevel.MEMORY_AND_DISK.html">
pyspark.StorageLevel.MEMORY_AND_DISK
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.StorageLevel.MEMORY_AND_DISK_2.html">
pyspark.StorageLevel.MEMORY_AND_DISK_2
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.StorageLevel.MEMORY_ONLY.html">
pyspark.StorageLevel.MEMORY_ONLY
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.StorageLevel.MEMORY_ONLY_2.html">
pyspark.StorageLevel.MEMORY_ONLY_2
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.StorageLevel.OFF_HEAP.html">
pyspark.StorageLevel.OFF_HEAP
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.TaskContext.attemptNumber.html">
pyspark.TaskContext.attemptNumber
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.TaskContext.get.html">
pyspark.TaskContext.get
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.TaskContext.getLocalProperty.html">
pyspark.TaskContext.getLocalProperty
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.TaskContext.partitionId.html">
pyspark.TaskContext.partitionId
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.TaskContext.resources.html">
pyspark.TaskContext.resources
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.TaskContext.stageId.html">
pyspark.TaskContext.stageId
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.TaskContext.taskAttemptId.html">
pyspark.TaskContext.taskAttemptId
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDDBarrier.mapPartitions.html">
pyspark.RDDBarrier.mapPartitions
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.RDDBarrier.mapPartitionsWithIndex.html">
pyspark.RDDBarrier.mapPartitionsWithIndex
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.BarrierTaskContext.allGather.html">
pyspark.BarrierTaskContext.allGather
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.BarrierTaskContext.attemptNumber.html">
pyspark.BarrierTaskContext.attemptNumber
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.BarrierTaskContext.barrier.html">
pyspark.BarrierTaskContext.barrier
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.BarrierTaskContext.get.html">
pyspark.BarrierTaskContext.get
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.BarrierTaskContext.getLocalProperty.html">
pyspark.BarrierTaskContext.getLocalProperty
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.BarrierTaskContext.getTaskInfos.html">
pyspark.BarrierTaskContext.getTaskInfos
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.BarrierTaskContext.partitionId.html">
pyspark.BarrierTaskContext.partitionId
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.BarrierTaskContext.resources.html">
pyspark.BarrierTaskContext.resources
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.BarrierTaskContext.stageId.html">
pyspark.BarrierTaskContext.stageId
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.BarrierTaskContext.taskAttemptId.html">
pyspark.BarrierTaskContext.taskAttemptId
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.util.VersionUtils.majorMinorVersion.html">
pyspark.util.VersionUtils.majorMinorVersion
</a>
</li>
</ul>
</li>
<li class="toctree-l1 has-children">
<a class="reference internal" href="pyspark.resource.html">
Resource Management
</a>
<input class="toctree-checkbox" id="toctree-checkbox-17" name="toctree-checkbox-17" type="checkbox"/>
<label for="toctree-checkbox-17">
<i class="fas fa-chevron-down">
</i>
</label>
<ul>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.resource.ResourceInformation.html">
pyspark.resource.ResourceInformation
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.resource.ResourceProfile.html">
pyspark.resource.ResourceProfile
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.resource.ResourceProfileBuilder.html">
pyspark.resource.ResourceProfileBuilder
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.resource.ExecutorResourceRequest.html">
pyspark.resource.ExecutorResourceRequest
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.resource.ExecutorResourceRequests.html">
pyspark.resource.ExecutorResourceRequests
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.resource.TaskResourceRequest.html">
pyspark.resource.TaskResourceRequest
</a>
</li>
<li class="toctree-l2">
<a class="reference internal" href="api/pyspark.resource.TaskResourceRequests.html">
pyspark.resource.TaskResourceRequests
</a>
</li>
</ul>
</li>
</ul>
</div>
</nav>
</div>
<div class="d-none d-xl-block col-xl-2 bd-toc">
<div class="toc-item">
<div class="tocsection onthispage pt-5 pb-3">
<i class="fas fa-list"></i> On this page
</div>
<nav id="bd-toc-nav">
<ul class="visible nav section-nav flex-column">
<li class="toc-h2 nav-item toc-entry">
<a class="reference internal nav-link" href="#core-classes">
Core Classes
</a>
</li>
<li class="toc-h2 nav-item toc-entry">
<a class="reference internal nav-link" href="#spark-session-apis">
Spark Session APIs
</a>
</li>
<li class="toc-h2 nav-item toc-entry">
<a class="reference internal nav-link" href="#configuration">
Configuration
</a>
</li>
<li class="toc-h2 nav-item toc-entry">
<a class="reference internal nav-link" href="#input-and-output">
Input and Output
</a>
</li>
<li class="toc-h2 nav-item toc-entry">
<a class="reference internal nav-link" href="#dataframe-apis">
DataFrame APIs
</a>
</li>
<li class="toc-h2 nav-item toc-entry">
<a class="reference internal nav-link" href="#column-apis">
Column APIs
</a>
</li>
<li class="toc-h2 nav-item toc-entry">
<a class="reference internal nav-link" href="#data-types">
Data Types
</a>
</li>
<li class="toc-h2 nav-item toc-entry">
<a class="reference internal nav-link" href="#row">
Row
</a>
</li>
<li class="toc-h2 nav-item toc-entry">
<a class="reference internal nav-link" href="#functions">
Functions
</a>
<ul class="nav section-nav flex-column">
</ul>
</li>
<li class="toc-h2 nav-item toc-entry">
<a class="reference internal nav-link" href="#window">
Window
</a>
</li>
<li class="toc-h2 nav-item toc-entry">
<a class="reference internal nav-link" href="#grouping">
Grouping
</a>
</li>
<li class="toc-h2 nav-item toc-entry">
<a class="reference internal nav-link" href="#catalog-apis">
Catalog APIs
</a>
</li>
</ul>
</nav>
</div>
<div class="toc-item">
</div>
</div>
<main class="col-12 col-md-9 col-xl-7 py-md-5 pl-md-5 pr-md-4 bd-content" role="main">
<div>
<section id="spark-sql">
<h1>Spark SQL<a class="headerlink" href="#spark-sql" title="Permalink to this headline"></a></h1>
<section id="core-classes">
<h2>Core Classes<a class="headerlink" href="#core-classes" title="Permalink to this headline"></a></h2>
<table class="longtable table autosummary">
<colgroup>
<col style="width: 10%" />
<col style="width: 90%" />
</colgroup>
<tbody>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.html#pyspark.sql.SparkSession" title="pyspark.sql.SparkSession"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession</span></code></a>(sparkContext[, jsparkSession])</p></td>
<td><p>The entry point to programming Spark with the Dataset and DataFrame API.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.html#pyspark.sql.Catalog" title="pyspark.sql.Catalog"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog</span></code></a>(sparkSession)</p></td>
<td><p>User-facing catalog API, accessible through <cite>SparkSession.catalog</cite>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame</span></code></a>(jdf, sql_ctx)</p></td>
<td><p>A distributed collection of data grouped into named columns.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.html#pyspark.sql.Column" title="pyspark.sql.Column"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column</span></code></a>(jc)</p></td>
<td><p>A column in a DataFrame.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Row.html#pyspark.sql.Row" title="pyspark.sql.Row"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Row</span></code></a></p></td>
<td><p>A row in <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.GroupedData.html#pyspark.sql.GroupedData" title="pyspark.sql.GroupedData"><code class="xref py py-obj docutils literal notranslate"><span class="pre">GroupedData</span></code></a>(jgd, df)</p></td>
<td><p>A set of methods for aggregations on a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>, created by <a class="reference internal" href="api/pyspark.sql.DataFrame.groupBy.html#pyspark.sql.DataFrame.groupBy" title="pyspark.sql.DataFrame.groupBy"><code class="xref py py-func docutils literal notranslate"><span class="pre">DataFrame.groupBy()</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.PandasCogroupedOps.html#pyspark.sql.PandasCogroupedOps" title="pyspark.sql.PandasCogroupedOps"><code class="xref py py-obj docutils literal notranslate"><span class="pre">PandasCogroupedOps</span></code></a>(gd1, gd2)</p></td>
<td><p>A logical grouping of two <a class="reference internal" href="api/pyspark.sql.GroupedData.html#pyspark.sql.GroupedData" title="pyspark.sql.GroupedData"><code class="xref py py-class docutils literal notranslate"><span class="pre">GroupedData</span></code></a>, created by <a class="reference internal" href="api/pyspark.sql.GroupedData.cogroup.html#pyspark.sql.GroupedData.cogroup" title="pyspark.sql.GroupedData.cogroup"><code class="xref py py-func docutils literal notranslate"><span class="pre">GroupedData.cogroup()</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameNaFunctions.html#pyspark.sql.DataFrameNaFunctions" title="pyspark.sql.DataFrameNaFunctions"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameNaFunctions</span></code></a>(df)</p></td>
<td><p>Functionality for working with missing data in <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.html#pyspark.sql.DataFrameStatFunctions" title="pyspark.sql.DataFrameStatFunctions"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameStatFunctions</span></code></a>(df)</p></td>
<td><p>Functionality for statistic functions with <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Window.html#pyspark.sql.Window" title="pyspark.sql.Window"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Window</span></code></a></p></td>
<td><p>Utility functions for defining window in DataFrames.</p></td>
</tr>
</tbody>
</table>
</section>
<section id="spark-session-apis">
<h2>Spark Session APIs<a class="headerlink" href="#spark-session-apis" title="Permalink to this headline"></a></h2>
<p>The entry point to programming Spark with the Dataset and DataFrame API.
To create a Spark session, you should use <code class="docutils literal notranslate"><span class="pre">SparkSession.builder</span></code> attribute.
See also <a class="reference internal" href="api/pyspark.sql.SparkSession.html#pyspark.sql.SparkSession" title="pyspark.sql.SparkSession"><code class="xref py py-class docutils literal notranslate"><span class="pre">SparkSession</span></code></a>.</p>
<table class="longtable table autosummary">
<colgroup>
<col style="width: 10%" />
<col style="width: 90%" />
</colgroup>
<tbody>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.builder.appName.html#pyspark.sql.SparkSession.builder.appName" title="pyspark.sql.SparkSession.builder.appName"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.builder.appName</span></code></a>(name)</p></td>
<td><p>Sets a name for the application, which will be shown in the Spark web UI.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.builder.config.html#pyspark.sql.SparkSession.builder.config" title="pyspark.sql.SparkSession.builder.config"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.builder.config</span></code></a>([key, value, conf])</p></td>
<td><p>Sets a config option.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.builder.enableHiveSupport.html#pyspark.sql.SparkSession.builder.enableHiveSupport" title="pyspark.sql.SparkSession.builder.enableHiveSupport"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.builder.enableHiveSupport</span></code></a>()</p></td>
<td><p>Enables Hive support, including connectivity to a persistent Hive metastore, support for Hive SerDes, and Hive user-defined functions.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.builder.getOrCreate.html#pyspark.sql.SparkSession.builder.getOrCreate" title="pyspark.sql.SparkSession.builder.getOrCreate"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.builder.getOrCreate</span></code></a>()</p></td>
<td><p>Gets an existing <a class="reference internal" href="api/pyspark.sql.SparkSession.html#pyspark.sql.SparkSession" title="pyspark.sql.SparkSession"><code class="xref py py-class docutils literal notranslate"><span class="pre">SparkSession</span></code></a> or, if there is no existing one, creates a new one based on the options set in this builder.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.builder.master.html#pyspark.sql.SparkSession.builder.master" title="pyspark.sql.SparkSession.builder.master"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.builder.master</span></code></a>(master)</p></td>
<td><p>Sets the Spark master URL to connect to, such as “local” to run locally, “local[4]” to run locally with 4 cores, or “spark://master:7077” to run on a Spark standalone cluster.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.catalog.html#pyspark.sql.SparkSession.catalog" title="pyspark.sql.SparkSession.catalog"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.catalog</span></code></a></p></td>
<td><p>Interface through which the user may create, drop, alter or query underlying databases, tables, functions, etc.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.conf.html#pyspark.sql.SparkSession.conf" title="pyspark.sql.SparkSession.conf"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.conf</span></code></a></p></td>
<td><p>Runtime configuration interface for Spark.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.createDataFrame.html#pyspark.sql.SparkSession.createDataFrame" title="pyspark.sql.SparkSession.createDataFrame"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.createDataFrame</span></code></a>(data[, schema, …])</p></td>
<td><p>Creates a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> from an <code class="xref py py-class docutils literal notranslate"><span class="pre">RDD</span></code>, a list or a <code class="xref py py-class docutils literal notranslate"><span class="pre">pandas.DataFrame</span></code>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.getActiveSession.html#pyspark.sql.SparkSession.getActiveSession" title="pyspark.sql.SparkSession.getActiveSession"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.getActiveSession</span></code></a>()</p></td>
<td><p>Returns the active <a class="reference internal" href="api/pyspark.sql.SparkSession.html#pyspark.sql.SparkSession" title="pyspark.sql.SparkSession"><code class="xref py py-class docutils literal notranslate"><span class="pre">SparkSession</span></code></a> for the current thread, returned by the builder</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.newSession.html#pyspark.sql.SparkSession.newSession" title="pyspark.sql.SparkSession.newSession"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.newSession</span></code></a>()</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.SparkSession.html#pyspark.sql.SparkSession" title="pyspark.sql.SparkSession"><code class="xref py py-class docutils literal notranslate"><span class="pre">SparkSession</span></code></a> as new session, that has separate SQLConf, registered temporary views and UDFs, but shared <code class="xref py py-class docutils literal notranslate"><span class="pre">SparkContext</span></code> and table cache.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.range.html#pyspark.sql.SparkSession.range" title="pyspark.sql.SparkSession.range"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.range</span></code></a>(start[, end, step, …])</p></td>
<td><p>Create a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> with single <a class="reference internal" href="api/pyspark.sql.types.LongType.html#pyspark.sql.types.LongType" title="pyspark.sql.types.LongType"><code class="xref py py-class docutils literal notranslate"><span class="pre">pyspark.sql.types.LongType</span></code></a> column named <code class="docutils literal notranslate"><span class="pre">id</span></code>, containing elements in a range from <code class="docutils literal notranslate"><span class="pre">start</span></code> to <code class="docutils literal notranslate"><span class="pre">end</span></code> (exclusive) with step value <code class="docutils literal notranslate"><span class="pre">step</span></code>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.read.html#pyspark.sql.SparkSession.read" title="pyspark.sql.SparkSession.read"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.read</span></code></a></p></td>
<td><p>Returns a <code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrameReader</span></code> that can be used to read data in as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.readStream.html#pyspark.sql.SparkSession.readStream" title="pyspark.sql.SparkSession.readStream"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.readStream</span></code></a></p></td>
<td><p>Returns a <code class="xref py py-class docutils literal notranslate"><span class="pre">DataStreamReader</span></code> that can be used to read data streams as a streaming <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.sparkContext.html#pyspark.sql.SparkSession.sparkContext" title="pyspark.sql.SparkSession.sparkContext"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.sparkContext</span></code></a></p></td>
<td><p>Returns the underlying <code class="xref py py-class docutils literal notranslate"><span class="pre">SparkContext</span></code>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.sql.html#pyspark.sql.SparkSession.sql" title="pyspark.sql.SparkSession.sql"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.sql</span></code></a>(sqlQuery)</p></td>
<td><p>Returns a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> representing the result of the given query.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.stop.html#pyspark.sql.SparkSession.stop" title="pyspark.sql.SparkSession.stop"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.stop</span></code></a>()</p></td>
<td><p>Stop the underlying <code class="xref py py-class docutils literal notranslate"><span class="pre">SparkContext</span></code>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.streams.html#pyspark.sql.SparkSession.streams" title="pyspark.sql.SparkSession.streams"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.streams</span></code></a></p></td>
<td><p>Returns a <code class="xref py py-class docutils literal notranslate"><span class="pre">StreamingQueryManager</span></code> that allows managing all the <code class="xref py py-class docutils literal notranslate"><span class="pre">StreamingQuery</span></code> instances active on <cite>this</cite> context.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.table.html#pyspark.sql.SparkSession.table" title="pyspark.sql.SparkSession.table"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.table</span></code></a>(tableName)</p></td>
<td><p>Returns the specified table as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.udf.html#pyspark.sql.SparkSession.udf" title="pyspark.sql.SparkSession.udf"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.udf</span></code></a></p></td>
<td><p>Returns a <code class="xref py py-class docutils literal notranslate"><span class="pre">UDFRegistration</span></code> for UDF registration.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.SparkSession.version.html#pyspark.sql.SparkSession.version" title="pyspark.sql.SparkSession.version"><code class="xref py py-obj docutils literal notranslate"><span class="pre">SparkSession.version</span></code></a></p></td>
<td><p>The version of Spark on which this application is running.</p></td>
</tr>
</tbody>
</table>
</section>
<section id="configuration">
<h2>Configuration<a class="headerlink" href="#configuration" title="Permalink to this headline"></a></h2>
<table class="longtable table autosummary">
<colgroup>
<col style="width: 10%" />
<col style="width: 90%" />
</colgroup>
<tbody>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.conf.RuntimeConfig.html#pyspark.sql.conf.RuntimeConfig" title="pyspark.sql.conf.RuntimeConfig"><code class="xref py py-obj docutils literal notranslate"><span class="pre">RuntimeConfig</span></code></a>(jconf)</p></td>
<td><p>User-facing configuration API, accessible through <cite>SparkSession.conf</cite>.</p></td>
</tr>
</tbody>
</table>
</section>
<section id="input-and-output">
<h2>Input and Output<a class="headerlink" href="#input-and-output" title="Permalink to this headline"></a></h2>
<table class="longtable table autosummary">
<colgroup>
<col style="width: 10%" />
<col style="width: 90%" />
</colgroup>
<tbody>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.csv.html#pyspark.sql.DataFrameReader.csv" title="pyspark.sql.DataFrameReader.csv"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.csv</span></code></a>(path[, schema, sep, …])</p></td>
<td><p>Loads a CSV file and returns the result as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.format.html#pyspark.sql.DataFrameReader.format" title="pyspark.sql.DataFrameReader.format"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.format</span></code></a>(source)</p></td>
<td><p>Specifies the input data source format.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.jdbc.html#pyspark.sql.DataFrameReader.jdbc" title="pyspark.sql.DataFrameReader.jdbc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.jdbc</span></code></a>(url, table[, column, …])</p></td>
<td><p>Construct a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> representing the database table named <code class="docutils literal notranslate"><span class="pre">table</span></code> accessible via JDBC URL <code class="docutils literal notranslate"><span class="pre">url</span></code> and connection <code class="docutils literal notranslate"><span class="pre">properties</span></code>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.json.html#pyspark.sql.DataFrameReader.json" title="pyspark.sql.DataFrameReader.json"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.json</span></code></a>(path[, schema, …])</p></td>
<td><p>Loads JSON files and returns the results as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.load.html#pyspark.sql.DataFrameReader.load" title="pyspark.sql.DataFrameReader.load"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.load</span></code></a>([path, format, schema])</p></td>
<td><p>Loads data from a data source and returns it as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.option.html#pyspark.sql.DataFrameReader.option" title="pyspark.sql.DataFrameReader.option"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.option</span></code></a>(key, value)</p></td>
<td><p>Adds an input option for the underlying data source.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.options.html#pyspark.sql.DataFrameReader.options" title="pyspark.sql.DataFrameReader.options"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.options</span></code></a>(**options)</p></td>
<td><p>Adds input options for the underlying data source.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.orc.html#pyspark.sql.DataFrameReader.orc" title="pyspark.sql.DataFrameReader.orc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.orc</span></code></a>(path[, mergeSchema, …])</p></td>
<td><p>Loads ORC files, returning the result as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.parquet.html#pyspark.sql.DataFrameReader.parquet" title="pyspark.sql.DataFrameReader.parquet"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.parquet</span></code></a>(*paths, **options)</p></td>
<td><p>Loads Parquet files, returning the result as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.schema.html#pyspark.sql.DataFrameReader.schema" title="pyspark.sql.DataFrameReader.schema"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.schema</span></code></a>(schema)</p></td>
<td><p>Specifies the input schema.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameReader.table.html#pyspark.sql.DataFrameReader.table" title="pyspark.sql.DataFrameReader.table"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameReader.table</span></code></a>(tableName)</p></td>
<td><p>Returns the specified table as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.bucketBy.html#pyspark.sql.DataFrameWriter.bucketBy" title="pyspark.sql.DataFrameWriter.bucketBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.bucketBy</span></code></a>(numBuckets, col, *cols)</p></td>
<td><p>Buckets the output by the given columns.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.csv.html#pyspark.sql.DataFrameWriter.csv" title="pyspark.sql.DataFrameWriter.csv"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.csv</span></code></a>(path[, mode, …])</p></td>
<td><p>Saves the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> in CSV format at the specified path.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.format.html#pyspark.sql.DataFrameWriter.format" title="pyspark.sql.DataFrameWriter.format"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.format</span></code></a>(source)</p></td>
<td><p>Specifies the underlying output data source.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.insertInto.html#pyspark.sql.DataFrameWriter.insertInto" title="pyspark.sql.DataFrameWriter.insertInto"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.insertInto</span></code></a>(tableName[, …])</p></td>
<td><p>Inserts the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> to the specified table.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.jdbc.html#pyspark.sql.DataFrameWriter.jdbc" title="pyspark.sql.DataFrameWriter.jdbc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.jdbc</span></code></a>(url, table[, mode, …])</p></td>
<td><p>Saves the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> to an external database table via JDBC.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.json.html#pyspark.sql.DataFrameWriter.json" title="pyspark.sql.DataFrameWriter.json"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.json</span></code></a>(path[, mode, …])</p></td>
<td><p>Saves the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> in JSON format (<a class="reference external" href="http://jsonlines.org/">JSON Lines text format or newline-delimited JSON</a>) at the specified path.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.mode.html#pyspark.sql.DataFrameWriter.mode" title="pyspark.sql.DataFrameWriter.mode"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.mode</span></code></a>(saveMode)</p></td>
<td><p>Specifies the behavior when data or table already exists.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.option.html#pyspark.sql.DataFrameWriter.option" title="pyspark.sql.DataFrameWriter.option"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.option</span></code></a>(key, value)</p></td>
<td><p>Adds an output option for the underlying data source.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.options.html#pyspark.sql.DataFrameWriter.options" title="pyspark.sql.DataFrameWriter.options"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.options</span></code></a>(**options)</p></td>
<td><p>Adds output options for the underlying data source.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.orc.html#pyspark.sql.DataFrameWriter.orc" title="pyspark.sql.DataFrameWriter.orc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.orc</span></code></a>(path[, mode, …])</p></td>
<td><p>Saves the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> in ORC format at the specified path.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.parquet.html#pyspark.sql.DataFrameWriter.parquet" title="pyspark.sql.DataFrameWriter.parquet"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.parquet</span></code></a>(path[, mode, …])</p></td>
<td><p>Saves the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> in Parquet format at the specified path.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.partitionBy.html#pyspark.sql.DataFrameWriter.partitionBy" title="pyspark.sql.DataFrameWriter.partitionBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.partitionBy</span></code></a>(*cols)</p></td>
<td><p>Partitions the output by the given columns on the file system.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.save.html#pyspark.sql.DataFrameWriter.save" title="pyspark.sql.DataFrameWriter.save"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.save</span></code></a>([path, format, mode, …])</p></td>
<td><p>Saves the contents of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> to a data source.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.saveAsTable.html#pyspark.sql.DataFrameWriter.saveAsTable" title="pyspark.sql.DataFrameWriter.saveAsTable"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.saveAsTable</span></code></a>(name[, format, …])</p></td>
<td><p>Saves the content of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> as the specified table.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.sortBy.html#pyspark.sql.DataFrameWriter.sortBy" title="pyspark.sql.DataFrameWriter.sortBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.sortBy</span></code></a>(col, *cols)</p></td>
<td><p>Sorts the output in each bucket by the given columns on the file system.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameWriter.text.html#pyspark.sql.DataFrameWriter.text" title="pyspark.sql.DataFrameWriter.text"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameWriter.text</span></code></a>(path[, compression, …])</p></td>
<td><p>Saves the content of the DataFrame in a text file at the specified path.</p></td>
</tr>
</tbody>
</table>
</section>
<section id="dataframe-apis">
<h2>DataFrame APIs<a class="headerlink" href="#dataframe-apis" title="Permalink to this headline"></a></h2>
<table class="longtable table autosummary">
<colgroup>
<col style="width: 10%" />
<col style="width: 90%" />
</colgroup>
<tbody>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.agg.html#pyspark.sql.DataFrame.agg" title="pyspark.sql.DataFrame.agg"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.agg</span></code></a>(*exprs)</p></td>
<td><p>Aggregate on the entire <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> without groups (shorthand for <code class="docutils literal notranslate"><span class="pre">df.groupBy().agg()</span></code>).</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.alias.html#pyspark.sql.DataFrame.alias" title="pyspark.sql.DataFrame.alias"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.alias</span></code></a>(alias)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> with an alias set.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.approxQuantile.html#pyspark.sql.DataFrame.approxQuantile" title="pyspark.sql.DataFrame.approxQuantile"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.approxQuantile</span></code></a>(col, probabilities, …)</p></td>
<td><p>Calculates the approximate quantiles of numerical columns of a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.cache.html#pyspark.sql.DataFrame.cache" title="pyspark.sql.DataFrame.cache"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.cache</span></code></a>()</p></td>
<td><p>Persists the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> with the default storage level (<cite>MEMORY_AND_DISK</cite>).</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.checkpoint.html#pyspark.sql.DataFrame.checkpoint" title="pyspark.sql.DataFrame.checkpoint"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.checkpoint</span></code></a>([eager])</p></td>
<td><p>Returns a checkpointed version of this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.coalesce.html#pyspark.sql.DataFrame.coalesce" title="pyspark.sql.DataFrame.coalesce"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.coalesce</span></code></a>(numPartitions)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> that has exactly <cite>numPartitions</cite> partitions.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.colRegex.html#pyspark.sql.DataFrame.colRegex" title="pyspark.sql.DataFrame.colRegex"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.colRegex</span></code></a>(colName)</p></td>
<td><p>Selects column based on the column name specified as a regex and returns it as <a class="reference internal" href="api/pyspark.sql.Column.html#pyspark.sql.Column" title="pyspark.sql.Column"><code class="xref py py-class docutils literal notranslate"><span class="pre">Column</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.collect.html#pyspark.sql.DataFrame.collect" title="pyspark.sql.DataFrame.collect"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.collect</span></code></a>()</p></td>
<td><p>Returns all the records as a list of <a class="reference internal" href="api/pyspark.sql.Row.html#pyspark.sql.Row" title="pyspark.sql.Row"><code class="xref py py-class docutils literal notranslate"><span class="pre">Row</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.columns.html#pyspark.sql.DataFrame.columns" title="pyspark.sql.DataFrame.columns"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.columns</span></code></a></p></td>
<td><p>Returns all column names as a list.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.corr.html#pyspark.sql.DataFrame.corr" title="pyspark.sql.DataFrame.corr"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.corr</span></code></a>(col1, col2[, method])</p></td>
<td><p>Calculates the correlation of two columns of a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> as a double value.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.count.html#pyspark.sql.DataFrame.count" title="pyspark.sql.DataFrame.count"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.count</span></code></a>()</p></td>
<td><p>Returns the number of rows in this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.cov.html#pyspark.sql.DataFrame.cov" title="pyspark.sql.DataFrame.cov"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.cov</span></code></a>(col1, col2)</p></td>
<td><p>Calculate the sample covariance for the given columns, specified by their names, as a double value.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.createGlobalTempView.html#pyspark.sql.DataFrame.createGlobalTempView" title="pyspark.sql.DataFrame.createGlobalTempView"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.createGlobalTempView</span></code></a>(name)</p></td>
<td><p>Creates a global temporary view with this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.createOrReplaceGlobalTempView.html#pyspark.sql.DataFrame.createOrReplaceGlobalTempView" title="pyspark.sql.DataFrame.createOrReplaceGlobalTempView"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.createOrReplaceGlobalTempView</span></code></a>(name)</p></td>
<td><p>Creates or replaces a global temporary view using the given name.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.createOrReplaceTempView.html#pyspark.sql.DataFrame.createOrReplaceTempView" title="pyspark.sql.DataFrame.createOrReplaceTempView"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.createOrReplaceTempView</span></code></a>(name)</p></td>
<td><p>Creates or replaces a local temporary view with this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.createTempView.html#pyspark.sql.DataFrame.createTempView" title="pyspark.sql.DataFrame.createTempView"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.createTempView</span></code></a>(name)</p></td>
<td><p>Creates a local temporary view with this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.crossJoin.html#pyspark.sql.DataFrame.crossJoin" title="pyspark.sql.DataFrame.crossJoin"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.crossJoin</span></code></a>(other)</p></td>
<td><p>Returns the cartesian product with another <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.crosstab.html#pyspark.sql.DataFrame.crosstab" title="pyspark.sql.DataFrame.crosstab"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.crosstab</span></code></a>(col1, col2)</p></td>
<td><p>Computes a pair-wise frequency table of the given columns.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.cube.html#pyspark.sql.DataFrame.cube" title="pyspark.sql.DataFrame.cube"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.cube</span></code></a>(*cols)</p></td>
<td><p>Create a multi-dimensional cube for the current <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> using the specified columns, so we can run aggregations on them.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.describe.html#pyspark.sql.DataFrame.describe" title="pyspark.sql.DataFrame.describe"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.describe</span></code></a>(*cols)</p></td>
<td><p>Computes basic statistics for numeric and string columns.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.distinct.html#pyspark.sql.DataFrame.distinct" title="pyspark.sql.DataFrame.distinct"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.distinct</span></code></a>()</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> containing the distinct rows in this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.drop.html#pyspark.sql.DataFrame.drop" title="pyspark.sql.DataFrame.drop"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.drop</span></code></a>(*cols)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> that drops the specified column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.dropDuplicates.html#pyspark.sql.DataFrame.dropDuplicates" title="pyspark.sql.DataFrame.dropDuplicates"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.dropDuplicates</span></code></a>([subset])</p></td>
<td><p>Return a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> with duplicate rows removed, optionally only considering certain columns.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.drop_duplicates.html#pyspark.sql.DataFrame.drop_duplicates" title="pyspark.sql.DataFrame.drop_duplicates"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.drop_duplicates</span></code></a>([subset])</p></td>
<td><p><code class="xref py py-func docutils literal notranslate"><span class="pre">drop_duplicates()</span></code> is an alias for <code class="xref py py-func docutils literal notranslate"><span class="pre">dropDuplicates()</span></code>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.dropna.html#pyspark.sql.DataFrame.dropna" title="pyspark.sql.DataFrame.dropna"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.dropna</span></code></a>([how, thresh, subset])</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> omitting rows with null values.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.dtypes.html#pyspark.sql.DataFrame.dtypes" title="pyspark.sql.DataFrame.dtypes"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.dtypes</span></code></a></p></td>
<td><p>Returns all column names and their data types as a list.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.exceptAll.html#pyspark.sql.DataFrame.exceptAll" title="pyspark.sql.DataFrame.exceptAll"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.exceptAll</span></code></a>(other)</p></td>
<td><p>Return a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> containing rows in this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> but not in another <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> while preserving duplicates.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.explain.html#pyspark.sql.DataFrame.explain" title="pyspark.sql.DataFrame.explain"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.explain</span></code></a>([extended, mode])</p></td>
<td><p>Prints the (logical and physical) plans to the console for debugging purpose.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.fillna.html#pyspark.sql.DataFrame.fillna" title="pyspark.sql.DataFrame.fillna"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.fillna</span></code></a>(value[, subset])</p></td>
<td><p>Replace null values, alias for <code class="docutils literal notranslate"><span class="pre">na.fill()</span></code>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.filter.html#pyspark.sql.DataFrame.filter" title="pyspark.sql.DataFrame.filter"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.filter</span></code></a>(condition)</p></td>
<td><p>Filters rows using the given condition.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.first.html#pyspark.sql.DataFrame.first" title="pyspark.sql.DataFrame.first"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.first</span></code></a>()</p></td>
<td><p>Returns the first row as a <a class="reference internal" href="api/pyspark.sql.Row.html#pyspark.sql.Row" title="pyspark.sql.Row"><code class="xref py py-class docutils literal notranslate"><span class="pre">Row</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.foreach.html#pyspark.sql.DataFrame.foreach" title="pyspark.sql.DataFrame.foreach"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.foreach</span></code></a>(f)</p></td>
<td><p>Applies the <code class="docutils literal notranslate"><span class="pre">f</span></code> function to all <a class="reference internal" href="api/pyspark.sql.Row.html#pyspark.sql.Row" title="pyspark.sql.Row"><code class="xref py py-class docutils literal notranslate"><span class="pre">Row</span></code></a> of this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.foreachPartition.html#pyspark.sql.DataFrame.foreachPartition" title="pyspark.sql.DataFrame.foreachPartition"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.foreachPartition</span></code></a>(f)</p></td>
<td><p>Applies the <code class="docutils literal notranslate"><span class="pre">f</span></code> function to each partition of this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.freqItems.html#pyspark.sql.DataFrame.freqItems" title="pyspark.sql.DataFrame.freqItems"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.freqItems</span></code></a>(cols[, support])</p></td>
<td><p>Finding frequent items for columns, possibly with false positives.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.groupBy.html#pyspark.sql.DataFrame.groupBy" title="pyspark.sql.DataFrame.groupBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.groupBy</span></code></a>(*cols)</p></td>
<td><p>Groups the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> using the specified columns, so we can run aggregation on them.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.head.html#pyspark.sql.DataFrame.head" title="pyspark.sql.DataFrame.head"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.head</span></code></a>([n])</p></td>
<td><p>Returns the first <code class="docutils literal notranslate"><span class="pre">n</span></code> rows.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.hint.html#pyspark.sql.DataFrame.hint" title="pyspark.sql.DataFrame.hint"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.hint</span></code></a>(name, *parameters)</p></td>
<td><p>Specifies some hint on the current <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.inputFiles.html#pyspark.sql.DataFrame.inputFiles" title="pyspark.sql.DataFrame.inputFiles"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.inputFiles</span></code></a>()</p></td>
<td><p>Returns a best-effort snapshot of the files that compose this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.intersect.html#pyspark.sql.DataFrame.intersect" title="pyspark.sql.DataFrame.intersect"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.intersect</span></code></a>(other)</p></td>
<td><p>Return a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> containing rows only in both this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> and another <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.intersectAll.html#pyspark.sql.DataFrame.intersectAll" title="pyspark.sql.DataFrame.intersectAll"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.intersectAll</span></code></a>(other)</p></td>
<td><p>Return a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> containing rows in both this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> and another <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> while preserving duplicates.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.isLocal.html#pyspark.sql.DataFrame.isLocal" title="pyspark.sql.DataFrame.isLocal"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.isLocal</span></code></a>()</p></td>
<td><p>Returns <code class="docutils literal notranslate"><span class="pre">True</span></code> if the <code class="xref py py-func docutils literal notranslate"><span class="pre">collect()</span></code> and <code class="xref py py-func docutils literal notranslate"><span class="pre">take()</span></code> methods can be run locally (without any Spark executors).</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.isStreaming.html#pyspark.sql.DataFrame.isStreaming" title="pyspark.sql.DataFrame.isStreaming"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.isStreaming</span></code></a></p></td>
<td><p>Returns <code class="docutils literal notranslate"><span class="pre">True</span></code> if this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> contains one or more sources that continuously return data as it arrives.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.join.html#pyspark.sql.DataFrame.join" title="pyspark.sql.DataFrame.join"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.join</span></code></a>(other[, on, how])</p></td>
<td><p>Joins with another <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>, using the given join expression.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.limit.html#pyspark.sql.DataFrame.limit" title="pyspark.sql.DataFrame.limit"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.limit</span></code></a>(num)</p></td>
<td><p>Limits the result count to the number specified.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.localCheckpoint.html#pyspark.sql.DataFrame.localCheckpoint" title="pyspark.sql.DataFrame.localCheckpoint"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.localCheckpoint</span></code></a>([eager])</p></td>
<td><p>Returns a locally checkpointed version of this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.mapInPandas.html#pyspark.sql.DataFrame.mapInPandas" title="pyspark.sql.DataFrame.mapInPandas"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.mapInPandas</span></code></a>(func, schema)</p></td>
<td><p>Maps an iterator of batches in the current <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> using a Python native function that takes and outputs a pandas DataFrame, and returns the result as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.na.html#pyspark.sql.DataFrame.na" title="pyspark.sql.DataFrame.na"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.na</span></code></a></p></td>
<td><p>Returns a <a class="reference internal" href="api/pyspark.sql.DataFrameNaFunctions.html#pyspark.sql.DataFrameNaFunctions" title="pyspark.sql.DataFrameNaFunctions"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrameNaFunctions</span></code></a> for handling missing values.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.orderBy.html#pyspark.sql.DataFrame.orderBy" title="pyspark.sql.DataFrame.orderBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.orderBy</span></code></a>(*cols, **kwargs)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> sorted by the specified column(s).</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.persist.html#pyspark.sql.DataFrame.persist" title="pyspark.sql.DataFrame.persist"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.persist</span></code></a>([storageLevel])</p></td>
<td><p>Sets the storage level to persist the contents of the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> across operations after the first time it is computed.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.printSchema.html#pyspark.sql.DataFrame.printSchema" title="pyspark.sql.DataFrame.printSchema"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.printSchema</span></code></a>()</p></td>
<td><p>Prints out the schema in the tree format.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.randomSplit.html#pyspark.sql.DataFrame.randomSplit" title="pyspark.sql.DataFrame.randomSplit"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.randomSplit</span></code></a>(weights[, seed])</p></td>
<td><p>Randomly splits this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> with the provided weights.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.rdd.html#pyspark.sql.DataFrame.rdd" title="pyspark.sql.DataFrame.rdd"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.rdd</span></code></a></p></td>
<td><p>Returns the content as an <a class="reference internal" href="api/pyspark.RDD.html#pyspark.RDD" title="pyspark.RDD"><code class="xref py py-class docutils literal notranslate"><span class="pre">pyspark.RDD</span></code></a> of <a class="reference internal" href="api/pyspark.sql.Row.html#pyspark.sql.Row" title="pyspark.sql.Row"><code class="xref py py-class docutils literal notranslate"><span class="pre">Row</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.registerTempTable.html#pyspark.sql.DataFrame.registerTempTable" title="pyspark.sql.DataFrame.registerTempTable"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.registerTempTable</span></code></a>(name)</p></td>
<td><p>Registers this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> as a temporary table using the given name.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.repartition.html#pyspark.sql.DataFrame.repartition" title="pyspark.sql.DataFrame.repartition"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.repartition</span></code></a>(numPartitions, *cols)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> partitioned by the given partitioning expressions.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.repartitionByRange.html#pyspark.sql.DataFrame.repartitionByRange" title="pyspark.sql.DataFrame.repartitionByRange"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.repartitionByRange</span></code></a>(numPartitions, …)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> partitioned by the given partitioning expressions.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.replace.html#pyspark.sql.DataFrame.replace" title="pyspark.sql.DataFrame.replace"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.replace</span></code></a>(to_replace[, value, subset])</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> replacing a value with another value.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.rollup.html#pyspark.sql.DataFrame.rollup" title="pyspark.sql.DataFrame.rollup"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.rollup</span></code></a>(*cols)</p></td>
<td><p>Create a multi-dimensional rollup for the current <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> using the specified columns, so we can run aggregation on them.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.sameSemantics.html#pyspark.sql.DataFrame.sameSemantics" title="pyspark.sql.DataFrame.sameSemantics"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.sameSemantics</span></code></a>(other)</p></td>
<td><p>Returns <cite>True</cite> when the logical query plans inside both <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>s are equal and therefore return same results.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.sample.html#pyspark.sql.DataFrame.sample" title="pyspark.sql.DataFrame.sample"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.sample</span></code></a>([withReplacement, …])</p></td>
<td><p>Returns a sampled subset of this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.sampleBy.html#pyspark.sql.DataFrame.sampleBy" title="pyspark.sql.DataFrame.sampleBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.sampleBy</span></code></a>(col, fractions[, seed])</p></td>
<td><p>Returns a stratified sample without replacement based on the fraction given on each stratum.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.schema.html#pyspark.sql.DataFrame.schema" title="pyspark.sql.DataFrame.schema"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.schema</span></code></a></p></td>
<td><p>Returns the schema of this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> as a <a class="reference internal" href="api/pyspark.sql.types.StructType.html#pyspark.sql.types.StructType" title="pyspark.sql.types.StructType"><code class="xref py py-class docutils literal notranslate"><span class="pre">pyspark.sql.types.StructType</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.select.html#pyspark.sql.DataFrame.select" title="pyspark.sql.DataFrame.select"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.select</span></code></a>(*cols)</p></td>
<td><p>Projects a set of expressions and returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.selectExpr.html#pyspark.sql.DataFrame.selectExpr" title="pyspark.sql.DataFrame.selectExpr"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.selectExpr</span></code></a>(*expr)</p></td>
<td><p>Projects a set of SQL expressions and returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.semanticHash.html#pyspark.sql.DataFrame.semanticHash" title="pyspark.sql.DataFrame.semanticHash"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.semanticHash</span></code></a>()</p></td>
<td><p>Returns a hash code of the logical query plan against this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.show.html#pyspark.sql.DataFrame.show" title="pyspark.sql.DataFrame.show"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.show</span></code></a>([n, truncate, vertical])</p></td>
<td><p>Prints the first <code class="docutils literal notranslate"><span class="pre">n</span></code> rows to the console.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.sort.html#pyspark.sql.DataFrame.sort" title="pyspark.sql.DataFrame.sort"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.sort</span></code></a>(*cols, **kwargs)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> sorted by the specified column(s).</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.sortWithinPartitions.html#pyspark.sql.DataFrame.sortWithinPartitions" title="pyspark.sql.DataFrame.sortWithinPartitions"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.sortWithinPartitions</span></code></a>(*cols, **kwargs)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> with each partition sorted by the specified column(s).</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.stat.html#pyspark.sql.DataFrame.stat" title="pyspark.sql.DataFrame.stat"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.stat</span></code></a></p></td>
<td><p>Returns a <a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.html#pyspark.sql.DataFrameStatFunctions" title="pyspark.sql.DataFrameStatFunctions"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrameStatFunctions</span></code></a> for statistic functions.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.storageLevel.html#pyspark.sql.DataFrame.storageLevel" title="pyspark.sql.DataFrame.storageLevel"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.storageLevel</span></code></a></p></td>
<td><p>Get the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>’s current storage level.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.subtract.html#pyspark.sql.DataFrame.subtract" title="pyspark.sql.DataFrame.subtract"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.subtract</span></code></a>(other)</p></td>
<td><p>Return a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> containing rows in this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> but not in another <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.summary.html#pyspark.sql.DataFrame.summary" title="pyspark.sql.DataFrame.summary"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.summary</span></code></a>(*statistics)</p></td>
<td><p>Computes specified statistics for numeric and string columns.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.tail.html#pyspark.sql.DataFrame.tail" title="pyspark.sql.DataFrame.tail"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.tail</span></code></a>(num)</p></td>
<td><p>Returns the last <code class="docutils literal notranslate"><span class="pre">num</span></code> rows as a <code class="xref py py-class docutils literal notranslate"><span class="pre">list</span></code> of <a class="reference internal" href="api/pyspark.sql.Row.html#pyspark.sql.Row" title="pyspark.sql.Row"><code class="xref py py-class docutils literal notranslate"><span class="pre">Row</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.take.html#pyspark.sql.DataFrame.take" title="pyspark.sql.DataFrame.take"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.take</span></code></a>(num)</p></td>
<td><p>Returns the first <code class="docutils literal notranslate"><span class="pre">num</span></code> rows as a <code class="xref py py-class docutils literal notranslate"><span class="pre">list</span></code> of <a class="reference internal" href="api/pyspark.sql.Row.html#pyspark.sql.Row" title="pyspark.sql.Row"><code class="xref py py-class docutils literal notranslate"><span class="pre">Row</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.toDF.html#pyspark.sql.DataFrame.toDF" title="pyspark.sql.DataFrame.toDF"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.toDF</span></code></a>(*cols)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> that with new specified column names</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.toJSON.html#pyspark.sql.DataFrame.toJSON" title="pyspark.sql.DataFrame.toJSON"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.toJSON</span></code></a>([use_unicode])</p></td>
<td><p>Converts a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> into a <code class="xref py py-class docutils literal notranslate"><span class="pre">RDD</span></code> of string.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.toLocalIterator.html#pyspark.sql.DataFrame.toLocalIterator" title="pyspark.sql.DataFrame.toLocalIterator"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.toLocalIterator</span></code></a>([prefetchPartitions])</p></td>
<td><p>Returns an iterator that contains all of the rows in this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.toPandas.html#pyspark.sql.DataFrame.toPandas" title="pyspark.sql.DataFrame.toPandas"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.toPandas</span></code></a>()</p></td>
<td><p>Returns the contents of this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> as Pandas <code class="docutils literal notranslate"><span class="pre">pandas.DataFrame</span></code>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.transform.html#pyspark.sql.DataFrame.transform" title="pyspark.sql.DataFrame.transform"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.transform</span></code></a>(func)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.union.html#pyspark.sql.DataFrame.union" title="pyspark.sql.DataFrame.union"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.union</span></code></a>(other)</p></td>
<td><p>Return a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> containing union of rows in this and another <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.unionAll.html#pyspark.sql.DataFrame.unionAll" title="pyspark.sql.DataFrame.unionAll"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.unionAll</span></code></a>(other)</p></td>
<td><p>Return a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> containing union of rows in this and another <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.unionByName.html#pyspark.sql.DataFrame.unionByName" title="pyspark.sql.DataFrame.unionByName"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.unionByName</span></code></a>(other[, …])</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> containing union of rows in this and another <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.unpersist.html#pyspark.sql.DataFrame.unpersist" title="pyspark.sql.DataFrame.unpersist"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.unpersist</span></code></a>([blocking])</p></td>
<td><p>Marks the <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> as non-persistent, and remove all blocks for it from memory and disk.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.where.html#pyspark.sql.DataFrame.where" title="pyspark.sql.DataFrame.where"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.where</span></code></a>(condition)</p></td>
<td><p><code class="xref py py-func docutils literal notranslate"><span class="pre">where()</span></code> is an alias for <code class="xref py py-func docutils literal notranslate"><span class="pre">filter()</span></code>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.withColumn.html#pyspark.sql.DataFrame.withColumn" title="pyspark.sql.DataFrame.withColumn"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.withColumn</span></code></a>(colName, col)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> by adding a column or replacing the existing column that has the same name.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.withColumnRenamed.html#pyspark.sql.DataFrame.withColumnRenamed" title="pyspark.sql.DataFrame.withColumnRenamed"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.withColumnRenamed</span></code></a>(existing, new)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> by renaming an existing column.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.withWatermark.html#pyspark.sql.DataFrame.withWatermark" title="pyspark.sql.DataFrame.withWatermark"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.withWatermark</span></code></a>(eventTime, …)</p></td>
<td><p>Defines an event time watermark for this <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.write.html#pyspark.sql.DataFrame.write" title="pyspark.sql.DataFrame.write"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.write</span></code></a></p></td>
<td><p>Interface for saving the content of the non-streaming <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> out into external storage.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.writeStream.html#pyspark.sql.DataFrame.writeStream" title="pyspark.sql.DataFrame.writeStream"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.writeStream</span></code></a></p></td>
<td><p>Interface for saving the content of the streaming <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> out into external storage.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.writeTo.html#pyspark.sql.DataFrame.writeTo" title="pyspark.sql.DataFrame.writeTo"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.writeTo</span></code></a>(table)</p></td>
<td><p>Create a write configuration builder for v2 sources.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrame.to_pandas_on_spark.html#pyspark.sql.DataFrame.to_pandas_on_spark" title="pyspark.sql.DataFrame.to_pandas_on_spark"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrame.to_pandas_on_spark</span></code></a>([index_col])</p></td>
<td><p>Converts the existing DataFrame into a pandas-on-Spark DataFrame.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameNaFunctions.drop.html#pyspark.sql.DataFrameNaFunctions.drop" title="pyspark.sql.DataFrameNaFunctions.drop"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameNaFunctions.drop</span></code></a>([how, thresh, subset])</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> omitting rows with null values.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameNaFunctions.fill.html#pyspark.sql.DataFrameNaFunctions.fill" title="pyspark.sql.DataFrameNaFunctions.fill"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameNaFunctions.fill</span></code></a>(value[, subset])</p></td>
<td><p>Replace null values, alias for <code class="docutils literal notranslate"><span class="pre">na.fill()</span></code>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameNaFunctions.replace.html#pyspark.sql.DataFrameNaFunctions.replace" title="pyspark.sql.DataFrameNaFunctions.replace"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameNaFunctions.replace</span></code></a>(to_replace[, …])</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> replacing a value with another value.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.approxQuantile.html#pyspark.sql.DataFrameStatFunctions.approxQuantile" title="pyspark.sql.DataFrameStatFunctions.approxQuantile"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameStatFunctions.approxQuantile</span></code></a>(col, …)</p></td>
<td><p>Calculates the approximate quantiles of numerical columns of a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.corr.html#pyspark.sql.DataFrameStatFunctions.corr" title="pyspark.sql.DataFrameStatFunctions.corr"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameStatFunctions.corr</span></code></a>(col1, col2[, method])</p></td>
<td><p>Calculates the correlation of two columns of a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> as a double value.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.cov.html#pyspark.sql.DataFrameStatFunctions.cov" title="pyspark.sql.DataFrameStatFunctions.cov"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameStatFunctions.cov</span></code></a>(col1, col2)</p></td>
<td><p>Calculate the sample covariance for the given columns, specified by their names, as a double value.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.crosstab.html#pyspark.sql.DataFrameStatFunctions.crosstab" title="pyspark.sql.DataFrameStatFunctions.crosstab"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameStatFunctions.crosstab</span></code></a>(col1, col2)</p></td>
<td><p>Computes a pair-wise frequency table of the given columns.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.freqItems.html#pyspark.sql.DataFrameStatFunctions.freqItems" title="pyspark.sql.DataFrameStatFunctions.freqItems"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameStatFunctions.freqItems</span></code></a>(cols[, support])</p></td>
<td><p>Finding frequent items for columns, possibly with false positives.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.DataFrameStatFunctions.sampleBy.html#pyspark.sql.DataFrameStatFunctions.sampleBy" title="pyspark.sql.DataFrameStatFunctions.sampleBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataFrameStatFunctions.sampleBy</span></code></a>(col, fractions)</p></td>
<td><p>Returns a stratified sample without replacement based on the fraction given on each stratum.</p></td>
</tr>
</tbody>
</table>
</section>
<section id="column-apis">
<h2>Column APIs<a class="headerlink" href="#column-apis" title="Permalink to this headline"></a></h2>
<table class="longtable table autosummary">
<colgroup>
<col style="width: 10%" />
<col style="width: 90%" />
</colgroup>
<tbody>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.alias.html#pyspark.sql.Column.alias" title="pyspark.sql.Column.alias"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.alias</span></code></a>(*alias, **kwargs)</p></td>
<td><p>Returns this column aliased with a new name or names (in the case of expressions that return more than one column, such as explode).</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.asc.html#pyspark.sql.Column.asc" title="pyspark.sql.Column.asc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.asc</span></code></a>()</p></td>
<td><p>Returns a sort expression based on ascending order of the column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.asc_nulls_first.html#pyspark.sql.Column.asc_nulls_first" title="pyspark.sql.Column.asc_nulls_first"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.asc_nulls_first</span></code></a>()</p></td>
<td><p>Returns a sort expression based on ascending order of the column, and null values return before non-null values.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.asc_nulls_last.html#pyspark.sql.Column.asc_nulls_last" title="pyspark.sql.Column.asc_nulls_last"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.asc_nulls_last</span></code></a>()</p></td>
<td><p>Returns a sort expression based on ascending order of the column, and null values appear after non-null values.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.astype.html#pyspark.sql.Column.astype" title="pyspark.sql.Column.astype"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.astype</span></code></a>(dataType)</p></td>
<td><p><code class="xref py py-func docutils literal notranslate"><span class="pre">astype()</span></code> is an alias for <code class="xref py py-func docutils literal notranslate"><span class="pre">cast()</span></code>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.between.html#pyspark.sql.Column.between" title="pyspark.sql.Column.between"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.between</span></code></a>(lowerBound, upperBound)</p></td>
<td><p>True if the current column is between the lower bound and upper bound, inclusive.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.bitwiseAND.html#pyspark.sql.Column.bitwiseAND" title="pyspark.sql.Column.bitwiseAND"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.bitwiseAND</span></code></a>(other)</p></td>
<td><p>Compute bitwise AND of this expression with another expression.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.bitwiseOR.html#pyspark.sql.Column.bitwiseOR" title="pyspark.sql.Column.bitwiseOR"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.bitwiseOR</span></code></a>(other)</p></td>
<td><p>Compute bitwise OR of this expression with another expression.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.bitwiseXOR.html#pyspark.sql.Column.bitwiseXOR" title="pyspark.sql.Column.bitwiseXOR"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.bitwiseXOR</span></code></a>(other)</p></td>
<td><p>Compute bitwise XOR of this expression with another expression.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.cast.html#pyspark.sql.Column.cast" title="pyspark.sql.Column.cast"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.cast</span></code></a>(dataType)</p></td>
<td><p>Casts the column into type <code class="docutils literal notranslate"><span class="pre">dataType</span></code>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.contains.html#pyspark.sql.Column.contains" title="pyspark.sql.Column.contains"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.contains</span></code></a>(other)</p></td>
<td><p>Contains the other element.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.desc.html#pyspark.sql.Column.desc" title="pyspark.sql.Column.desc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.desc</span></code></a>()</p></td>
<td><p>Returns a sort expression based on the descending order of the column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.desc_nulls_first.html#pyspark.sql.Column.desc_nulls_first" title="pyspark.sql.Column.desc_nulls_first"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.desc_nulls_first</span></code></a>()</p></td>
<td><p>Returns a sort expression based on the descending order of the column, and null values appear before non-null values.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.desc_nulls_last.html#pyspark.sql.Column.desc_nulls_last" title="pyspark.sql.Column.desc_nulls_last"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.desc_nulls_last</span></code></a>()</p></td>
<td><p>Returns a sort expression based on the descending order of the column, and null values appear after non-null values.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.dropFields.html#pyspark.sql.Column.dropFields" title="pyspark.sql.Column.dropFields"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.dropFields</span></code></a>(*fieldNames)</p></td>
<td><p>An expression that drops fields in <code class="xref py py-class docutils literal notranslate"><span class="pre">StructType</span></code> by name.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.endswith.html#pyspark.sql.Column.endswith" title="pyspark.sql.Column.endswith"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.endswith</span></code></a>(other)</p></td>
<td><p>String ends with.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.eqNullSafe.html#pyspark.sql.Column.eqNullSafe" title="pyspark.sql.Column.eqNullSafe"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.eqNullSafe</span></code></a>(other)</p></td>
<td><p>Equality test that is safe for null values.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.getField.html#pyspark.sql.Column.getField" title="pyspark.sql.Column.getField"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.getField</span></code></a>(name)</p></td>
<td><p>An expression that gets a field by name in a <code class="xref py py-class docutils literal notranslate"><span class="pre">StructType</span></code>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.getItem.html#pyspark.sql.Column.getItem" title="pyspark.sql.Column.getItem"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.getItem</span></code></a>(key)</p></td>
<td><p>An expression that gets an item at position <code class="docutils literal notranslate"><span class="pre">ordinal</span></code> out of a list, or gets an item by key out of a dict.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.isNotNull.html#pyspark.sql.Column.isNotNull" title="pyspark.sql.Column.isNotNull"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.isNotNull</span></code></a>()</p></td>
<td><p>True if the current expression is NOT null.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.isNull.html#pyspark.sql.Column.isNull" title="pyspark.sql.Column.isNull"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.isNull</span></code></a>()</p></td>
<td><p>True if the current expression is null.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.isin.html#pyspark.sql.Column.isin" title="pyspark.sql.Column.isin"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.isin</span></code></a>(*cols)</p></td>
<td><p>A boolean expression that is evaluated to true if the value of this expression is contained by the evaluated values of the arguments.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.like.html#pyspark.sql.Column.like" title="pyspark.sql.Column.like"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.like</span></code></a>(other)</p></td>
<td><p>SQL like expression.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.name.html#pyspark.sql.Column.name" title="pyspark.sql.Column.name"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.name</span></code></a>(*alias, **kwargs)</p></td>
<td><p><code class="xref py py-func docutils literal notranslate"><span class="pre">name()</span></code> is an alias for <code class="xref py py-func docutils literal notranslate"><span class="pre">alias()</span></code>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.otherwise.html#pyspark.sql.Column.otherwise" title="pyspark.sql.Column.otherwise"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.otherwise</span></code></a>(value)</p></td>
<td><p>Evaluates a list of conditions and returns one of multiple possible result expressions.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.over.html#pyspark.sql.Column.over" title="pyspark.sql.Column.over"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.over</span></code></a>(window)</p></td>
<td><p>Define a windowing column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.rlike.html#pyspark.sql.Column.rlike" title="pyspark.sql.Column.rlike"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.rlike</span></code></a>(other)</p></td>
<td><p>SQL RLIKE expression (LIKE with Regex).</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.startswith.html#pyspark.sql.Column.startswith" title="pyspark.sql.Column.startswith"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.startswith</span></code></a>(other)</p></td>
<td><p>String starts with.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.substr.html#pyspark.sql.Column.substr" title="pyspark.sql.Column.substr"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.substr</span></code></a>(startPos, length)</p></td>
<td><p>Return a <a class="reference internal" href="api/pyspark.sql.Column.html#pyspark.sql.Column" title="pyspark.sql.Column"><code class="xref py py-class docutils literal notranslate"><span class="pre">Column</span></code></a> which is a substring of the column.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Column.when.html#pyspark.sql.Column.when" title="pyspark.sql.Column.when"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.when</span></code></a>(condition, value)</p></td>
<td><p>Evaluates a list of conditions and returns one of multiple possible result expressions.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Column.withField.html#pyspark.sql.Column.withField" title="pyspark.sql.Column.withField"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Column.withField</span></code></a>(fieldName, col)</p></td>
<td><p>An expression that adds/replaces a field in <code class="xref py py-class docutils literal notranslate"><span class="pre">StructType</span></code> by name.</p></td>
</tr>
</tbody>
</table>
</section>
<section id="data-types">
<h2>Data Types<a class="headerlink" href="#data-types" title="Permalink to this headline"></a></h2>
<table class="longtable table autosummary">
<colgroup>
<col style="width: 10%" />
<col style="width: 90%" />
</colgroup>
<tbody>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.types.ArrayType.html#pyspark.sql.types.ArrayType" title="pyspark.sql.types.ArrayType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">ArrayType</span></code></a>(elementType[, containsNull])</p></td>
<td><p>Array data type.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.types.BinaryType.html#pyspark.sql.types.BinaryType" title="pyspark.sql.types.BinaryType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">BinaryType</span></code></a></p></td>
<td><p>Binary (byte array) data type.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.types.BooleanType.html#pyspark.sql.types.BooleanType" title="pyspark.sql.types.BooleanType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">BooleanType</span></code></a></p></td>
<td><p>Boolean data type.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.types.ByteType.html#pyspark.sql.types.ByteType" title="pyspark.sql.types.ByteType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">ByteType</span></code></a></p></td>
<td><p>Byte data type, i.e.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.types.DataType.html#pyspark.sql.types.DataType" title="pyspark.sql.types.DataType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DataType</span></code></a></p></td>
<td><p>Base class for data types.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.types.DateType.html#pyspark.sql.types.DateType" title="pyspark.sql.types.DateType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DateType</span></code></a></p></td>
<td><p>Date (datetime.date) data type.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.types.DecimalType.html#pyspark.sql.types.DecimalType" title="pyspark.sql.types.DecimalType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DecimalType</span></code></a>([precision, scale])</p></td>
<td><p>Decimal (decimal.Decimal) data type.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.types.DoubleType.html#pyspark.sql.types.DoubleType" title="pyspark.sql.types.DoubleType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">DoubleType</span></code></a></p></td>
<td><p>Double data type, representing double precision floats.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.types.FloatType.html#pyspark.sql.types.FloatType" title="pyspark.sql.types.FloatType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">FloatType</span></code></a></p></td>
<td><p>Float data type, representing single precision floats.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.types.IntegerType.html#pyspark.sql.types.IntegerType" title="pyspark.sql.types.IntegerType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">IntegerType</span></code></a></p></td>
<td><p>Int data type, i.e.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.types.LongType.html#pyspark.sql.types.LongType" title="pyspark.sql.types.LongType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">LongType</span></code></a></p></td>
<td><p>Long data type, i.e.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.types.MapType.html#pyspark.sql.types.MapType" title="pyspark.sql.types.MapType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">MapType</span></code></a>(keyType, valueType[, valueContainsNull])</p></td>
<td><p>Map data type.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.types.NullType.html#pyspark.sql.types.NullType" title="pyspark.sql.types.NullType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">NullType</span></code></a></p></td>
<td><p>Null type.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.types.ShortType.html#pyspark.sql.types.ShortType" title="pyspark.sql.types.ShortType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">ShortType</span></code></a></p></td>
<td><p>Short data type, i.e.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.types.StringType.html#pyspark.sql.types.StringType" title="pyspark.sql.types.StringType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">StringType</span></code></a></p></td>
<td><p>String data type.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.types.StructField.html#pyspark.sql.types.StructField" title="pyspark.sql.types.StructField"><code class="xref py py-obj docutils literal notranslate"><span class="pre">StructField</span></code></a>(name, dataType[, nullable, metadata])</p></td>
<td><p>A field in <a class="reference internal" href="api/pyspark.sql.types.StructType.html#pyspark.sql.types.StructType" title="pyspark.sql.types.StructType"><code class="xref py py-class docutils literal notranslate"><span class="pre">StructType</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.types.StructType.html#pyspark.sql.types.StructType" title="pyspark.sql.types.StructType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">StructType</span></code></a>([fields])</p></td>
<td><p>Struct type, consisting of a list of <a class="reference internal" href="api/pyspark.sql.types.StructField.html#pyspark.sql.types.StructField" title="pyspark.sql.types.StructField"><code class="xref py py-class docutils literal notranslate"><span class="pre">StructField</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.types.TimestampType.html#pyspark.sql.types.TimestampType" title="pyspark.sql.types.TimestampType"><code class="xref py py-obj docutils literal notranslate"><span class="pre">TimestampType</span></code></a></p></td>
<td><p>Timestamp (datetime.datetime) data type.</p></td>
</tr>
</tbody>
</table>
</section>
<section id="row">
<h2>Row<a class="headerlink" href="#row" title="Permalink to this headline"></a></h2>
<table class="longtable table autosummary">
<colgroup>
<col style="width: 10%" />
<col style="width: 90%" />
</colgroup>
<tbody>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Row.asDict.html#pyspark.sql.Row.asDict" title="pyspark.sql.Row.asDict"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Row.asDict</span></code></a>([recursive])</p></td>
<td><p>Return as a dict</p></td>
</tr>
</tbody>
</table>
</section>
<section id="functions">
<h2>Functions<a class="headerlink" href="#functions" title="Permalink to this headline"></a></h2>
<table class="longtable table autosummary">
<colgroup>
<col style="width: 10%" />
<col style="width: 90%" />
</colgroup>
<tbody>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.abs.html#pyspark.sql.functions.abs" title="pyspark.sql.functions.abs"><code class="xref py py-obj docutils literal notranslate"><span class="pre">abs</span></code></a>(col)</p></td>
<td><p>Computes the absolute value.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.acos.html#pyspark.sql.functions.acos" title="pyspark.sql.functions.acos"><code class="xref py py-obj docutils literal notranslate"><span class="pre">acos</span></code></a>(col)</p></td>
<td><p><div class="versionadded">
<p><span class="versionmodified added">New in version 1.4.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.acosh.html#pyspark.sql.functions.acosh" title="pyspark.sql.functions.acosh"><code class="xref py py-obj docutils literal notranslate"><span class="pre">acosh</span></code></a>(col)</p></td>
<td><p>Computes inverse hyperbolic cosine of the input column.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.add_months.html#pyspark.sql.functions.add_months" title="pyspark.sql.functions.add_months"><code class="xref py py-obj docutils literal notranslate"><span class="pre">add_months</span></code></a>(start, months)</p></td>
<td><p>Returns the date that is <cite>months</cite> months after <cite>start</cite></p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.aggregate.html#pyspark.sql.functions.aggregate" title="pyspark.sql.functions.aggregate"><code class="xref py py-obj docutils literal notranslate"><span class="pre">aggregate</span></code></a>(col, initialValue, merge[, finish])</p></td>
<td><p>Applies a binary operator to an initial state and all elements in the array, and reduces this to a single state.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.approxCountDistinct.html#pyspark.sql.functions.approxCountDistinct" title="pyspark.sql.functions.approxCountDistinct"><code class="xref py py-obj docutils literal notranslate"><span class="pre">approxCountDistinct</span></code></a>(col[, rsd])</p></td>
<td><p><div class="deprecated">
<p><span class="versionmodified deprecated">Deprecated since version 2.1.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.approx_count_distinct.html#pyspark.sql.functions.approx_count_distinct" title="pyspark.sql.functions.approx_count_distinct"><code class="xref py py-obj docutils literal notranslate"><span class="pre">approx_count_distinct</span></code></a>(col[, rsd])</p></td>
<td><p>Aggregate function: returns a new <a class="reference internal" href="api/pyspark.sql.Column.html#pyspark.sql.Column" title="pyspark.sql.Column"><code class="xref py py-class docutils literal notranslate"><span class="pre">Column</span></code></a> for approximate distinct count of column <cite>col</cite>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.array.html#pyspark.sql.functions.array" title="pyspark.sql.functions.array"><code class="xref py py-obj docutils literal notranslate"><span class="pre">array</span></code></a>(*cols)</p></td>
<td><p>Creates a new array column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.array_contains.html#pyspark.sql.functions.array_contains" title="pyspark.sql.functions.array_contains"><code class="xref py py-obj docutils literal notranslate"><span class="pre">array_contains</span></code></a>(col, value)</p></td>
<td><p>Collection function: returns null if the array is null, true if the array contains the given value, and false otherwise.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.array_distinct.html#pyspark.sql.functions.array_distinct" title="pyspark.sql.functions.array_distinct"><code class="xref py py-obj docutils literal notranslate"><span class="pre">array_distinct</span></code></a>(col)</p></td>
<td><p>Collection function: removes duplicate values from the array.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.array_except.html#pyspark.sql.functions.array_except" title="pyspark.sql.functions.array_except"><code class="xref py py-obj docutils literal notranslate"><span class="pre">array_except</span></code></a>(col1, col2)</p></td>
<td><p>Collection function: returns an array of the elements in col1 but not in col2, without duplicates.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.array_intersect.html#pyspark.sql.functions.array_intersect" title="pyspark.sql.functions.array_intersect"><code class="xref py py-obj docutils literal notranslate"><span class="pre">array_intersect</span></code></a>(col1, col2)</p></td>
<td><p>Collection function: returns an array of the elements in the intersection of col1 and col2, without duplicates.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.array_join.html#pyspark.sql.functions.array_join" title="pyspark.sql.functions.array_join"><code class="xref py py-obj docutils literal notranslate"><span class="pre">array_join</span></code></a>(col, delimiter[, null_replacement])</p></td>
<td><p>Concatenates the elements of <cite>column</cite> using the <cite>delimiter</cite>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.array_max.html#pyspark.sql.functions.array_max" title="pyspark.sql.functions.array_max"><code class="xref py py-obj docutils literal notranslate"><span class="pre">array_max</span></code></a>(col)</p></td>
<td><p>Collection function: returns the maximum value of the array.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.array_min.html#pyspark.sql.functions.array_min" title="pyspark.sql.functions.array_min"><code class="xref py py-obj docutils literal notranslate"><span class="pre">array_min</span></code></a>(col)</p></td>
<td><p>Collection function: returns the minimum value of the array.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.array_position.html#pyspark.sql.functions.array_position" title="pyspark.sql.functions.array_position"><code class="xref py py-obj docutils literal notranslate"><span class="pre">array_position</span></code></a>(col, value)</p></td>
<td><p>Collection function: Locates the position of the first occurrence of the given value in the given array.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.array_remove.html#pyspark.sql.functions.array_remove" title="pyspark.sql.functions.array_remove"><code class="xref py py-obj docutils literal notranslate"><span class="pre">array_remove</span></code></a>(col, element)</p></td>
<td><p>Collection function: Remove all elements that equal to element from the given array.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.array_repeat.html#pyspark.sql.functions.array_repeat" title="pyspark.sql.functions.array_repeat"><code class="xref py py-obj docutils literal notranslate"><span class="pre">array_repeat</span></code></a>(col, count)</p></td>
<td><p>Collection function: creates an array containing a column repeated count times.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.array_sort.html#pyspark.sql.functions.array_sort" title="pyspark.sql.functions.array_sort"><code class="xref py py-obj docutils literal notranslate"><span class="pre">array_sort</span></code></a>(col)</p></td>
<td><p>Collection function: sorts the input array in ascending order.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.array_union.html#pyspark.sql.functions.array_union" title="pyspark.sql.functions.array_union"><code class="xref py py-obj docutils literal notranslate"><span class="pre">array_union</span></code></a>(col1, col2)</p></td>
<td><p>Collection function: returns an array of the elements in the union of col1 and col2, without duplicates.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.arrays_overlap.html#pyspark.sql.functions.arrays_overlap" title="pyspark.sql.functions.arrays_overlap"><code class="xref py py-obj docutils literal notranslate"><span class="pre">arrays_overlap</span></code></a>(a1, a2)</p></td>
<td><p>Collection function: returns true if the arrays contain any common non-null element; if not, returns null if both the arrays are non-empty and any of them contains a null element; returns false otherwise.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.arrays_zip.html#pyspark.sql.functions.arrays_zip" title="pyspark.sql.functions.arrays_zip"><code class="xref py py-obj docutils literal notranslate"><span class="pre">arrays_zip</span></code></a>(*cols)</p></td>
<td><p>Collection function: Returns a merged array of structs in which the N-th struct contains all N-th values of input arrays.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.asc.html#pyspark.sql.functions.asc" title="pyspark.sql.functions.asc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">asc</span></code></a>(col)</p></td>
<td><p>Returns a sort expression based on the ascending order of the given column name.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.asc_nulls_first.html#pyspark.sql.functions.asc_nulls_first" title="pyspark.sql.functions.asc_nulls_first"><code class="xref py py-obj docutils literal notranslate"><span class="pre">asc_nulls_first</span></code></a>(col)</p></td>
<td><p>Returns a sort expression based on the ascending order of the given column name, and null values return before non-null values.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.asc_nulls_last.html#pyspark.sql.functions.asc_nulls_last" title="pyspark.sql.functions.asc_nulls_last"><code class="xref py py-obj docutils literal notranslate"><span class="pre">asc_nulls_last</span></code></a>(col)</p></td>
<td><p>Returns a sort expression based on the ascending order of the given column name, and null values appear after non-null values.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.ascii.html#pyspark.sql.functions.ascii" title="pyspark.sql.functions.ascii"><code class="xref py py-obj docutils literal notranslate"><span class="pre">ascii</span></code></a>(col)</p></td>
<td><p>Computes the numeric value of the first character of the string column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.asin.html#pyspark.sql.functions.asin" title="pyspark.sql.functions.asin"><code class="xref py py-obj docutils literal notranslate"><span class="pre">asin</span></code></a>(col)</p></td>
<td><p><div class="versionadded">
<p><span class="versionmodified added">New in version 1.3.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.asinh.html#pyspark.sql.functions.asinh" title="pyspark.sql.functions.asinh"><code class="xref py py-obj docutils literal notranslate"><span class="pre">asinh</span></code></a>(col)</p></td>
<td><p>Computes inverse hyperbolic sine of the input column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.assert_true.html#pyspark.sql.functions.assert_true" title="pyspark.sql.functions.assert_true"><code class="xref py py-obj docutils literal notranslate"><span class="pre">assert_true</span></code></a>(col[, errMsg])</p></td>
<td><p>Returns null if the input column is true; throws an exception with the provided error message otherwise.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.atan.html#pyspark.sql.functions.atan" title="pyspark.sql.functions.atan"><code class="xref py py-obj docutils literal notranslate"><span class="pre">atan</span></code></a>(col)</p></td>
<td><p><div class="versionadded">
<p><span class="versionmodified added">New in version 1.4.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.atanh.html#pyspark.sql.functions.atanh" title="pyspark.sql.functions.atanh"><code class="xref py py-obj docutils literal notranslate"><span class="pre">atanh</span></code></a>(col)</p></td>
<td><p>Computes inverse hyperbolic tangent of the input column.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.atan2.html#pyspark.sql.functions.atan2" title="pyspark.sql.functions.atan2"><code class="xref py py-obj docutils literal notranslate"><span class="pre">atan2</span></code></a>(col1, col2)</p></td>
<td><p><div class="versionadded">
<p><span class="versionmodified added">New in version 1.4.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.avg.html#pyspark.sql.functions.avg" title="pyspark.sql.functions.avg"><code class="xref py py-obj docutils literal notranslate"><span class="pre">avg</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the average of the values in a group.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.base64.html#pyspark.sql.functions.base64" title="pyspark.sql.functions.base64"><code class="xref py py-obj docutils literal notranslate"><span class="pre">base64</span></code></a>(col)</p></td>
<td><p>Computes the BASE64 encoding of a binary column and returns it as a string column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.bin.html#pyspark.sql.functions.bin" title="pyspark.sql.functions.bin"><code class="xref py py-obj docutils literal notranslate"><span class="pre">bin</span></code></a>(col)</p></td>
<td><p>Returns the string representation of the binary value of the given column.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.bitwise_not.html#pyspark.sql.functions.bitwise_not" title="pyspark.sql.functions.bitwise_not"><code class="xref py py-obj docutils literal notranslate"><span class="pre">bitwise_not</span></code></a>(col)</p></td>
<td><p>Computes bitwise not.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.bitwiseNOT.html#pyspark.sql.functions.bitwiseNOT" title="pyspark.sql.functions.bitwiseNOT"><code class="xref py py-obj docutils literal notranslate"><span class="pre">bitwiseNOT</span></code></a>(col)</p></td>
<td><p>Computes bitwise not.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.broadcast.html#pyspark.sql.functions.broadcast" title="pyspark.sql.functions.broadcast"><code class="xref py py-obj docutils literal notranslate"><span class="pre">broadcast</span></code></a>(df)</p></td>
<td><p>Marks a DataFrame as small enough for use in broadcast joins.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.bround.html#pyspark.sql.functions.bround" title="pyspark.sql.functions.bround"><code class="xref py py-obj docutils literal notranslate"><span class="pre">bround</span></code></a>(col[, scale])</p></td>
<td><p>Round the given value to <cite>scale</cite> decimal places using HALF_EVEN rounding mode if <cite>scale</cite> &gt;= 0 or at integral part when <cite>scale</cite> &lt; 0.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.bucket.html#pyspark.sql.functions.bucket" title="pyspark.sql.functions.bucket"><code class="xref py py-obj docutils literal notranslate"><span class="pre">bucket</span></code></a>(numBuckets, col)</p></td>
<td><p>Partition transform function: A transform for any type that partitions by a hash of the input column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.cbrt.html#pyspark.sql.functions.cbrt" title="pyspark.sql.functions.cbrt"><code class="xref py py-obj docutils literal notranslate"><span class="pre">cbrt</span></code></a>(col)</p></td>
<td><p>Computes the cube-root of the given value.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.ceil.html#pyspark.sql.functions.ceil" title="pyspark.sql.functions.ceil"><code class="xref py py-obj docutils literal notranslate"><span class="pre">ceil</span></code></a>(col)</p></td>
<td><p>Computes the ceiling of the given value.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.coalesce.html#pyspark.sql.functions.coalesce" title="pyspark.sql.functions.coalesce"><code class="xref py py-obj docutils literal notranslate"><span class="pre">coalesce</span></code></a>(*cols)</p></td>
<td><p>Returns the first column that is not null.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.col.html#pyspark.sql.functions.col" title="pyspark.sql.functions.col"><code class="xref py py-obj docutils literal notranslate"><span class="pre">col</span></code></a>(col)</p></td>
<td><p>Returns a <a class="reference internal" href="api/pyspark.sql.Column.html#pyspark.sql.Column" title="pyspark.sql.Column"><code class="xref py py-class docutils literal notranslate"><span class="pre">Column</span></code></a> based on the given column name.’ Examples ——– &gt;&gt;&gt; col(‘x’) Column&lt;’x’&gt; &gt;&gt;&gt; column(‘x’) Column&lt;’x’&gt;</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.collect_list.html#pyspark.sql.functions.collect_list" title="pyspark.sql.functions.collect_list"><code class="xref py py-obj docutils literal notranslate"><span class="pre">collect_list</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns a list of objects with duplicates.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.collect_set.html#pyspark.sql.functions.collect_set" title="pyspark.sql.functions.collect_set"><code class="xref py py-obj docutils literal notranslate"><span class="pre">collect_set</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns a set of objects with duplicate elements eliminated.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.column.html#pyspark.sql.functions.column" title="pyspark.sql.functions.column"><code class="xref py py-obj docutils literal notranslate"><span class="pre">column</span></code></a>(col)</p></td>
<td><p>Returns a <a class="reference internal" href="api/pyspark.sql.Column.html#pyspark.sql.Column" title="pyspark.sql.Column"><code class="xref py py-class docutils literal notranslate"><span class="pre">Column</span></code></a> based on the given column name.’ Examples ——– &gt;&gt;&gt; col(‘x’) Column&lt;’x’&gt; &gt;&gt;&gt; column(‘x’) Column&lt;’x’&gt;</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.concat.html#pyspark.sql.functions.concat" title="pyspark.sql.functions.concat"><code class="xref py py-obj docutils literal notranslate"><span class="pre">concat</span></code></a>(*cols)</p></td>
<td><p>Concatenates multiple input columns together into a single column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.concat_ws.html#pyspark.sql.functions.concat_ws" title="pyspark.sql.functions.concat_ws"><code class="xref py py-obj docutils literal notranslate"><span class="pre">concat_ws</span></code></a>(sep, *cols)</p></td>
<td><p>Concatenates multiple input string columns together into a single string column, using the given separator.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.conv.html#pyspark.sql.functions.conv" title="pyspark.sql.functions.conv"><code class="xref py py-obj docutils literal notranslate"><span class="pre">conv</span></code></a>(col, fromBase, toBase)</p></td>
<td><p>Convert a number in a string column from one base to another.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.corr.html#pyspark.sql.functions.corr" title="pyspark.sql.functions.corr"><code class="xref py py-obj docutils literal notranslate"><span class="pre">corr</span></code></a>(col1, col2)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.Column.html#pyspark.sql.Column" title="pyspark.sql.Column"><code class="xref py py-class docutils literal notranslate"><span class="pre">Column</span></code></a> for the Pearson Correlation Coefficient for <code class="docutils literal notranslate"><span class="pre">col1</span></code> and <code class="docutils literal notranslate"><span class="pre">col2</span></code>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.cos.html#pyspark.sql.functions.cos" title="pyspark.sql.functions.cos"><code class="xref py py-obj docutils literal notranslate"><span class="pre">cos</span></code></a>(col)</p></td>
<td><p><div class="versionadded">
<p><span class="versionmodified added">New in version 1.4.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.cosh.html#pyspark.sql.functions.cosh" title="pyspark.sql.functions.cosh"><code class="xref py py-obj docutils literal notranslate"><span class="pre">cosh</span></code></a>(col)</p></td>
<td><p><div class="versionadded">
<p><span class="versionmodified added">New in version 1.4.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.count.html#pyspark.sql.functions.count" title="pyspark.sql.functions.count"><code class="xref py py-obj docutils literal notranslate"><span class="pre">count</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the number of items in a group.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.count_distinct.html#pyspark.sql.functions.count_distinct" title="pyspark.sql.functions.count_distinct"><code class="xref py py-obj docutils literal notranslate"><span class="pre">count_distinct</span></code></a>(col, *cols)</p></td>
<td><p>Returns a new <code class="xref py py-class docutils literal notranslate"><span class="pre">Column</span></code> for distinct count of <code class="docutils literal notranslate"><span class="pre">col</span></code> or <code class="docutils literal notranslate"><span class="pre">cols</span></code>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.countDistinct.html#pyspark.sql.functions.countDistinct" title="pyspark.sql.functions.countDistinct"><code class="xref py py-obj docutils literal notranslate"><span class="pre">countDistinct</span></code></a>(col, *cols)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.Column.html#pyspark.sql.Column" title="pyspark.sql.Column"><code class="xref py py-class docutils literal notranslate"><span class="pre">Column</span></code></a> for distinct count of <code class="docutils literal notranslate"><span class="pre">col</span></code> or <code class="docutils literal notranslate"><span class="pre">cols</span></code>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.covar_pop.html#pyspark.sql.functions.covar_pop" title="pyspark.sql.functions.covar_pop"><code class="xref py py-obj docutils literal notranslate"><span class="pre">covar_pop</span></code></a>(col1, col2)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.Column.html#pyspark.sql.Column" title="pyspark.sql.Column"><code class="xref py py-class docutils literal notranslate"><span class="pre">Column</span></code></a> for the population covariance of <code class="docutils literal notranslate"><span class="pre">col1</span></code> and <code class="docutils literal notranslate"><span class="pre">col2</span></code>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.covar_samp.html#pyspark.sql.functions.covar_samp" title="pyspark.sql.functions.covar_samp"><code class="xref py py-obj docutils literal notranslate"><span class="pre">covar_samp</span></code></a>(col1, col2)</p></td>
<td><p>Returns a new <a class="reference internal" href="api/pyspark.sql.Column.html#pyspark.sql.Column" title="pyspark.sql.Column"><code class="xref py py-class docutils literal notranslate"><span class="pre">Column</span></code></a> for the sample covariance of <code class="docutils literal notranslate"><span class="pre">col1</span></code> and <code class="docutils literal notranslate"><span class="pre">col2</span></code>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.crc32.html#pyspark.sql.functions.crc32" title="pyspark.sql.functions.crc32"><code class="xref py py-obj docutils literal notranslate"><span class="pre">crc32</span></code></a>(col)</p></td>
<td><p>Calculates the cyclic redundancy check value (CRC32) of a binary column and returns the value as a bigint.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.create_map.html#pyspark.sql.functions.create_map" title="pyspark.sql.functions.create_map"><code class="xref py py-obj docutils literal notranslate"><span class="pre">create_map</span></code></a>(*cols)</p></td>
<td><p>Creates a new map column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.cume_dist.html#pyspark.sql.functions.cume_dist" title="pyspark.sql.functions.cume_dist"><code class="xref py py-obj docutils literal notranslate"><span class="pre">cume_dist</span></code></a>()</p></td>
<td><p>Window function: returns the cumulative distribution of values within a window partition, i.e.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.current_date.html#pyspark.sql.functions.current_date" title="pyspark.sql.functions.current_date"><code class="xref py py-obj docutils literal notranslate"><span class="pre">current_date</span></code></a>()</p></td>
<td><p>Returns the current date at the start of query evaluation as a <code class="xref py py-class docutils literal notranslate"><span class="pre">DateType</span></code> column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.current_timestamp.html#pyspark.sql.functions.current_timestamp" title="pyspark.sql.functions.current_timestamp"><code class="xref py py-obj docutils literal notranslate"><span class="pre">current_timestamp</span></code></a>()</p></td>
<td><p>Returns the current timestamp at the start of query evaluation as a <code class="xref py py-class docutils literal notranslate"><span class="pre">TimestampType</span></code> column.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.date_add.html#pyspark.sql.functions.date_add" title="pyspark.sql.functions.date_add"><code class="xref py py-obj docutils literal notranslate"><span class="pre">date_add</span></code></a>(start, days)</p></td>
<td><p>Returns the date that is <cite>days</cite> days after <cite>start</cite></p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.date_format.html#pyspark.sql.functions.date_format" title="pyspark.sql.functions.date_format"><code class="xref py py-obj docutils literal notranslate"><span class="pre">date_format</span></code></a>(date, format)</p></td>
<td><p>Converts a date/timestamp/string to a value of string in the format specified by the date format given by the second argument.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.date_sub.html#pyspark.sql.functions.date_sub" title="pyspark.sql.functions.date_sub"><code class="xref py py-obj docutils literal notranslate"><span class="pre">date_sub</span></code></a>(start, days)</p></td>
<td><p>Returns the date that is <cite>days</cite> days before <cite>start</cite></p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.date_trunc.html#pyspark.sql.functions.date_trunc" title="pyspark.sql.functions.date_trunc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">date_trunc</span></code></a>(format, timestamp)</p></td>
<td><p>Returns timestamp truncated to the unit specified by the format.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.datediff.html#pyspark.sql.functions.datediff" title="pyspark.sql.functions.datediff"><code class="xref py py-obj docutils literal notranslate"><span class="pre">datediff</span></code></a>(end, start)</p></td>
<td><p>Returns the number of days from <cite>start</cite> to <cite>end</cite>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.dayofmonth.html#pyspark.sql.functions.dayofmonth" title="pyspark.sql.functions.dayofmonth"><code class="xref py py-obj docutils literal notranslate"><span class="pre">dayofmonth</span></code></a>(col)</p></td>
<td><p>Extract the day of the month of a given date as integer.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.dayofweek.html#pyspark.sql.functions.dayofweek" title="pyspark.sql.functions.dayofweek"><code class="xref py py-obj docutils literal notranslate"><span class="pre">dayofweek</span></code></a>(col)</p></td>
<td><p>Extract the day of the week of a given date as integer.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.dayofyear.html#pyspark.sql.functions.dayofyear" title="pyspark.sql.functions.dayofyear"><code class="xref py py-obj docutils literal notranslate"><span class="pre">dayofyear</span></code></a>(col)</p></td>
<td><p>Extract the day of the year of a given date as integer.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.days.html#pyspark.sql.functions.days" title="pyspark.sql.functions.days"><code class="xref py py-obj docutils literal notranslate"><span class="pre">days</span></code></a>(col)</p></td>
<td><p>Partition transform function: A transform for timestamps and dates to partition data into days.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.decode.html#pyspark.sql.functions.decode" title="pyspark.sql.functions.decode"><code class="xref py py-obj docutils literal notranslate"><span class="pre">decode</span></code></a>(col, charset)</p></td>
<td><p>Computes the first argument into a string from a binary using the provided character set (one of ‘US-ASCII’, ‘ISO-8859-1’, ‘UTF-8’, ‘UTF-16BE’, ‘UTF-16LE’, ‘UTF-16’).</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.degrees.html#pyspark.sql.functions.degrees" title="pyspark.sql.functions.degrees"><code class="xref py py-obj docutils literal notranslate"><span class="pre">degrees</span></code></a>(col)</p></td>
<td><p>Converts an angle measured in radians to an approximately equivalent angle measured in degrees.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.dense_rank.html#pyspark.sql.functions.dense_rank" title="pyspark.sql.functions.dense_rank"><code class="xref py py-obj docutils literal notranslate"><span class="pre">dense_rank</span></code></a>()</p></td>
<td><p>Window function: returns the rank of rows within a window partition, without any gaps.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.desc.html#pyspark.sql.functions.desc" title="pyspark.sql.functions.desc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">desc</span></code></a>(col)</p></td>
<td><p>Returns a sort expression based on the descending order of the given column name.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.desc_nulls_first.html#pyspark.sql.functions.desc_nulls_first" title="pyspark.sql.functions.desc_nulls_first"><code class="xref py py-obj docutils literal notranslate"><span class="pre">desc_nulls_first</span></code></a>(col)</p></td>
<td><p>Returns a sort expression based on the descending order of the given column name, and null values appear before non-null values.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.desc_nulls_last.html#pyspark.sql.functions.desc_nulls_last" title="pyspark.sql.functions.desc_nulls_last"><code class="xref py py-obj docutils literal notranslate"><span class="pre">desc_nulls_last</span></code></a>(col)</p></td>
<td><p>Returns a sort expression based on the descending order of the given column name, and null values appear after non-null values.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.element_at.html#pyspark.sql.functions.element_at" title="pyspark.sql.functions.element_at"><code class="xref py py-obj docutils literal notranslate"><span class="pre">element_at</span></code></a>(col, extraction)</p></td>
<td><p>Collection function: Returns element of array at given index in extraction if col is array.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.encode.html#pyspark.sql.functions.encode" title="pyspark.sql.functions.encode"><code class="xref py py-obj docutils literal notranslate"><span class="pre">encode</span></code></a>(col, charset)</p></td>
<td><p>Computes the first argument into a binary from a string using the provided character set (one of ‘US-ASCII’, ‘ISO-8859-1’, ‘UTF-8’, ‘UTF-16BE’, ‘UTF-16LE’, ‘UTF-16’).</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.exists.html#pyspark.sql.functions.exists" title="pyspark.sql.functions.exists"><code class="xref py py-obj docutils literal notranslate"><span class="pre">exists</span></code></a>(col, f)</p></td>
<td><p>Returns whether a predicate holds for one or more elements in the array.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.exp.html#pyspark.sql.functions.exp" title="pyspark.sql.functions.exp"><code class="xref py py-obj docutils literal notranslate"><span class="pre">exp</span></code></a>(col)</p></td>
<td><p>Computes the exponential of the given value.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.explode.html#pyspark.sql.functions.explode" title="pyspark.sql.functions.explode"><code class="xref py py-obj docutils literal notranslate"><span class="pre">explode</span></code></a>(col)</p></td>
<td><p>Returns a new row for each element in the given array or map.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.explode_outer.html#pyspark.sql.functions.explode_outer" title="pyspark.sql.functions.explode_outer"><code class="xref py py-obj docutils literal notranslate"><span class="pre">explode_outer</span></code></a>(col)</p></td>
<td><p>Returns a new row for each element in the given array or map.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.expm1.html#pyspark.sql.functions.expm1" title="pyspark.sql.functions.expm1"><code class="xref py py-obj docutils literal notranslate"><span class="pre">expm1</span></code></a>(col)</p></td>
<td><p>Computes the exponential of the given value minus one.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.expr.html#pyspark.sql.functions.expr" title="pyspark.sql.functions.expr"><code class="xref py py-obj docutils literal notranslate"><span class="pre">expr</span></code></a>(str)</p></td>
<td><p>Parses the expression string into the column that it represents</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.factorial.html#pyspark.sql.functions.factorial" title="pyspark.sql.functions.factorial"><code class="xref py py-obj docutils literal notranslate"><span class="pre">factorial</span></code></a>(col)</p></td>
<td><p>Computes the factorial of the given value.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.filter.html#pyspark.sql.functions.filter" title="pyspark.sql.functions.filter"><code class="xref py py-obj docutils literal notranslate"><span class="pre">filter</span></code></a>(col, f)</p></td>
<td><p>Returns an array of elements for which a predicate holds in a given array.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.first.html#pyspark.sql.functions.first" title="pyspark.sql.functions.first"><code class="xref py py-obj docutils literal notranslate"><span class="pre">first</span></code></a>(col[, ignorenulls])</p></td>
<td><p>Aggregate function: returns the first value in a group.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.flatten.html#pyspark.sql.functions.flatten" title="pyspark.sql.functions.flatten"><code class="xref py py-obj docutils literal notranslate"><span class="pre">flatten</span></code></a>(col)</p></td>
<td><p>Collection function: creates a single array from an array of arrays.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.floor.html#pyspark.sql.functions.floor" title="pyspark.sql.functions.floor"><code class="xref py py-obj docutils literal notranslate"><span class="pre">floor</span></code></a>(col)</p></td>
<td><p>Computes the floor of the given value.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.forall.html#pyspark.sql.functions.forall" title="pyspark.sql.functions.forall"><code class="xref py py-obj docutils literal notranslate"><span class="pre">forall</span></code></a>(col, f)</p></td>
<td><p>Returns whether a predicate holds for every element in the array.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.format_number.html#pyspark.sql.functions.format_number" title="pyspark.sql.functions.format_number"><code class="xref py py-obj docutils literal notranslate"><span class="pre">format_number</span></code></a>(col, d)</p></td>
<td><p>Formats the number X to a format like ‘#,–#,–#.–’, rounded to d decimal places with HALF_EVEN round mode, and returns the result as a string.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.format_string.html#pyspark.sql.functions.format_string" title="pyspark.sql.functions.format_string"><code class="xref py py-obj docutils literal notranslate"><span class="pre">format_string</span></code></a>(format, *cols)</p></td>
<td><p>Formats the arguments in printf-style and returns the result as a string column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.from_csv.html#pyspark.sql.functions.from_csv" title="pyspark.sql.functions.from_csv"><code class="xref py py-obj docutils literal notranslate"><span class="pre">from_csv</span></code></a>(col, schema[, options])</p></td>
<td><p>Parses a column containing a CSV string to a row with the specified schema.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.from_json.html#pyspark.sql.functions.from_json" title="pyspark.sql.functions.from_json"><code class="xref py py-obj docutils literal notranslate"><span class="pre">from_json</span></code></a>(col, schema[, options])</p></td>
<td><p>Parses a column containing a JSON string into a <code class="xref py py-class docutils literal notranslate"><span class="pre">MapType</span></code> with <code class="xref py py-class docutils literal notranslate"><span class="pre">StringType</span></code> as keys type, <code class="xref py py-class docutils literal notranslate"><span class="pre">StructType</span></code> or <code class="xref py py-class docutils literal notranslate"><span class="pre">ArrayType</span></code> with the specified schema.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.from_unixtime.html#pyspark.sql.functions.from_unixtime" title="pyspark.sql.functions.from_unixtime"><code class="xref py py-obj docutils literal notranslate"><span class="pre">from_unixtime</span></code></a>(timestamp[, format])</p></td>
<td><p>Converts the number of seconds from unix epoch (1970-01-01 00:00:00 UTC) to a string representing the timestamp of that moment in the current system time zone in the given format.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.from_utc_timestamp.html#pyspark.sql.functions.from_utc_timestamp" title="pyspark.sql.functions.from_utc_timestamp"><code class="xref py py-obj docutils literal notranslate"><span class="pre">from_utc_timestamp</span></code></a>(timestamp, tz)</p></td>
<td><p>This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.get_json_object.html#pyspark.sql.functions.get_json_object" title="pyspark.sql.functions.get_json_object"><code class="xref py py-obj docutils literal notranslate"><span class="pre">get_json_object</span></code></a>(col, path)</p></td>
<td><p>Extracts json object from a json string based on json path specified, and returns json string of the extracted json object.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.greatest.html#pyspark.sql.functions.greatest" title="pyspark.sql.functions.greatest"><code class="xref py py-obj docutils literal notranslate"><span class="pre">greatest</span></code></a>(*cols)</p></td>
<td><p>Returns the greatest value of the list of column names, skipping null values.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.grouping.html#pyspark.sql.functions.grouping" title="pyspark.sql.functions.grouping"><code class="xref py py-obj docutils literal notranslate"><span class="pre">grouping</span></code></a>(col)</p></td>
<td><p>Aggregate function: indicates whether a specified column in a GROUP BY list is aggregated or not, returns 1 for aggregated or 0 for not aggregated in the result set.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.grouping_id.html#pyspark.sql.functions.grouping_id" title="pyspark.sql.functions.grouping_id"><code class="xref py py-obj docutils literal notranslate"><span class="pre">grouping_id</span></code></a>(*cols)</p></td>
<td><p>Aggregate function: returns the level of grouping, equals to</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.hash.html#pyspark.sql.functions.hash" title="pyspark.sql.functions.hash"><code class="xref py py-obj docutils literal notranslate"><span class="pre">hash</span></code></a>(*cols)</p></td>
<td><p>Calculates the hash code of given columns, and returns the result as an int column.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.hex.html#pyspark.sql.functions.hex" title="pyspark.sql.functions.hex"><code class="xref py py-obj docutils literal notranslate"><span class="pre">hex</span></code></a>(col)</p></td>
<td><p>Computes hex value of the given column, which could be <a class="reference internal" href="api/pyspark.sql.types.StringType.html#pyspark.sql.types.StringType" title="pyspark.sql.types.StringType"><code class="xref py py-class docutils literal notranslate"><span class="pre">pyspark.sql.types.StringType</span></code></a>, <a class="reference internal" href="api/pyspark.sql.types.BinaryType.html#pyspark.sql.types.BinaryType" title="pyspark.sql.types.BinaryType"><code class="xref py py-class docutils literal notranslate"><span class="pre">pyspark.sql.types.BinaryType</span></code></a>, <a class="reference internal" href="api/pyspark.sql.types.IntegerType.html#pyspark.sql.types.IntegerType" title="pyspark.sql.types.IntegerType"><code class="xref py py-class docutils literal notranslate"><span class="pre">pyspark.sql.types.IntegerType</span></code></a> or <a class="reference internal" href="api/pyspark.sql.types.LongType.html#pyspark.sql.types.LongType" title="pyspark.sql.types.LongType"><code class="xref py py-class docutils literal notranslate"><span class="pre">pyspark.sql.types.LongType</span></code></a>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.hour.html#pyspark.sql.functions.hour" title="pyspark.sql.functions.hour"><code class="xref py py-obj docutils literal notranslate"><span class="pre">hour</span></code></a>(col)</p></td>
<td><p>Extract the hours of a given date as integer.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.hours.html#pyspark.sql.functions.hours" title="pyspark.sql.functions.hours"><code class="xref py py-obj docutils literal notranslate"><span class="pre">hours</span></code></a>(col)</p></td>
<td><p>Partition transform function: A transform for timestamps to partition data into hours.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.hypot.html#pyspark.sql.functions.hypot" title="pyspark.sql.functions.hypot"><code class="xref py py-obj docutils literal notranslate"><span class="pre">hypot</span></code></a>(col1, col2)</p></td>
<td><p>Computes <code class="docutils literal notranslate"><span class="pre">sqrt(a^2</span> <span class="pre">+</span> <span class="pre">b^2)</span></code> without intermediate overflow or underflow.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.initcap.html#pyspark.sql.functions.initcap" title="pyspark.sql.functions.initcap"><code class="xref py py-obj docutils literal notranslate"><span class="pre">initcap</span></code></a>(col)</p></td>
<td><p>Translate the first letter of each word to upper case in the sentence.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.input_file_name.html#pyspark.sql.functions.input_file_name" title="pyspark.sql.functions.input_file_name"><code class="xref py py-obj docutils literal notranslate"><span class="pre">input_file_name</span></code></a>()</p></td>
<td><p>Creates a string column for the file name of the current Spark task.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.instr.html#pyspark.sql.functions.instr" title="pyspark.sql.functions.instr"><code class="xref py py-obj docutils literal notranslate"><span class="pre">instr</span></code></a>(str, substr)</p></td>
<td><p>Locate the position of the first occurrence of substr column in the given string.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.isnan.html#pyspark.sql.functions.isnan" title="pyspark.sql.functions.isnan"><code class="xref py py-obj docutils literal notranslate"><span class="pre">isnan</span></code></a>(col)</p></td>
<td><p>An expression that returns true iff the column is NaN.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.isnull.html#pyspark.sql.functions.isnull" title="pyspark.sql.functions.isnull"><code class="xref py py-obj docutils literal notranslate"><span class="pre">isnull</span></code></a>(col)</p></td>
<td><p>An expression that returns true iff the column is null.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.json_tuple.html#pyspark.sql.functions.json_tuple" title="pyspark.sql.functions.json_tuple"><code class="xref py py-obj docutils literal notranslate"><span class="pre">json_tuple</span></code></a>(col, *fields)</p></td>
<td><p>Creates a new row for a json column according to the given field names.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.kurtosis.html#pyspark.sql.functions.kurtosis" title="pyspark.sql.functions.kurtosis"><code class="xref py py-obj docutils literal notranslate"><span class="pre">kurtosis</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the kurtosis of the values in a group.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.lag.html#pyspark.sql.functions.lag" title="pyspark.sql.functions.lag"><code class="xref py py-obj docutils literal notranslate"><span class="pre">lag</span></code></a>(col[, offset, default])</p></td>
<td><p>Window function: returns the value that is <cite>offset</cite> rows before the current row, and <cite>default</cite> if there is less than <cite>offset</cite> rows before the current row.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.last.html#pyspark.sql.functions.last" title="pyspark.sql.functions.last"><code class="xref py py-obj docutils literal notranslate"><span class="pre">last</span></code></a>(col[, ignorenulls])</p></td>
<td><p>Aggregate function: returns the last value in a group.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.last_day.html#pyspark.sql.functions.last_day" title="pyspark.sql.functions.last_day"><code class="xref py py-obj docutils literal notranslate"><span class="pre">last_day</span></code></a>(date)</p></td>
<td><p>Returns the last day of the month which the given date belongs to.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.lead.html#pyspark.sql.functions.lead" title="pyspark.sql.functions.lead"><code class="xref py py-obj docutils literal notranslate"><span class="pre">lead</span></code></a>(col[, offset, default])</p></td>
<td><p>Window function: returns the value that is <cite>offset</cite> rows after the current row, and <cite>default</cite> if there is less than <cite>offset</cite> rows after the current row.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.least.html#pyspark.sql.functions.least" title="pyspark.sql.functions.least"><code class="xref py py-obj docutils literal notranslate"><span class="pre">least</span></code></a>(*cols)</p></td>
<td><p>Returns the least value of the list of column names, skipping null values.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.length.html#pyspark.sql.functions.length" title="pyspark.sql.functions.length"><code class="xref py py-obj docutils literal notranslate"><span class="pre">length</span></code></a>(col)</p></td>
<td><p>Computes the character length of string data or number of bytes of binary data.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.levenshtein.html#pyspark.sql.functions.levenshtein" title="pyspark.sql.functions.levenshtein"><code class="xref py py-obj docutils literal notranslate"><span class="pre">levenshtein</span></code></a>(left, right)</p></td>
<td><p>Computes the Levenshtein distance of the two given strings.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.lit.html#pyspark.sql.functions.lit" title="pyspark.sql.functions.lit"><code class="xref py py-obj docutils literal notranslate"><span class="pre">lit</span></code></a>(col)</p></td>
<td><p>Creates a <a class="reference internal" href="api/pyspark.sql.Column.html#pyspark.sql.Column" title="pyspark.sql.Column"><code class="xref py py-class docutils literal notranslate"><span class="pre">Column</span></code></a> of literal value.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.locate.html#pyspark.sql.functions.locate" title="pyspark.sql.functions.locate"><code class="xref py py-obj docutils literal notranslate"><span class="pre">locate</span></code></a>(substr, str[, pos])</p></td>
<td><p>Locate the position of the first occurrence of substr in a string column, after position pos.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.log.html#pyspark.sql.functions.log" title="pyspark.sql.functions.log"><code class="xref py py-obj docutils literal notranslate"><span class="pre">log</span></code></a>(arg1[, arg2])</p></td>
<td><p>Returns the first argument-based logarithm of the second argument.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.log10.html#pyspark.sql.functions.log10" title="pyspark.sql.functions.log10"><code class="xref py py-obj docutils literal notranslate"><span class="pre">log10</span></code></a>(col)</p></td>
<td><p>Computes the logarithm of the given value in Base 10.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.log1p.html#pyspark.sql.functions.log1p" title="pyspark.sql.functions.log1p"><code class="xref py py-obj docutils literal notranslate"><span class="pre">log1p</span></code></a>(col)</p></td>
<td><p>Computes the natural logarithm of the given value plus one.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.log2.html#pyspark.sql.functions.log2" title="pyspark.sql.functions.log2"><code class="xref py py-obj docutils literal notranslate"><span class="pre">log2</span></code></a>(col)</p></td>
<td><p>Returns the base-2 logarithm of the argument.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.lower.html#pyspark.sql.functions.lower" title="pyspark.sql.functions.lower"><code class="xref py py-obj docutils literal notranslate"><span class="pre">lower</span></code></a>(col)</p></td>
<td><p>Converts a string expression to lower case.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.lpad.html#pyspark.sql.functions.lpad" title="pyspark.sql.functions.lpad"><code class="xref py py-obj docutils literal notranslate"><span class="pre">lpad</span></code></a>(col, len, pad)</p></td>
<td><p>Left-pad the string column to width <cite>len</cite> with <cite>pad</cite>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.ltrim.html#pyspark.sql.functions.ltrim" title="pyspark.sql.functions.ltrim"><code class="xref py py-obj docutils literal notranslate"><span class="pre">ltrim</span></code></a>(col)</p></td>
<td><p>Trim the spaces from left end for the specified string value.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.map_concat.html#pyspark.sql.functions.map_concat" title="pyspark.sql.functions.map_concat"><code class="xref py py-obj docutils literal notranslate"><span class="pre">map_concat</span></code></a>(*cols)</p></td>
<td><p>Returns the union of all the given maps.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.map_entries.html#pyspark.sql.functions.map_entries" title="pyspark.sql.functions.map_entries"><code class="xref py py-obj docutils literal notranslate"><span class="pre">map_entries</span></code></a>(col)</p></td>
<td><p>Collection function: Returns an unordered array of all entries in the given map.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.map_filter.html#pyspark.sql.functions.map_filter" title="pyspark.sql.functions.map_filter"><code class="xref py py-obj docutils literal notranslate"><span class="pre">map_filter</span></code></a>(col, f)</p></td>
<td><p>Returns a map whose key-value pairs satisfy a predicate.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.map_from_arrays.html#pyspark.sql.functions.map_from_arrays" title="pyspark.sql.functions.map_from_arrays"><code class="xref py py-obj docutils literal notranslate"><span class="pre">map_from_arrays</span></code></a>(col1, col2)</p></td>
<td><p>Creates a new map from two arrays.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.map_from_entries.html#pyspark.sql.functions.map_from_entries" title="pyspark.sql.functions.map_from_entries"><code class="xref py py-obj docutils literal notranslate"><span class="pre">map_from_entries</span></code></a>(col)</p></td>
<td><p>Collection function: Returns a map created from the given array of entries.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.map_keys.html#pyspark.sql.functions.map_keys" title="pyspark.sql.functions.map_keys"><code class="xref py py-obj docutils literal notranslate"><span class="pre">map_keys</span></code></a>(col)</p></td>
<td><p>Collection function: Returns an unordered array containing the keys of the map.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.map_values.html#pyspark.sql.functions.map_values" title="pyspark.sql.functions.map_values"><code class="xref py py-obj docutils literal notranslate"><span class="pre">map_values</span></code></a>(col)</p></td>
<td><p>Collection function: Returns an unordered array containing the values of the map.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.map_zip_with.html#pyspark.sql.functions.map_zip_with" title="pyspark.sql.functions.map_zip_with"><code class="xref py py-obj docutils literal notranslate"><span class="pre">map_zip_with</span></code></a>(col1, col2, f)</p></td>
<td><p>Merge two given maps, key-wise into a single map using a function.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.max.html#pyspark.sql.functions.max" title="pyspark.sql.functions.max"><code class="xref py py-obj docutils literal notranslate"><span class="pre">max</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the maximum value of the expression in a group.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.md5.html#pyspark.sql.functions.md5" title="pyspark.sql.functions.md5"><code class="xref py py-obj docutils literal notranslate"><span class="pre">md5</span></code></a>(col)</p></td>
<td><p>Calculates the MD5 digest and returns the value as a 32 character hex string.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.mean.html#pyspark.sql.functions.mean" title="pyspark.sql.functions.mean"><code class="xref py py-obj docutils literal notranslate"><span class="pre">mean</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the average of the values in a group.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.min.html#pyspark.sql.functions.min" title="pyspark.sql.functions.min"><code class="xref py py-obj docutils literal notranslate"><span class="pre">min</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the minimum value of the expression in a group.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.minute.html#pyspark.sql.functions.minute" title="pyspark.sql.functions.minute"><code class="xref py py-obj docutils literal notranslate"><span class="pre">minute</span></code></a>(col)</p></td>
<td><p>Extract the minutes of a given date as integer.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.monotonically_increasing_id.html#pyspark.sql.functions.monotonically_increasing_id" title="pyspark.sql.functions.monotonically_increasing_id"><code class="xref py py-obj docutils literal notranslate"><span class="pre">monotonically_increasing_id</span></code></a>()</p></td>
<td><p>A column that generates monotonically increasing 64-bit integers.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.month.html#pyspark.sql.functions.month" title="pyspark.sql.functions.month"><code class="xref py py-obj docutils literal notranslate"><span class="pre">month</span></code></a>(col)</p></td>
<td><p>Extract the month of a given date as integer.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.months.html#pyspark.sql.functions.months" title="pyspark.sql.functions.months"><code class="xref py py-obj docutils literal notranslate"><span class="pre">months</span></code></a>(col)</p></td>
<td><p>Partition transform function: A transform for timestamps and dates to partition data into months.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.months_between.html#pyspark.sql.functions.months_between" title="pyspark.sql.functions.months_between"><code class="xref py py-obj docutils literal notranslate"><span class="pre">months_between</span></code></a>(date1, date2[, roundOff])</p></td>
<td><p>Returns number of months between dates date1 and date2.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.nanvl.html#pyspark.sql.functions.nanvl" title="pyspark.sql.functions.nanvl"><code class="xref py py-obj docutils literal notranslate"><span class="pre">nanvl</span></code></a>(col1, col2)</p></td>
<td><p>Returns col1 if it is not NaN, or col2 if col1 is NaN.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.next_day.html#pyspark.sql.functions.next_day" title="pyspark.sql.functions.next_day"><code class="xref py py-obj docutils literal notranslate"><span class="pre">next_day</span></code></a>(date, dayOfWeek)</p></td>
<td><p>Returns the first date which is later than the value of the date column.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.nth_value.html#pyspark.sql.functions.nth_value" title="pyspark.sql.functions.nth_value"><code class="xref py py-obj docutils literal notranslate"><span class="pre">nth_value</span></code></a>(col, offset[, ignoreNulls])</p></td>
<td><p>Window function: returns the value that is the <cite>offset</cite>th row of the window frame (counting from 1), and <cite>null</cite> if the size of window frame is less than <cite>offset</cite> rows.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.ntile.html#pyspark.sql.functions.ntile" title="pyspark.sql.functions.ntile"><code class="xref py py-obj docutils literal notranslate"><span class="pre">ntile</span></code></a>(n)</p></td>
<td><p>Window function: returns the ntile group id (from 1 to <cite>n</cite> inclusive) in an ordered window partition.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.overlay.html#pyspark.sql.functions.overlay" title="pyspark.sql.functions.overlay"><code class="xref py py-obj docutils literal notranslate"><span class="pre">overlay</span></code></a>(src, replace, pos[, len])</p></td>
<td><p>Overlay the specified portion of <cite>src</cite> with <cite>replace</cite>, starting from byte position <cite>pos</cite> of <cite>src</cite> and proceeding for <cite>len</cite> bytes.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.pandas_udf.html#pyspark.sql.functions.pandas_udf" title="pyspark.sql.functions.pandas_udf"><code class="xref py py-obj docutils literal notranslate"><span class="pre">pandas_udf</span></code></a>([f, returnType, functionType])</p></td>
<td><p>Creates a pandas user defined function (a.k.a.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.percent_rank.html#pyspark.sql.functions.percent_rank" title="pyspark.sql.functions.percent_rank"><code class="xref py py-obj docutils literal notranslate"><span class="pre">percent_rank</span></code></a>()</p></td>
<td><p>Window function: returns the relative rank (i.e.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.percentile_approx.html#pyspark.sql.functions.percentile_approx" title="pyspark.sql.functions.percentile_approx"><code class="xref py py-obj docutils literal notranslate"><span class="pre">percentile_approx</span></code></a>(col, percentage[, accuracy])</p></td>
<td><p>Returns the approximate <cite>percentile</cite> of the numeric column <cite>col</cite> which is the smallest value in the ordered <cite>col</cite> values (sorted from least to greatest) such that no more than <cite>percentage</cite> of <cite>col</cite> values is less than the value or equal to that value.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.posexplode.html#pyspark.sql.functions.posexplode" title="pyspark.sql.functions.posexplode"><code class="xref py py-obj docutils literal notranslate"><span class="pre">posexplode</span></code></a>(col)</p></td>
<td><p>Returns a new row for each element with position in the given array or map.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.posexplode_outer.html#pyspark.sql.functions.posexplode_outer" title="pyspark.sql.functions.posexplode_outer"><code class="xref py py-obj docutils literal notranslate"><span class="pre">posexplode_outer</span></code></a>(col)</p></td>
<td><p>Returns a new row for each element with position in the given array or map.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.pow.html#pyspark.sql.functions.pow" title="pyspark.sql.functions.pow"><code class="xref py py-obj docutils literal notranslate"><span class="pre">pow</span></code></a>(col1, col2)</p></td>
<td><p>Returns the value of the first argument raised to the power of the second argument.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.product.html#pyspark.sql.functions.product" title="pyspark.sql.functions.product"><code class="xref py py-obj docutils literal notranslate"><span class="pre">product</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the product of the values in a group.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.quarter.html#pyspark.sql.functions.quarter" title="pyspark.sql.functions.quarter"><code class="xref py py-obj docutils literal notranslate"><span class="pre">quarter</span></code></a>(col)</p></td>
<td><p>Extract the quarter of a given date as integer.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.radians.html#pyspark.sql.functions.radians" title="pyspark.sql.functions.radians"><code class="xref py py-obj docutils literal notranslate"><span class="pre">radians</span></code></a>(col)</p></td>
<td><p>Converts an angle measured in degrees to an approximately equivalent angle measured in radians.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.raise_error.html#pyspark.sql.functions.raise_error" title="pyspark.sql.functions.raise_error"><code class="xref py py-obj docutils literal notranslate"><span class="pre">raise_error</span></code></a>(errMsg)</p></td>
<td><p>Throws an exception with the provided error message.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.rand.html#pyspark.sql.functions.rand" title="pyspark.sql.functions.rand"><code class="xref py py-obj docutils literal notranslate"><span class="pre">rand</span></code></a>([seed])</p></td>
<td><p>Generates a random column with independent and identically distributed (i.i.d.) samples uniformly distributed in [0.0, 1.0).</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.randn.html#pyspark.sql.functions.randn" title="pyspark.sql.functions.randn"><code class="xref py py-obj docutils literal notranslate"><span class="pre">randn</span></code></a>([seed])</p></td>
<td><p>Generates a column with independent and identically distributed (i.i.d.) samples from the standard normal distribution.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.rank.html#pyspark.sql.functions.rank" title="pyspark.sql.functions.rank"><code class="xref py py-obj docutils literal notranslate"><span class="pre">rank</span></code></a>()</p></td>
<td><p>Window function: returns the rank of rows within a window partition.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.regexp_extract.html#pyspark.sql.functions.regexp_extract" title="pyspark.sql.functions.regexp_extract"><code class="xref py py-obj docutils literal notranslate"><span class="pre">regexp_extract</span></code></a>(str, pattern, idx)</p></td>
<td><p>Extract a specific group matched by a Java regex, from the specified string column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.regexp_replace.html#pyspark.sql.functions.regexp_replace" title="pyspark.sql.functions.regexp_replace"><code class="xref py py-obj docutils literal notranslate"><span class="pre">regexp_replace</span></code></a>(str, pattern, replacement)</p></td>
<td><p>Replace all substrings of the specified string value that match regexp with rep.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.repeat.html#pyspark.sql.functions.repeat" title="pyspark.sql.functions.repeat"><code class="xref py py-obj docutils literal notranslate"><span class="pre">repeat</span></code></a>(col, n)</p></td>
<td><p>Repeats a string column n times, and returns it as a new string column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.reverse.html#pyspark.sql.functions.reverse" title="pyspark.sql.functions.reverse"><code class="xref py py-obj docutils literal notranslate"><span class="pre">reverse</span></code></a>(col)</p></td>
<td><p>Collection function: returns a reversed string or an array with reverse order of elements.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.rint.html#pyspark.sql.functions.rint" title="pyspark.sql.functions.rint"><code class="xref py py-obj docutils literal notranslate"><span class="pre">rint</span></code></a>(col)</p></td>
<td><p>Returns the double value that is closest in value to the argument and is equal to a mathematical integer.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.round.html#pyspark.sql.functions.round" title="pyspark.sql.functions.round"><code class="xref py py-obj docutils literal notranslate"><span class="pre">round</span></code></a>(col[, scale])</p></td>
<td><p>Round the given value to <cite>scale</cite> decimal places using HALF_UP rounding mode if <cite>scale</cite> &gt;= 0 or at integral part when <cite>scale</cite> &lt; 0.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.row_number.html#pyspark.sql.functions.row_number" title="pyspark.sql.functions.row_number"><code class="xref py py-obj docutils literal notranslate"><span class="pre">row_number</span></code></a>()</p></td>
<td><p>Window function: returns a sequential number starting at 1 within a window partition.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.rpad.html#pyspark.sql.functions.rpad" title="pyspark.sql.functions.rpad"><code class="xref py py-obj docutils literal notranslate"><span class="pre">rpad</span></code></a>(col, len, pad)</p></td>
<td><p>Right-pad the string column to width <cite>len</cite> with <cite>pad</cite>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.rtrim.html#pyspark.sql.functions.rtrim" title="pyspark.sql.functions.rtrim"><code class="xref py py-obj docutils literal notranslate"><span class="pre">rtrim</span></code></a>(col)</p></td>
<td><p>Trim the spaces from right end for the specified string value.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.schema_of_csv.html#pyspark.sql.functions.schema_of_csv" title="pyspark.sql.functions.schema_of_csv"><code class="xref py py-obj docutils literal notranslate"><span class="pre">schema_of_csv</span></code></a>(csv[, options])</p></td>
<td><p>Parses a CSV string and infers its schema in DDL format.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.schema_of_json.html#pyspark.sql.functions.schema_of_json" title="pyspark.sql.functions.schema_of_json"><code class="xref py py-obj docutils literal notranslate"><span class="pre">schema_of_json</span></code></a>(json[, options])</p></td>
<td><p>Parses a JSON string and infers its schema in DDL format.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.second.html#pyspark.sql.functions.second" title="pyspark.sql.functions.second"><code class="xref py py-obj docutils literal notranslate"><span class="pre">second</span></code></a>(col)</p></td>
<td><p>Extract the seconds of a given date as integer.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.sentences.html#pyspark.sql.functions.sentences" title="pyspark.sql.functions.sentences"><code class="xref py py-obj docutils literal notranslate"><span class="pre">sentences</span></code></a>(string[, language, country])</p></td>
<td><p>Splits a string into arrays of sentences, where each sentence is an array of words.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.sequence.html#pyspark.sql.functions.sequence" title="pyspark.sql.functions.sequence"><code class="xref py py-obj docutils literal notranslate"><span class="pre">sequence</span></code></a>(start, stop[, step])</p></td>
<td><p>Generate a sequence of integers from <cite>start</cite> to <cite>stop</cite>, incrementing by <cite>step</cite>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.session_window.html#pyspark.sql.functions.session_window" title="pyspark.sql.functions.session_window"><code class="xref py py-obj docutils literal notranslate"><span class="pre">session_window</span></code></a>(timeColumn, gapDuration)</p></td>
<td><p>Generates session window given a timestamp specifying column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.sha1.html#pyspark.sql.functions.sha1" title="pyspark.sql.functions.sha1"><code class="xref py py-obj docutils literal notranslate"><span class="pre">sha1</span></code></a>(col)</p></td>
<td><p>Returns the hex string result of SHA-1.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.sha2.html#pyspark.sql.functions.sha2" title="pyspark.sql.functions.sha2"><code class="xref py py-obj docutils literal notranslate"><span class="pre">sha2</span></code></a>(col, numBits)</p></td>
<td><p>Returns the hex string result of SHA-2 family of hash functions (SHA-224, SHA-256, SHA-384, and SHA-512).</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.shiftleft.html#pyspark.sql.functions.shiftleft" title="pyspark.sql.functions.shiftleft"><code class="xref py py-obj docutils literal notranslate"><span class="pre">shiftleft</span></code></a>(col, numBits)</p></td>
<td><p>Shift the given value numBits left.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.shiftright.html#pyspark.sql.functions.shiftright" title="pyspark.sql.functions.shiftright"><code class="xref py py-obj docutils literal notranslate"><span class="pre">shiftright</span></code></a>(col, numBits)</p></td>
<td><p>(Signed) shift the given value numBits right.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.shiftrightunsigned.html#pyspark.sql.functions.shiftrightunsigned" title="pyspark.sql.functions.shiftrightunsigned"><code class="xref py py-obj docutils literal notranslate"><span class="pre">shiftrightunsigned</span></code></a>(col, numBits)</p></td>
<td><p>Unsigned shift the given value numBits right.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.shuffle.html#pyspark.sql.functions.shuffle" title="pyspark.sql.functions.shuffle"><code class="xref py py-obj docutils literal notranslate"><span class="pre">shuffle</span></code></a>(col)</p></td>
<td><p>Collection function: Generates a random permutation of the given array.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.signum.html#pyspark.sql.functions.signum" title="pyspark.sql.functions.signum"><code class="xref py py-obj docutils literal notranslate"><span class="pre">signum</span></code></a>(col)</p></td>
<td><p>Computes the signum of the given value.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.sin.html#pyspark.sql.functions.sin" title="pyspark.sql.functions.sin"><code class="xref py py-obj docutils literal notranslate"><span class="pre">sin</span></code></a>(col)</p></td>
<td><p><div class="versionadded">
<p><span class="versionmodified added">New in version 1.4.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.sinh.html#pyspark.sql.functions.sinh" title="pyspark.sql.functions.sinh"><code class="xref py py-obj docutils literal notranslate"><span class="pre">sinh</span></code></a>(col)</p></td>
<td><p><div class="versionadded">
<p><span class="versionmodified added">New in version 1.4.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.size.html#pyspark.sql.functions.size" title="pyspark.sql.functions.size"><code class="xref py py-obj docutils literal notranslate"><span class="pre">size</span></code></a>(col)</p></td>
<td><p>Collection function: returns the length of the array or map stored in the column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.skewness.html#pyspark.sql.functions.skewness" title="pyspark.sql.functions.skewness"><code class="xref py py-obj docutils literal notranslate"><span class="pre">skewness</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the skewness of the values in a group.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.slice.html#pyspark.sql.functions.slice" title="pyspark.sql.functions.slice"><code class="xref py py-obj docutils literal notranslate"><span class="pre">slice</span></code></a>(x, start, length)</p></td>
<td><p>Collection function: returns an array containing all the elements in <cite>x</cite> from index <cite>start</cite> (array indices start at 1, or from the end if <cite>start</cite> is negative) with the specified <cite>length</cite>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.sort_array.html#pyspark.sql.functions.sort_array" title="pyspark.sql.functions.sort_array"><code class="xref py py-obj docutils literal notranslate"><span class="pre">sort_array</span></code></a>(col[, asc])</p></td>
<td><p>Collection function: sorts the input array in ascending or descending order according to the natural ordering of the array elements.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.soundex.html#pyspark.sql.functions.soundex" title="pyspark.sql.functions.soundex"><code class="xref py py-obj docutils literal notranslate"><span class="pre">soundex</span></code></a>(col)</p></td>
<td><p>Returns the SoundEx encoding for a string</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.spark_partition_id.html#pyspark.sql.functions.spark_partition_id" title="pyspark.sql.functions.spark_partition_id"><code class="xref py py-obj docutils literal notranslate"><span class="pre">spark_partition_id</span></code></a>()</p></td>
<td><p>A column for partition ID.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.split.html#pyspark.sql.functions.split" title="pyspark.sql.functions.split"><code class="xref py py-obj docutils literal notranslate"><span class="pre">split</span></code></a>(str, pattern[, limit])</p></td>
<td><p>Splits str around matches of the given pattern.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.sqrt.html#pyspark.sql.functions.sqrt" title="pyspark.sql.functions.sqrt"><code class="xref py py-obj docutils literal notranslate"><span class="pre">sqrt</span></code></a>(col)</p></td>
<td><p>Computes the square root of the specified float value.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.stddev.html#pyspark.sql.functions.stddev" title="pyspark.sql.functions.stddev"><code class="xref py py-obj docutils literal notranslate"><span class="pre">stddev</span></code></a>(col)</p></td>
<td><p>Aggregate function: alias for stddev_samp.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.stddev_pop.html#pyspark.sql.functions.stddev_pop" title="pyspark.sql.functions.stddev_pop"><code class="xref py py-obj docutils literal notranslate"><span class="pre">stddev_pop</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns population standard deviation of the expression in a group.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.stddev_samp.html#pyspark.sql.functions.stddev_samp" title="pyspark.sql.functions.stddev_samp"><code class="xref py py-obj docutils literal notranslate"><span class="pre">stddev_samp</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the unbiased sample standard deviation of the expression in a group.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.struct.html#pyspark.sql.functions.struct" title="pyspark.sql.functions.struct"><code class="xref py py-obj docutils literal notranslate"><span class="pre">struct</span></code></a>(*cols)</p></td>
<td><p>Creates a new struct column.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.substring.html#pyspark.sql.functions.substring" title="pyspark.sql.functions.substring"><code class="xref py py-obj docutils literal notranslate"><span class="pre">substring</span></code></a>(str, pos, len)</p></td>
<td><p>Substring starts at <cite>pos</cite> and is of length <cite>len</cite> when str is String type or returns the slice of byte array that starts at <cite>pos</cite> in byte and is of length <cite>len</cite> when str is Binary type.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.substring_index.html#pyspark.sql.functions.substring_index" title="pyspark.sql.functions.substring_index"><code class="xref py py-obj docutils literal notranslate"><span class="pre">substring_index</span></code></a>(str, delim, count)</p></td>
<td><p>Returns the substring from string str before count occurrences of the delimiter delim.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.sum.html#pyspark.sql.functions.sum" title="pyspark.sql.functions.sum"><code class="xref py py-obj docutils literal notranslate"><span class="pre">sum</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the sum of all values in the expression.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.sum_distinct.html#pyspark.sql.functions.sum_distinct" title="pyspark.sql.functions.sum_distinct"><code class="xref py py-obj docutils literal notranslate"><span class="pre">sum_distinct</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the sum of distinct values in the expression.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.sumDistinct.html#pyspark.sql.functions.sumDistinct" title="pyspark.sql.functions.sumDistinct"><code class="xref py py-obj docutils literal notranslate"><span class="pre">sumDistinct</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the sum of distinct values in the expression.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.tan.html#pyspark.sql.functions.tan" title="pyspark.sql.functions.tan"><code class="xref py py-obj docutils literal notranslate"><span class="pre">tan</span></code></a>(col)</p></td>
<td><p><div class="versionadded">
<p><span class="versionmodified added">New in version 1.4.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.tanh.html#pyspark.sql.functions.tanh" title="pyspark.sql.functions.tanh"><code class="xref py py-obj docutils literal notranslate"><span class="pre">tanh</span></code></a>(col)</p></td>
<td><p><div class="versionadded">
<p><span class="versionmodified added">New in version 1.4.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.timestamp_seconds.html#pyspark.sql.functions.timestamp_seconds" title="pyspark.sql.functions.timestamp_seconds"><code class="xref py py-obj docutils literal notranslate"><span class="pre">timestamp_seconds</span></code></a>(col)</p></td>
<td><p><div class="versionadded">
<p><span class="versionmodified added">New in version 3.1.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.toDegrees.html#pyspark.sql.functions.toDegrees" title="pyspark.sql.functions.toDegrees"><code class="xref py py-obj docutils literal notranslate"><span class="pre">toDegrees</span></code></a>(col)</p></td>
<td><p><div class="deprecated">
<p><span class="versionmodified deprecated">Deprecated since version 2.1.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.toRadians.html#pyspark.sql.functions.toRadians" title="pyspark.sql.functions.toRadians"><code class="xref py py-obj docutils literal notranslate"><span class="pre">toRadians</span></code></a>(col)</p></td>
<td><p><div class="deprecated">
<p><span class="versionmodified deprecated">Deprecated since version 2.1.0.</span></p>
</div>
</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.to_csv.html#pyspark.sql.functions.to_csv" title="pyspark.sql.functions.to_csv"><code class="xref py py-obj docutils literal notranslate"><span class="pre">to_csv</span></code></a>(col[, options])</p></td>
<td><p>Converts a column containing a <code class="xref py py-class docutils literal notranslate"><span class="pre">StructType</span></code> into a CSV string.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.to_date.html#pyspark.sql.functions.to_date" title="pyspark.sql.functions.to_date"><code class="xref py py-obj docutils literal notranslate"><span class="pre">to_date</span></code></a>(col[, format])</p></td>
<td><p>Converts a <a class="reference internal" href="api/pyspark.sql.Column.html#pyspark.sql.Column" title="pyspark.sql.Column"><code class="xref py py-class docutils literal notranslate"><span class="pre">Column</span></code></a> into <a class="reference internal" href="api/pyspark.sql.types.DateType.html#pyspark.sql.types.DateType" title="pyspark.sql.types.DateType"><code class="xref py py-class docutils literal notranslate"><span class="pre">pyspark.sql.types.DateType</span></code></a> using the optionally specified format.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.to_json.html#pyspark.sql.functions.to_json" title="pyspark.sql.functions.to_json"><code class="xref py py-obj docutils literal notranslate"><span class="pre">to_json</span></code></a>(col[, options])</p></td>
<td><p>Converts a column containing a <code class="xref py py-class docutils literal notranslate"><span class="pre">StructType</span></code>, <code class="xref py py-class docutils literal notranslate"><span class="pre">ArrayType</span></code> or a <code class="xref py py-class docutils literal notranslate"><span class="pre">MapType</span></code> into a JSON string.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.to_timestamp.html#pyspark.sql.functions.to_timestamp" title="pyspark.sql.functions.to_timestamp"><code class="xref py py-obj docutils literal notranslate"><span class="pre">to_timestamp</span></code></a>(col[, format])</p></td>
<td><p>Converts a <a class="reference internal" href="api/pyspark.sql.Column.html#pyspark.sql.Column" title="pyspark.sql.Column"><code class="xref py py-class docutils literal notranslate"><span class="pre">Column</span></code></a> into <a class="reference internal" href="api/pyspark.sql.types.TimestampType.html#pyspark.sql.types.TimestampType" title="pyspark.sql.types.TimestampType"><code class="xref py py-class docutils literal notranslate"><span class="pre">pyspark.sql.types.TimestampType</span></code></a> using the optionally specified format.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.to_utc_timestamp.html#pyspark.sql.functions.to_utc_timestamp" title="pyspark.sql.functions.to_utc_timestamp"><code class="xref py py-obj docutils literal notranslate"><span class="pre">to_utc_timestamp</span></code></a>(timestamp, tz)</p></td>
<td><p>This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.transform.html#pyspark.sql.functions.transform" title="pyspark.sql.functions.transform"><code class="xref py py-obj docutils literal notranslate"><span class="pre">transform</span></code></a>(col, f)</p></td>
<td><p>Returns an array of elements after applying a transformation to each element in the input array.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.transform_keys.html#pyspark.sql.functions.transform_keys" title="pyspark.sql.functions.transform_keys"><code class="xref py py-obj docutils literal notranslate"><span class="pre">transform_keys</span></code></a>(col, f)</p></td>
<td><p>Applies a function to every key-value pair in a map and returns a map with the results of those applications as the new keys for the pairs.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.transform_values.html#pyspark.sql.functions.transform_values" title="pyspark.sql.functions.transform_values"><code class="xref py py-obj docutils literal notranslate"><span class="pre">transform_values</span></code></a>(col, f)</p></td>
<td><p>Applies a function to every key-value pair in a map and returns a map with the results of those applications as the new values for the pairs.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.translate.html#pyspark.sql.functions.translate" title="pyspark.sql.functions.translate"><code class="xref py py-obj docutils literal notranslate"><span class="pre">translate</span></code></a>(srcCol, matching, replace)</p></td>
<td><p>A function translate any character in the <cite>srcCol</cite> by a character in <cite>matching</cite>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.trim.html#pyspark.sql.functions.trim" title="pyspark.sql.functions.trim"><code class="xref py py-obj docutils literal notranslate"><span class="pre">trim</span></code></a>(col)</p></td>
<td><p>Trim the spaces from both ends for the specified string column.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.trunc.html#pyspark.sql.functions.trunc" title="pyspark.sql.functions.trunc"><code class="xref py py-obj docutils literal notranslate"><span class="pre">trunc</span></code></a>(date, format)</p></td>
<td><p>Returns date truncated to the unit specified by the format.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.udf.html#pyspark.sql.functions.udf" title="pyspark.sql.functions.udf"><code class="xref py py-obj docutils literal notranslate"><span class="pre">udf</span></code></a>([f, returnType])</p></td>
<td><p>Creates a user defined function (UDF).</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.unbase64.html#pyspark.sql.functions.unbase64" title="pyspark.sql.functions.unbase64"><code class="xref py py-obj docutils literal notranslate"><span class="pre">unbase64</span></code></a>(col)</p></td>
<td><p>Decodes a BASE64 encoded string column and returns it as a binary column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.unhex.html#pyspark.sql.functions.unhex" title="pyspark.sql.functions.unhex"><code class="xref py py-obj docutils literal notranslate"><span class="pre">unhex</span></code></a>(col)</p></td>
<td><p>Inverse of hex.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.unix_timestamp.html#pyspark.sql.functions.unix_timestamp" title="pyspark.sql.functions.unix_timestamp"><code class="xref py py-obj docutils literal notranslate"><span class="pre">unix_timestamp</span></code></a>([timestamp, format])</p></td>
<td><p>Convert time string with given pattern (‘yyyy-MM-dd HH:mm:ss’, by default) to Unix time stamp (in seconds), using the default timezone and the default locale, return null if fail.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.upper.html#pyspark.sql.functions.upper" title="pyspark.sql.functions.upper"><code class="xref py py-obj docutils literal notranslate"><span class="pre">upper</span></code></a>(col)</p></td>
<td><p>Converts a string expression to upper case.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.var_pop.html#pyspark.sql.functions.var_pop" title="pyspark.sql.functions.var_pop"><code class="xref py py-obj docutils literal notranslate"><span class="pre">var_pop</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the population variance of the values in a group.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.var_samp.html#pyspark.sql.functions.var_samp" title="pyspark.sql.functions.var_samp"><code class="xref py py-obj docutils literal notranslate"><span class="pre">var_samp</span></code></a>(col)</p></td>
<td><p>Aggregate function: returns the unbiased sample variance of the values in a group.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.variance.html#pyspark.sql.functions.variance" title="pyspark.sql.functions.variance"><code class="xref py py-obj docutils literal notranslate"><span class="pre">variance</span></code></a>(col)</p></td>
<td><p>Aggregate function: alias for var_samp</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.weekofyear.html#pyspark.sql.functions.weekofyear" title="pyspark.sql.functions.weekofyear"><code class="xref py py-obj docutils literal notranslate"><span class="pre">weekofyear</span></code></a>(col)</p></td>
<td><p>Extract the week number of a given date as integer.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.when.html#pyspark.sql.functions.when" title="pyspark.sql.functions.when"><code class="xref py py-obj docutils literal notranslate"><span class="pre">when</span></code></a>(condition, value)</p></td>
<td><p>Evaluates a list of conditions and returns one of multiple possible result expressions.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.window.html#pyspark.sql.functions.window" title="pyspark.sql.functions.window"><code class="xref py py-obj docutils literal notranslate"><span class="pre">window</span></code></a>(timeColumn, windowDuration[, …])</p></td>
<td><p>Bucketize rows into one or more time windows given a timestamp specifying column.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.xxhash64.html#pyspark.sql.functions.xxhash64" title="pyspark.sql.functions.xxhash64"><code class="xref py py-obj docutils literal notranslate"><span class="pre">xxhash64</span></code></a>(*cols)</p></td>
<td><p>Calculates the hash code of given columns using the 64-bit variant of the xxHash algorithm, and returns the result as a long column.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.year.html#pyspark.sql.functions.year" title="pyspark.sql.functions.year"><code class="xref py py-obj docutils literal notranslate"><span class="pre">year</span></code></a>(col)</p></td>
<td><p>Extract the year of a given date as integer.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.functions.years.html#pyspark.sql.functions.years" title="pyspark.sql.functions.years"><code class="xref py py-obj docutils literal notranslate"><span class="pre">years</span></code></a>(col)</p></td>
<td><p>Partition transform function: A transform for timestamps and dates to partition data into years.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.functions.zip_with.html#pyspark.sql.functions.zip_with" title="pyspark.sql.functions.zip_with"><code class="xref py py-obj docutils literal notranslate"><span class="pre">zip_with</span></code></a>(left, right, f)</p></td>
<td><p>Merge two given arrays, element-wise, into a single array using a function.</p></td>
</tr>
</tbody>
</table>
<table class="longtable table autosummary">
<colgroup>
<col style="width: 10%" />
<col style="width: 90%" />
</colgroup>
<tbody>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.avro.functions.from_avro.html#pyspark.sql.avro.functions.from_avro" title="pyspark.sql.avro.functions.from_avro"><code class="xref py py-obj docutils literal notranslate"><span class="pre">from_avro</span></code></a>(data, jsonFormatSchema[, options])</p></td>
<td><p>Converts a binary column of Avro format into its corresponding catalyst value.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.avro.functions.to_avro.html#pyspark.sql.avro.functions.to_avro" title="pyspark.sql.avro.functions.to_avro"><code class="xref py py-obj docutils literal notranslate"><span class="pre">to_avro</span></code></a>(data[, jsonFormatSchema])</p></td>
<td><p>Converts a column into binary of avro format.</p></td>
</tr>
</tbody>
</table>
</section>
<section id="window">
<h2>Window<a class="headerlink" href="#window" title="Permalink to this headline"></a></h2>
<table class="longtable table autosummary">
<colgroup>
<col style="width: 10%" />
<col style="width: 90%" />
</colgroup>
<tbody>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Window.currentRow.html#pyspark.sql.Window.currentRow" title="pyspark.sql.Window.currentRow"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Window.currentRow</span></code></a></p></td>
<td><p></p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Window.orderBy.html#pyspark.sql.Window.orderBy" title="pyspark.sql.Window.orderBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Window.orderBy</span></code></a>(*cols)</p></td>
<td><p>Creates a <code class="xref py py-class docutils literal notranslate"><span class="pre">WindowSpec</span></code> with the ordering defined.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Window.partitionBy.html#pyspark.sql.Window.partitionBy" title="pyspark.sql.Window.partitionBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Window.partitionBy</span></code></a>(*cols)</p></td>
<td><p>Creates a <code class="xref py py-class docutils literal notranslate"><span class="pre">WindowSpec</span></code> with the partitioning defined.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Window.rangeBetween.html#pyspark.sql.Window.rangeBetween" title="pyspark.sql.Window.rangeBetween"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Window.rangeBetween</span></code></a>(start, end)</p></td>
<td><p>Creates a <code class="xref py py-class docutils literal notranslate"><span class="pre">WindowSpec</span></code> with the frame boundaries defined, from <cite>start</cite> (inclusive) to <cite>end</cite> (inclusive).</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Window.rowsBetween.html#pyspark.sql.Window.rowsBetween" title="pyspark.sql.Window.rowsBetween"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Window.rowsBetween</span></code></a>(start, end)</p></td>
<td><p>Creates a <code class="xref py py-class docutils literal notranslate"><span class="pre">WindowSpec</span></code> with the frame boundaries defined, from <cite>start</cite> (inclusive) to <cite>end</cite> (inclusive).</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Window.unboundedFollowing.html#pyspark.sql.Window.unboundedFollowing" title="pyspark.sql.Window.unboundedFollowing"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Window.unboundedFollowing</span></code></a></p></td>
<td><p></p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Window.unboundedPreceding.html#pyspark.sql.Window.unboundedPreceding" title="pyspark.sql.Window.unboundedPreceding"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Window.unboundedPreceding</span></code></a></p></td>
<td><p></p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.WindowSpec.orderBy.html#pyspark.sql.WindowSpec.orderBy" title="pyspark.sql.WindowSpec.orderBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">WindowSpec.orderBy</span></code></a>(*cols)</p></td>
<td><p>Defines the ordering columns in a <code class="xref py py-class docutils literal notranslate"><span class="pre">WindowSpec</span></code>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.WindowSpec.partitionBy.html#pyspark.sql.WindowSpec.partitionBy" title="pyspark.sql.WindowSpec.partitionBy"><code class="xref py py-obj docutils literal notranslate"><span class="pre">WindowSpec.partitionBy</span></code></a>(*cols)</p></td>
<td><p>Defines the partitioning columns in a <code class="xref py py-class docutils literal notranslate"><span class="pre">WindowSpec</span></code>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.WindowSpec.rangeBetween.html#pyspark.sql.WindowSpec.rangeBetween" title="pyspark.sql.WindowSpec.rangeBetween"><code class="xref py py-obj docutils literal notranslate"><span class="pre">WindowSpec.rangeBetween</span></code></a>(start, end)</p></td>
<td><p>Defines the frame boundaries, from <cite>start</cite> (inclusive) to <cite>end</cite> (inclusive).</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.WindowSpec.rowsBetween.html#pyspark.sql.WindowSpec.rowsBetween" title="pyspark.sql.WindowSpec.rowsBetween"><code class="xref py py-obj docutils literal notranslate"><span class="pre">WindowSpec.rowsBetween</span></code></a>(start, end)</p></td>
<td><p>Defines the frame boundaries, from <cite>start</cite> (inclusive) to <cite>end</cite> (inclusive).</p></td>
</tr>
</tbody>
</table>
</section>
<section id="grouping">
<h2>Grouping<a class="headerlink" href="#grouping" title="Permalink to this headline"></a></h2>
<table class="longtable table autosummary">
<colgroup>
<col style="width: 10%" />
<col style="width: 90%" />
</colgroup>
<tbody>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.GroupedData.agg.html#pyspark.sql.GroupedData.agg" title="pyspark.sql.GroupedData.agg"><code class="xref py py-obj docutils literal notranslate"><span class="pre">GroupedData.agg</span></code></a>(*exprs)</p></td>
<td><p>Compute aggregates and returns the result as a <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.GroupedData.apply.html#pyspark.sql.GroupedData.apply" title="pyspark.sql.GroupedData.apply"><code class="xref py py-obj docutils literal notranslate"><span class="pre">GroupedData.apply</span></code></a>(udf)</p></td>
<td><p>It is an alias of <a class="reference internal" href="api/pyspark.sql.GroupedData.applyInPandas.html#pyspark.sql.GroupedData.applyInPandas" title="pyspark.sql.GroupedData.applyInPandas"><code class="xref py py-meth docutils literal notranslate"><span class="pre">pyspark.sql.GroupedData.applyInPandas()</span></code></a>; however, it takes a <a class="reference internal" href="api/pyspark.sql.functions.pandas_udf.html#pyspark.sql.functions.pandas_udf" title="pyspark.sql.functions.pandas_udf"><code class="xref py py-meth docutils literal notranslate"><span class="pre">pyspark.sql.functions.pandas_udf()</span></code></a> whereas <a class="reference internal" href="api/pyspark.sql.GroupedData.applyInPandas.html#pyspark.sql.GroupedData.applyInPandas" title="pyspark.sql.GroupedData.applyInPandas"><code class="xref py py-meth docutils literal notranslate"><span class="pre">pyspark.sql.GroupedData.applyInPandas()</span></code></a> takes a Python native function.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.GroupedData.applyInPandas.html#pyspark.sql.GroupedData.applyInPandas" title="pyspark.sql.GroupedData.applyInPandas"><code class="xref py py-obj docutils literal notranslate"><span class="pre">GroupedData.applyInPandas</span></code></a>(func, schema)</p></td>
<td><p>Maps each group of the current <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> using a pandas udf and returns the result as a <cite>DataFrame</cite>.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.GroupedData.avg.html#pyspark.sql.GroupedData.avg" title="pyspark.sql.GroupedData.avg"><code class="xref py py-obj docutils literal notranslate"><span class="pre">GroupedData.avg</span></code></a>(*cols)</p></td>
<td><p>Computes average values for each numeric columns for each group.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.GroupedData.cogroup.html#pyspark.sql.GroupedData.cogroup" title="pyspark.sql.GroupedData.cogroup"><code class="xref py py-obj docutils literal notranslate"><span class="pre">GroupedData.cogroup</span></code></a>(other)</p></td>
<td><p>Cogroups this group with another group so that we can run cogrouped operations.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.GroupedData.count.html#pyspark.sql.GroupedData.count" title="pyspark.sql.GroupedData.count"><code class="xref py py-obj docutils literal notranslate"><span class="pre">GroupedData.count</span></code></a>()</p></td>
<td><p>Counts the number of records for each group.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.GroupedData.max.html#pyspark.sql.GroupedData.max" title="pyspark.sql.GroupedData.max"><code class="xref py py-obj docutils literal notranslate"><span class="pre">GroupedData.max</span></code></a>(*cols)</p></td>
<td><p>Computes the max value for each numeric columns for each group.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.GroupedData.mean.html#pyspark.sql.GroupedData.mean" title="pyspark.sql.GroupedData.mean"><code class="xref py py-obj docutils literal notranslate"><span class="pre">GroupedData.mean</span></code></a>(*cols)</p></td>
<td><p>Computes average values for each numeric columns for each group.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.GroupedData.min.html#pyspark.sql.GroupedData.min" title="pyspark.sql.GroupedData.min"><code class="xref py py-obj docutils literal notranslate"><span class="pre">GroupedData.min</span></code></a>(*cols)</p></td>
<td><p>Computes the min value for each numeric column for each group.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.GroupedData.pivot.html#pyspark.sql.GroupedData.pivot" title="pyspark.sql.GroupedData.pivot"><code class="xref py py-obj docutils literal notranslate"><span class="pre">GroupedData.pivot</span></code></a>(pivot_col[, values])</p></td>
<td><p>Pivots a column of the current <a class="reference internal" href="api/pyspark.sql.DataFrame.html#pyspark.sql.DataFrame" title="pyspark.sql.DataFrame"><code class="xref py py-class docutils literal notranslate"><span class="pre">DataFrame</span></code></a> and perform the specified aggregation.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.GroupedData.sum.html#pyspark.sql.GroupedData.sum" title="pyspark.sql.GroupedData.sum"><code class="xref py py-obj docutils literal notranslate"><span class="pre">GroupedData.sum</span></code></a>(*cols)</p></td>
<td><p>Computes the sum for each numeric columns for each group.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.PandasCogroupedOps.applyInPandas.html#pyspark.sql.PandasCogroupedOps.applyInPandas" title="pyspark.sql.PandasCogroupedOps.applyInPandas"><code class="xref py py-obj docutils literal notranslate"><span class="pre">PandasCogroupedOps.applyInPandas</span></code></a>(func, schema)</p></td>
<td><p>Applies a function to each cogroup using pandas and returns the result as a <cite>DataFrame</cite>.</p></td>
</tr>
</tbody>
</table>
</section>
<section id="catalog-apis">
<h2>Catalog APIs<a class="headerlink" href="#catalog-apis" title="Permalink to this headline"></a></h2>
<table class="longtable table autosummary">
<colgroup>
<col style="width: 10%" />
<col style="width: 90%" />
</colgroup>
<tbody>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.cacheTable.html#pyspark.sql.Catalog.cacheTable" title="pyspark.sql.Catalog.cacheTable"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.cacheTable</span></code></a>(tableName)</p></td>
<td><p>Caches the specified table in-memory.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.clearCache.html#pyspark.sql.Catalog.clearCache" title="pyspark.sql.Catalog.clearCache"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.clearCache</span></code></a>()</p></td>
<td><p>Removes all cached tables from the in-memory cache.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.createExternalTable.html#pyspark.sql.Catalog.createExternalTable" title="pyspark.sql.Catalog.createExternalTable"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.createExternalTable</span></code></a>(tableName[, …])</p></td>
<td><p>Creates a table based on the dataset in a data source.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.createTable.html#pyspark.sql.Catalog.createTable" title="pyspark.sql.Catalog.createTable"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.createTable</span></code></a>(tableName[, path, …])</p></td>
<td><p>Creates a table based on the dataset in a data source.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.currentDatabase.html#pyspark.sql.Catalog.currentDatabase" title="pyspark.sql.Catalog.currentDatabase"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.currentDatabase</span></code></a>()</p></td>
<td><p>Returns the current default database in this session.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.dropGlobalTempView.html#pyspark.sql.Catalog.dropGlobalTempView" title="pyspark.sql.Catalog.dropGlobalTempView"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.dropGlobalTempView</span></code></a>(viewName)</p></td>
<td><p>Drops the global temporary view with the given view name in the catalog.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.dropTempView.html#pyspark.sql.Catalog.dropTempView" title="pyspark.sql.Catalog.dropTempView"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.dropTempView</span></code></a>(viewName)</p></td>
<td><p>Drops the local temporary view with the given view name in the catalog.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.isCached.html#pyspark.sql.Catalog.isCached" title="pyspark.sql.Catalog.isCached"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.isCached</span></code></a>(tableName)</p></td>
<td><p>Returns true if the table is currently cached in-memory.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.listColumns.html#pyspark.sql.Catalog.listColumns" title="pyspark.sql.Catalog.listColumns"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.listColumns</span></code></a>(tableName[, dbName])</p></td>
<td><p>Returns a list of columns for the given table/view in the specified database.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.listDatabases.html#pyspark.sql.Catalog.listDatabases" title="pyspark.sql.Catalog.listDatabases"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.listDatabases</span></code></a>()</p></td>
<td><p>Returns a list of databases available across all sessions.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.listFunctions.html#pyspark.sql.Catalog.listFunctions" title="pyspark.sql.Catalog.listFunctions"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.listFunctions</span></code></a>([dbName])</p></td>
<td><p>Returns a list of functions registered in the specified database.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.listTables.html#pyspark.sql.Catalog.listTables" title="pyspark.sql.Catalog.listTables"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.listTables</span></code></a>([dbName])</p></td>
<td><p>Returns a list of tables/views in the specified database.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.recoverPartitions.html#pyspark.sql.Catalog.recoverPartitions" title="pyspark.sql.Catalog.recoverPartitions"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.recoverPartitions</span></code></a>(tableName)</p></td>
<td><p>Recovers all the partitions of the given table and update the catalog.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.refreshByPath.html#pyspark.sql.Catalog.refreshByPath" title="pyspark.sql.Catalog.refreshByPath"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.refreshByPath</span></code></a>(path)</p></td>
<td><p>Invalidates and refreshes all the cached data (and the associated metadata) for any DataFrame that contains the given data source path.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.refreshTable.html#pyspark.sql.Catalog.refreshTable" title="pyspark.sql.Catalog.refreshTable"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.refreshTable</span></code></a>(tableName)</p></td>
<td><p>Invalidates and refreshes all the cached data and metadata of the given table.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.registerFunction.html#pyspark.sql.Catalog.registerFunction" title="pyspark.sql.Catalog.registerFunction"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.registerFunction</span></code></a>(name, f[, returnType])</p></td>
<td><p>An alias for <code class="xref py py-func docutils literal notranslate"><span class="pre">spark.udf.register()</span></code>.</p></td>
</tr>
<tr class="row-odd"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.setCurrentDatabase.html#pyspark.sql.Catalog.setCurrentDatabase" title="pyspark.sql.Catalog.setCurrentDatabase"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.setCurrentDatabase</span></code></a>(dbName)</p></td>
<td><p>Sets the current default database in this session.</p></td>
</tr>
<tr class="row-even"><td><p><a class="reference internal" href="api/pyspark.sql.Catalog.uncacheTable.html#pyspark.sql.Catalog.uncacheTable" title="pyspark.sql.Catalog.uncacheTable"><code class="xref py py-obj docutils literal notranslate"><span class="pre">Catalog.uncacheTable</span></code></a>(tableName)</p></td>
<td><p>Removes the specified table from the in-memory cache.</p></td>
</tr>
</tbody>
</table>
</section>
</section>
</div>
<!-- Previous / next buttons -->
<div class='prev-next-area'>
<a class='left-prev' id="prev-link" href="index.html" title="previous page">
<i class="fas fa-angle-left"></i>
<div class="prev-next-info">
<p class="prev-next-subtitle">previous</p>
<p class="prev-next-title">API Reference</p>
</div>
</a>
<a class='right-next' id="next-link" href="api/pyspark.sql.SparkSession.html" title="next page">
<div class="prev-next-info">
<p class="prev-next-subtitle">next</p>
<p class="prev-next-title">pyspark.sql.SparkSession</p>
</div>
<i class="fas fa-angle-right"></i>
</a>
</div>
</main>
</div>
</div>
<script src="../_static/js/index.be7d3bbb2ef33a8344ce.js"></script>
<footer class="footer mt-5 mt-md-0">
<div class="container">
<div class="footer-item">
<p class="copyright">
&copy; Copyright .<br>
</p>
</div>
<div class="footer-item">
<p class="sphinx-version">
Created using <a href="http://sphinx-doc.org/">Sphinx</a> 3.0.4.<br>
</p>
</div>
</div>
</footer>
</body>
</html>