blob: 9cd9167fe26fc658afb9af9f93af0352482f2f64 [file] [log] [blame] [view]
---
layout: page
title: "Zeppelin-Context"
description: "The Zeppelin-Context is a system-wide container for a variety of user-specific settings and parameters that are accessible across notebooks, cells, and interpreters."
group: usage/other_features
---
<!--
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
-->
{% include JB/setup %}
# Zeppelin-Context
<div id="toc"></div>
The zeppelin-context is a system-wide container for common utility functions and
user-specific data. It implements functions for data input, data display, etc. that are
often needed but are not uniformly available in all interpreters.
Its single per-user instance is accessible across all of the user's notebooks and cells,
enabling data exchange between cells - even in different notebooks.
But the way in which the zeppelin-context is used, and the functionality available differs
depending on whether or not the associated interpreter is based on a programming language.
Details of how the zeppelin-context is used for different purposes and in different
environments is described below.
## Usage in Programming Language Cells
In many programming-language interpreters (e.g. Apache Spark, Python, R) the zeppelin-context is available
as a predefined variable `z` that can be used by directly invoking its methods.
The methods available on the `z` object are described below.
Other interpreters based on programming languages. also provide the predefined variable `z`.
### Exploring Spark DataFrames
In the Apache Spark interpreter, the zeppelin-context provides a `show` method, which,
using Zeppelin's `table` feature, can be used to nicely display a Spark DataFrame:
```scala
df = spark.read.csv('/path/to/csv')
z.show(df)
```
This display functionality using the `show` method is planned to be extended uniformly to
other interpreters that can access the `z` object (Flink already support to show table too).
### Object Exchange
`ZeppelinContext` extends map and it's shared between the Apache Spark and Python environments.
So you can put some objects using Scala (in an Apache Spark cell) and read it from Python, and vice versa.
<div class="codetabs">
<div data-lang="scala" markdown="1">
{% highlight scala %}
// Put/Get object from scala
%spark
val myObject = "hello"
z.put("objName", myObject)
z.get("objName")
{% endhighlight %}
</div>
<div data-lang="python" markdown="1">
{% highlight python %}
# Put/Get object from python
%spark.pyspark
val myObject = "hello"
z.put("objName", myObject)
myObject = z.get("objName")
# df is Python pandas DataFrame
# "table_name" must be table type. Currently only sql interpreter (%spark.sql or %jdbc) result is supported.
df = z.getAsDataFrame("table_name")
{% endhighlight %}
</div>
<div data-lang="R" markdown="1">
{% highlight python %}
# Get/Put object from R
%spark.r
z.put("objName", myObject)
myObject <- z.get("objName")
# df is R DataFrame
# "table_name" must be table type. Currently only sql interpreter (%spark.sql or %jdbc) result is supported.
df <- z.getAsDataFrame("table_name")
{% endhighlight %}
</div>
</div>
Currently, there're two types of data could be shared across interpreters:
* String Data
* Table Data
#### Share String Object
Here's one example we share one String object `maxAge` between Spark interpreter and jdbc interpreter.
```scala
%spark
z.put("maxAge", 83)
```
```sql
%jdbc(interpolate=true)
select * from bank where age = {maxAge}
```
<img src="{{BASE_PATH}}/assets/themes/zeppelin/img/docs-img/zeppelin_context_share_string.png" height="100%" width="100%">
#### Share Table Object
Here's one example we share one Table object between jdbc interpreter and python interpreter.
```sql
%jdbc(saveAs=bank)
select * from bank
```
```python
%python.ipython
%matplotlib inline
import warnings
warnings.filterwarnings("ignore")
from plotnine import ggplot, geom_histogram, aes, facet_wrap
bank = z.getAsDataFrame('bank')
(ggplot(bank, aes(x='age'))
```
<img src="{{BASE_PATH}}/assets/themes/zeppelin/img/docs-img/zeppelin_context_share_table.png" height="100%" width="100%">
### Form Creation
`ZeppelinContext` provides functions for creating forms.
In Scala and Python environments, you can create forms programmatically.
<div class="codetabs">
<div data-lang="scala" markdown="1">
{% highlight scala %}
%spark
/* Create text input form */
z.input("input_1")
/* Create text input form with default value */
z.input("input_2", "defaultValue")
/* Create select form */
z.select("select_1", Seq(("option1", "option1DisplayName"),
("option2", "option2DisplayName")))
/* Create select form with default value*/
z.select("select_2", "option1", Seq(("option1", "option1DisplayName"),
("option2", "option2DisplayName")))
{% endhighlight %}
</div>
<div data-lang="python" markdown="1">
{% highlight python %}
%spark.pyspark
# Create text input form
z.input("input_1")
# Create text input form with default value
z.input("input_2", "defaultValue")
# Create select form
z.select("select_1", [("option1", "option1DisplayName"),
("option2", "option2DisplayName")])
# Create select form with default value
z.select("select_2", [("option1", "option1DisplayName"),
("option2", "option2DisplayName")], "option1")
{% endhighlight %}
</div>
</div>
Patterns of the form ${ ... } are used to dynamically create additional HTML elements
for requesting user input (that replaces the corresponding pattern in the paragraph text).
Currently only [text](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/input/text),
[select](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/select) with
[options](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/option), and
[checkbox](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/input/checkbox) are supported.
Dynamic forms are described in detail here: [Dynamic Form](../usage/dynamic_form/intro.html).
In sql environment, you can create dynamic form in simple template.
```sql
%spark.sql
select * from ${table=defaultTableName} where text like '%${search}%'
```
To learn more about dynamic form, checkout [Dynamic Form](../usage/dynamic_form/intro.html).
## Usage with Embedded Commands
In certain interpreters (see table below) zeppelin-context features may be invoked by embedding
command strings into the paragraph text. Such embedded command strings are used to invoke
dynamic-forms and object-interpolation as described below.
| Interpreters that use Embedded Commands |
|-------------------------------------------------------------------|
|spark.sql (\*), bigquery, cassandra, elasticsearch, file, hbase, jdbc (\*), livy, markdown, neo4j, python, shell (\*), zengine |
Dynamic forms are available in all of the interpreters in the table above,
but object interpolation is only available in a small, but growing, list of interpreters
(marked with an asterisk in the table above).
Both these zeppelin-context features are described below.
### Object Interpolation
Some interpreters can interpolate object values from `z` into the paragraph text by using the
`{variable-name}` syntax. The value of any object previously `put` into `z` can be
interpolated into a paragraph text by using such a pattern containing the object's name.
The following example shows one use of this facility:
####In Scala cell:
```scala
%spark
z.put("minAge", 35)
```
####In later SQL cell:
```sql
%spark.sql
select * from members where age >= {minAge}
```
The interpolation of a `{var-name}` pattern is performed only when `z` contains an object with the specified name.
But the pattern is left unchanged if the named object does not exist in `z`.
Further, all `{var-name}` patterns within the paragraph text must be translatable for any interpolation to occur --
translation of only some of the patterns in a paragraph text is never done.
In some situations, it is necessary to use { and } characters in a paragraph text without invoking the
object interpolation mechanism. For these cases an escaping mechanism is available --
doubled braces {{ and }} should be used. The following example shows the use of {{ and }} for passing a
regular expression containing just { and } into the paragraph text.
```sql
%spark.sql
{% raw %}
select * from members where name rlike '[aeiou]{{3}}'
{% endraw %}
```
To summarize, patterns of the form `{var-name}` within the paragraph text will be interpolated only if a predefined
object of the specified name exists. Additionally, all such patterns within the paragraph text should also
be translatable for any interpolation to occur. Patterns of the form {% raw %} `{{any-text}}` {% endraw %} are translated into `{any-text}`.
These translations are performed only when all occurrences of `{`, `}`, {% raw %} `{{`, and `}}`{% endraw %} in the paragraph text conform
to one of the two forms described above. Paragraph text containing `{` and/or `}` characters used in any other way
(than `{var-name}` and {% raw %} `{{any-text}}` {% endraw %} ) is used as-is without any changes.
No error is flagged in any case. This behavior is identical to the implementation of a similar feature in
Jupyter's shell invocation using the `!` magic command.
This feature is disabled by default, and must be explicitly turned on for each interpreter independently
by setting the value of an interpreter-specific property to `true`.
Consult the _Configuration_ section of each interpreter's documentation
to find out if object interpolation is implemented, and the name of the parameter that must be set to `true` to
enable the feature. The name of the parameter used to enable this feature is different for each interpreter.
For example, the SparkSQL and Shell interpreters use the parameter names `zeppelin.spark.sql.interpolation` and
`zeppelin.shell.interpolation` respectively.
At present only the SparkSQL, JDBC, and Shell interpreters support object interpolation.
### Interpreter-Specific Functions
Some interpreters use a subclass of `BaseZepplinContext` augmented with interpreter-specific functions.
Such interpreter-specific functions are described within each interpreter's documentation.