{% include JB/setup %}
Cloudera has officially provided CDH Docker Hub in their own container. Please check this guide page for more information.
You can import the Docker image by pulling it from Cloudera Docker Hub.
docker pull cloudera/quickstart:latest
docker run -it \ -p 80:80 \ -p 4040:4040 \ -p 8020:8020 \ -p 8022:8022 \ -p 8030:8030 \ -p 8032:8032 \ -p 8033:8033 \ -p 8040:8040 \ -p 8042:8042 \ -p 8088:8088 \ -p 8480:8480 \ -p 8485:8485 \ -p 8888:8888 \ -p 9083:9083 \ -p 10020:10020 \ -p 10033:10033 \ -p 18088:18088 \ -p 19888:19888 \ -p 25000:25000 \ -p 25010:25010 \ -p 25020:25020 \ -p 50010:50010 \ -p 50020:50020 \ -p 50070:50070 \ -p 50075:50075 \ -h quickstart.cloudera --privileged=true \ agitated_payne_backup /usr/bin/docker-quickstart;
To verify the application is running well, check the web UI for HDFS on http://<hostname>:50070/
and YARN on http://<hostname>:8088/cluster
.
Set following configurations to conf/zeppelin-env.sh
.
export MASTER=yarn-client export HADOOP_CONF_DIR=[your_hadoop_conf_path] export SPARK_HOME=[your_spark_home_path]
HADOOP_CONF_DIR
(Hadoop configuration path) is defined in /scripts/docker/spark-cluster-managers/cdh/hdfs_conf
.
Don't forget to set Spark master
as yarn-client
in Zeppelin Interpreters setting page like below.
After running a single paragraph with Spark interpreter in Zeppelin,
browse http://<hostname>:8088/cluster/apps
to check Zeppelin application is running well or not.