You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/install/spark_cluster_mode.md
+14-15Lines changed: 14 additions & 15 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,7 +1,7 @@
1
1
---
2
2
layout: page
3
3
title: "Apache Zeppelin on Spark cluster mode"
4
-
description: ""
4
+
description: "This document will guide you how you can build and configure the environment on 3 types of Spark cluster manager with Apache Zeppelin using docker scripts."
5
5
group: install
6
6
---
7
7
<!--
@@ -56,12 +56,12 @@ spark_standalone bash;
56
56
```
57
57
58
58
### 3. Configure Spark interpreter in Zeppelin
59
-
Set Spark master as `spark://localhost:7077` in Zeppelin **Interpreters** setting page.
59
+
Set Spark master as `spark://<hostname>:7077` in Zeppelin **Interpreters** setting page.
After running single paragraph with Spark interpreter in Zeppelin, browse `https://localhost:8080` and check whether Spark cluster is running well or not.
64
+
After running single paragraph with Spark interpreter in Zeppelin, browse `https://<hostname>:8080` and check whether Spark cluster is running well or not.
You can simply set up [Spark on Yarn](http://spark.apache.org/docs/latest/running-on-yarn.html)docker environment with below steps.
76
+
You can simply set up [Spark on Yarn](http://spark.apache.org/docs/latest/running-on-yarn.html) environment with below steps using Docker scripts.
77
77
78
78
> **Note :** Since Apache Zeppelin and Spark use same `8080` port for their web UI, you might need to change `zeppelin.server.port` in `conf/zeppelin-site.xml`.
79
79
@@ -113,31 +113,30 @@ docker run -it \
113
113
114
114
### 3. Verify running Spark on Yarn.
115
115
116
-
You can simply verify the processes of Spark and Yarn is running well in Docker with below command.
117
-
116
+
You can simply verify the processes of Spark and Yarn are running well in Docker with below command.
118
117
119
118
```
120
-
ps -ef
119
+
ps -ef | grep spark
121
120
```
122
121
123
-
and also You can see the hdfs web ui on `http://localhost:50070/` and yarn ui on `http://localhost:8088/cluster` and spark ui on `http://localhost:8080/`.
124
-
122
+
You can also check each application web UI for HDFS on `http://<hostname>:50070/`, YARN on `http://<hostname>:8088/cluster` and Spark on `http://<hostname>:8080/`.
125
123
126
124
### 4. Configure Spark interpreter in Zeppelin
127
-
Set following configurations to the `zeppelin-env.sh`.
125
+
Set following configurations to `conf/zeppelin-env.sh`.
128
126
129
127
```
130
128
export MASTER=yarn-client
131
-
export HADOOP_CONF_DIR=[PATH OF HADOOP CONF]
132
-
export SPARK_HOME=[PATH OF SPARK HOME]
129
+
export HADOOP_CONF_DIR=[your_hadoop_conf_path]
130
+
export SPARK_HOME=[your_spark_home_path]
133
131
```
134
-
Hadoop configuration path(HADOOP_CONF_DIR) is the `scripts/docker/spark-cluster-managers/spark_yarn_cluster/hdfs_conf`.
135
132
136
-
Please make sure the spark master as `yarn-client` in Zeppelin **Interpreters** setting page.
133
+
`HADOOP_CONF_DIR`(Hadoop configuration path) is defined in `/scripts/docker/spark-cluster-managers/spark_yarn_cluster/hdfs_conf`.
134
+
135
+
Don't forget to set Spark `master` as `yarn-client` in Zeppelin **Interpreters** setting page like below.
After running single paragraph with Spark interpreter in Zeppelin, browse `http://localhost:8088/cluster/apps` and check zeppelin application running well or not.
140
+
After running a single paragraph with Spark interpreter in Zeppelin, browse `http://<hostname>:8088/cluster/apps` and check Zeppelin application is running well or not.
0 commit comments