Skip to content

Commit 1702cc5

Browse files
authored
Merge pull request #1 from apache/master
merge latest commits
2 parents 5d5d758 + 865d10f commit 1702cc5

File tree

291 files changed

+11395
-3641
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

291 files changed

+11395
-3641
lines changed

.gitignore

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,14 @@
11
*.class
2+
*.pyc
23

34
# Package Files #
45
*.jar
56
*.war
67
*.ear
78

89
# interpreter
9-
/interpreter/
10+
/interpreter/*
11+
!/interpreter/lib
1012

1113
# interpreter temp files
1214
spark/derby.log
@@ -21,6 +23,7 @@ lens/lens-cli-hist.log
2123
conf/zeppelin-env.sh
2224
conf/zeppelin-env.cmd
2325
conf/zeppelin-site.xml
26+
conf/shiro.ini
2427
conf/keystore
2528
conf/truststore
2629
conf/interpreter.json

.travis.yml

Lines changed: 1 addition & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ matrix:
3636
include:
3737
# Test License compliance using RAT tool
3838
- jdk: "oraclejdk7"
39-
env: SCALA_VER="2.11" SPARK_VER="2.0.0" HADOOP_VER="2.3" PROFILE="-Prat" BUILD_FLAG="clean" TEST_FLAG="org.apache.rat:apache-rat-plugin:check" TEST_PROJECTS=""
39+
env: SCALA_VER="2.11" PROFILE="-Prat" BUILD_FLAG="clean" TEST_FLAG="org.apache.rat:apache-rat-plugin:check" TEST_PROJECTS=""
4040

4141
# Test all modules with spark 2.0.0 and scala 2.11
4242
- jdk: "oraclejdk7"
@@ -58,18 +58,6 @@ matrix:
5858
- jdk: "oraclejdk7"
5959
env: SCALA_VER="2.10" SPARK_VER="1.4.1" HADOOP_VER="2.3" PROFILE="-Pspark-1.4 -Pr -Phadoop-2.3 -Ppyspark -Psparkr" BUILD_FLAG="package -DskipTests -DskipRat" TEST_FLAG="verify -DskipRat" TEST_PROJECTS="-pl zeppelin-interpreter,zeppelin-zengine,zeppelin-server,zeppelin-display,spark-dependencies,spark,r -Dtest=org.apache.zeppelin.rest.*Test,org.apache.zeppelin.spark.* -DfailIfNoTests=false"
6060

61-
# Test spark module for 1.3.1
62-
- jdk: "oraclejdk7"
63-
env: SCALA_VER="2.10" SPARK_VER="1.3.1" HADOOP_VER="2.3" PROFILE="-Pspark-1.3 -Phadoop-2.3 -Ppyspark" BUILD_FLAG="package -DskipTests -DskipRat" TEST_FLAG="verify -DskipRat" TEST_PROJECTS="-pl zeppelin-interpreter,zeppelin-zengine,zeppelin-server,zeppelin-display,spark-dependencies,spark -Dtest=org.apache.zeppelin.rest.*Test,org.apache.zeppelin.spark.* -DfailIfNoTests=false"
64-
65-
# Test spark module for 1.2.2
66-
- jdk: "oraclejdk7"
67-
env: SCALA_VER="2.10" SPARK_VER="1.2.2" HADOOP_VER="2.3" PROFILE="-Pspark-1.2 -Phadoop-2.3 -Ppyspark" BUILD_FLAG="package -DskipTests -DskipRat" TEST_FLAG="verify -DskipRat" TEST_PROJECTS="-pl zeppelin-interpreter,zeppelin-zengine,zeppelin-server,zeppelin-display,spark-dependencies,spark -Dtest=org.apache.zeppelin.rest.*Test,org.apache.zeppelin.spark.* -DfailIfNoTests=false"
68-
69-
# Test spark module for 1.1.1
70-
- jdk: "oraclejdk7"
71-
env: SCALA_VER="2.10" SPARK_VER="1.1.1" HADOOP_VER="2.3" PROFILE="-Pspark-1.1 -Phadoop-2.3 -Ppyspark" BUILD_FLAG="package -DskipTests -DskipRat" TEST_FLAG="verify -DskipRat" TEST_PROJECTS="-pl zeppelin-interpreter,zeppelin-zengine,zeppelin-server,zeppelin-display,spark-dependencies,spark -Dtest=org.apache.zeppelin.rest.*Test,org.apache.zeppelin.spark.* -DfailIfNoTests=false"
72-
7361
# Test selenium with spark module for 1.6.1
7462
- jdk: "oraclejdk7"
7563
env: TEST_SELENIUM="true" SCALA_VER="2.10" SPARK_VER="1.6.1" HADOOP_VER="2.3" PROFILE="-Pspark-1.6 -Phadoop-2.3 -Ppyspark -Pexamples" BUILD_FLAG="package -DskipTests -DskipRat" TEST_FLAG="verify -DskipRat" TEST_PROJECTS="-pl zeppelin-interpreter,zeppelin-zengine,zeppelin-server,zeppelin-display,spark-dependencies,spark -Dtest=org.apache.zeppelin.AbstractFunctionalSuite -DfailIfNoTests=false"

README.md

Lines changed: 2 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
**Documentation:** [User Guide](http://zeppelin.apache.org/docs/latest/index.html)<br/>
44
**Mailing Lists:** [User and Dev mailing list](http://zeppelin.apache.org/community.html)<br/>
5-
**Continuous Integration:** [![Build Status](https://secure.travis-ci.org/apache/zeppelin.png?branch=master)](https://travis-ci.org/apache/zeppelin) <br/>
5+
**Continuous Integration:** [![Build Status](https://travis-ci.org/apache/zeppelin.svg?branch=master)](https://travis-ci.org/apache/zeppelin) <br/>
66
**Contributing:** [Contribution Guide](https://zeppelin.apache.org/contribution/contributions.html)<br/>
77
**Issue Tracker:** [Jira](https://issues.apache.org/jira/browse/ZEPPELIN)<br/>
88
**License:** [Apache 2.0](https://github.com/apache/zeppelin/blob/master/LICENSE)
@@ -128,9 +128,6 @@ Available profiles are
128128
-Pspark-1.6
129129
-Pspark-1.5
130130
-Pspark-1.4
131-
-Pspark-1.3
132-
-Pspark-1.2
133-
-Pspark-1.1
134131
-Pcassandra-spark-1.5
135132
-Pcassandra-spark-1.4
136133
-Pcassandra-spark-1.3
@@ -192,7 +189,7 @@ enable 3rd party vendor repository (cloudera)
192189

193190
##### `-Pmapr[version]` (optional)
194191

195-
For the MapR Hadoop Distribution, these profiles will handle the Hadoop version. As MapR allows different versions of Spark to be installed, you should specify which version of Spark is installed on the cluster by adding a Spark profile (`-Pspark-1.2`, `-Pspark-1.3`, etc.) as needed.
192+
For the MapR Hadoop Distribution, these profiles will handle the Hadoop version. As MapR allows different versions of Spark to be installed, you should specify which version of Spark is installed on the cluster by adding a Spark profile (`-Pspark-1.6`, `-Pspark-2.0`, etc.) as needed.
196193
The correct Maven artifacts can be found for every version of MapR at http://doc.mapr.com
197194

198195
Available profiles are

alluxio/src/main/java/org/apache/zeppelin/alluxio/AlluxioInterpreter.java

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -135,7 +135,7 @@ private InterpreterResult interpret(String[] commands, InterpreterContext contex
135135

136136
private String[] splitAndRemoveEmpty(String st, String splitSeparator) {
137137
String[] voices = st.split(splitSeparator);
138-
ArrayList<String> result = new ArrayList<String>();
138+
ArrayList<String> result = new ArrayList<>();
139139
for (String voice : voices) {
140140
if (!voice.trim().isEmpty()) {
141141
result.add(voice);
@@ -145,7 +145,7 @@ private String[] splitAndRemoveEmpty(String st, String splitSeparator) {
145145
}
146146

147147
private String[] splitAndRemoveEmpty(String[] sts, String splitSeparator) {
148-
ArrayList<String> result = new ArrayList<String>();
148+
ArrayList<String> result = new ArrayList<>();
149149
for (String st : sts) {
150150
result.addAll(Arrays.asList(splitAndRemoveEmpty(st, splitSeparator)));
151151
}

alluxio/src/main/resources/interpreter-setting.json

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,6 +16,9 @@
1616
"defaultValue": "19998",
1717
"description": "Alluxio master port"
1818
}
19+
},
20+
"editor": {
21+
"editOnDblClick": false
1922
}
2023
}
21-
]
24+
]

alluxio/src/test/java/org/apache/zeppelin/alluxio/AlluxioInterpreterTest.java

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -93,7 +93,7 @@ public void testCompletion() {
9393
List expectedResultThree = Arrays.asList(
9494
new InterpreterCompletion("copyFromLocal", "copyFromLocal"),
9595
new InterpreterCompletion("copyToLocal", "copyToLocal"));
96-
List expectedResultNone = new ArrayList<String>();
96+
List expectedResultNone = new ArrayList<>();
9797

9898
List<InterpreterCompletion> resultOne = alluxioInterpreter.completion("c", 0);
9999
List<InterpreterCompletion> resultTwo = alluxioInterpreter.completion("co", 0);

angular/src/main/resources/interpreter-setting.json

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,9 @@
44
"name": "angular",
55
"className": "org.apache.zeppelin.angular.AngularInterpreter",
66
"properties": {
7-
7+
},
8+
"editor": {
9+
"editOnDblClick": true
810
}
911
}
1012
]

beam/src/main/resources/interpreter-setting.json

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,9 @@
55
"className": "org.apache.zeppelin.beam.BeamInterpreter",
66
"defaultInterpreter": true,
77
"properties": {
8-
8+
},
9+
"editor": {
10+
"editOnDblClick": false
911
}
1012
}
1113
]

bigquery/src/main/resources/interpreter-setting.json

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,8 @@
2424
}
2525
},
2626
"editor": {
27-
"language": "sql"
27+
"language": "sql",
28+
"editOnDblClick": false
2829
}
2930
}
3031
]

bin/interpreter.sh

Lines changed: 22 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -149,6 +149,28 @@ elif [[ "${INTERPRETER_ID}" == "hbase" ]]; then
149149
else
150150
echo "HBASE_HOME and HBASE_CONF_DIR are not set, configuration might not be loaded"
151151
fi
152+
elif [[ "${INTERPRETER_ID}" == "pig" ]]; then
153+
# autodetect HADOOP_CONF_HOME by heuristic
154+
if [[ -n "${HADOOP_HOME}" ]] && [[ -z "${HADOOP_CONF_DIR}" ]]; then
155+
if [[ -d "${HADOOP_HOME}/etc/hadoop" ]]; then
156+
export HADOOP_CONF_DIR="${HADOOP_HOME}/etc/hadoop"
157+
elif [[ -d "/etc/hadoop/conf" ]]; then
158+
export HADOOP_CONF_DIR="/etc/hadoop/conf"
159+
fi
160+
fi
161+
162+
if [[ -n "${HADOOP_CONF_DIR}" ]] && [[ -d "${HADOOP_CONF_DIR}" ]]; then
163+
ZEPPELIN_INTP_CLASSPATH+=":${HADOOP_CONF_DIR}"
164+
fi
165+
166+
# autodetect TEZ_CONF_DIR
167+
if [[ -n "${TEZ_CONF_DIR}" ]]; then
168+
ZEPPELIN_INTP_CLASSPATH+=":${TEZ_CONF_DIR}"
169+
elif [[ -d "/etc/tez/conf" ]]; then
170+
ZEPPELIN_INTP_CLASSPATH+=":/etc/tez/conf"
171+
else
172+
echo "TEZ_CONF_DIR is not set, configuration might not be loaded"
173+
fi
152174
fi
153175

154176
addJarInDirForIntp "${LOCAL_INTERPRETER_REPO}"

0 commit comments

Comments
 (0)