You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Contributions via GitHub pull requests are gladly accepted from their original
4
+
author. Along with any pull requests, please state that the contribution is
5
+
your original work and that you license the work to the project under the
6
+
project's open source license. Whether or not you state this explicitly, by
7
+
submitting any copyrighted material via pull request, email, or other means
8
+
you agree to license the material under the project's open source license and
9
+
warrant that you have the legal authority to do so.
10
+
11
+
Please see [Contributing to Spark wiki page](https://cwiki.apache.org/SPARK/Contributing+to+Spark)
Copy file name to clipboardExpand all lines: README.md
+16-62Lines changed: 16 additions & 62 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -13,16 +13,19 @@ and Spark Streaming for stream processing.
13
13
## Online Documentation
14
14
15
15
You can find the latest Spark documentation, including a programming
16
-
guide, on the project webpage at <http://spark.apache.org/documentation.html>.
16
+
guide, on the [project web page](http://spark.apache.org/documentation.html).
17
17
This README file only contains basic setup instructions.
18
18
19
19
## Building Spark
20
20
21
-
Spark is built on Scala 2.10. To build Spark and its example programs, run:
21
+
Spark is built using [Apache Maven](http://maven.apache.org/).
22
+
To build Spark and its example programs, run:
22
23
23
-
./sbt/sbt assembly
24
+
mvn -DskipTests clean package
24
25
25
26
(You do not need to do this if you downloaded a pre-built package.)
27
+
More detailed documentation is available from the project site, at
[run all automated tests](https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark#ContributingtoSpark-AutomatedTesting).
79
+
74
80
## A Note About Hadoop Versions
75
81
76
82
Spark uses the Hadoop core library to talk to HDFS and other Hadoop-supported
77
83
storage systems. Because the protocols have changed in different versions of
78
84
Hadoop, you must build Spark against the same version that your cluster runs.
79
-
You can change the version by setting `-Dhadoop.version` when building Spark.
80
-
81
-
For Apache Hadoop versions 1.x, Cloudera CDH MRv1, and other Hadoop
When developing a Spark application, specify the Hadoop version by adding the
103
-
"hadoop-client" artifact to your project's dependencies. For example, if you're
104
-
using Hadoop 1.2.1 and build your application using SBT, add this entry to
105
-
`libraryDependencies`:
106
-
107
-
"org.apache.hadoop" % "hadoop-client" % "1.2.1"
108
85
109
-
If your project is built with Maven, add this to your POM file's `<dependencies>` section:
110
-
111
-
<dependency>
112
-
<groupId>org.apache.hadoop</groupId>
113
-
<artifactId>hadoop-client</artifactId>
114
-
<version>1.2.1</version>
115
-
</dependency>
116
-
117
-
118
-
## A Note About Thrift JDBC server and CLI for Spark SQL
119
-
120
-
Spark SQL supports Thrift JDBC server and CLI.
121
-
See sql-programming-guide.md for more information about using the JDBC server and CLI.
122
-
You can use those features by setting `-Phive` when building Spark as follows.
123
-
124
-
$ sbt/sbt -Phive assembly
86
+
Please refer to the build documentation at
87
+
["Specifying the Hadoop Version"](http://spark.apache.org/docs/latest/building-spark.html#specifying-the-hadoop-version)
88
+
for detailed guidance on building for a particular distribution of Hadoop, including
89
+
building for particular Hive and Hive Thriftserver distributions. See also
90
+
["Third Party Hadoop Distributions"](http://spark.apache.org/docs/latest/hadoop-third-party-distributions.html)
91
+
for guidance on building a Spark application that works with a particular
92
+
distribution.
125
93
126
94
## Configuration
127
95
128
96
Please refer to the [Configuration guide](http://spark.apache.org/docs/latest/configuration.html)
129
97
in the online documentation for an overview on how to configure Spark.
130
-
131
-
132
-
## Contributing to Spark
133
-
134
-
Contributions via GitHub pull requests are gladly accepted from their original
135
-
author. Along with any pull requests, please state that the contribution is
136
-
your original work and that you license the work to the project under the
137
-
project's open source license. Whether or not you state this explicitly, by
138
-
submitting any copyrighted material via pull request, email, or other means
139
-
you agree to license the material under the project's open source license and
140
-
warrant that you have the legal authority to do so.
141
-
142
-
Please see [Contributing to Spark wiki page](https://cwiki.apache.org/SPARK/Contributing+to+Spark)
0 commit comments