Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Won't Fix
-
2.1.0, 2.1.1
-
None
-
None
Description
Spark returning empty result from when querying remote hadoop cluster.
All firewall settings removed.
Querying using JDBC working properly using hive-jdbc driver from version 1.1.1
Code snippet is:
val spark = SparkSession.builder
.appName("RemoteSparkTest")
.master("local")
.getOrCreate()
val df = spark.read
.option("url", "jdbc:hive2://remote.hive.local:10000/default")
.option("user", "user")
.option("password", "pass")
.option("dbtable", "test_table")
.option("driver", "org.apache.hive.jdbc.HiveDriver")
.format("jdbc")
.load()
df.show()
Result:
+-------------------+ |test_table.test_col| +-------------------+ +-------------------+
All manipulations like:
df.select(*).show()
returns empty result too.
Attachments
Issue Links
- is duplicated by
-
SPARK-31457 spark jdbc read hive created the wrong PreparedStatement
-
- Resolved
-
- relates to
-
SPARK-22016 Add HiveDialect for JDBC connection to Hive
-
- Resolved
-