|
18 | 18 | package org.apache.spark.sql.execution |
19 | 19 |
|
20 | 20 | import org.apache.spark.sql.{Column, Dataset, Row} |
21 | | -import org.apache.spark.sql.internal.SQLConf |
22 | 21 | import org.apache.spark.sql.catalyst.analysis.UnresolvedAttribute |
23 | 22 | import org.apache.spark.sql.catalyst.expressions.{Add, Literal, Stack} |
24 | 23 | import org.apache.spark.sql.execution.aggregate.HashAggregateExec |
25 | 24 | import org.apache.spark.sql.execution.joins.BroadcastHashJoinExec |
26 | 25 | import org.apache.spark.sql.execution.joins.SortMergeJoinExec |
27 | 26 | import org.apache.spark.sql.expressions.scalalang.typed |
28 | 27 | import org.apache.spark.sql.functions.{avg, broadcast, col, max} |
| 28 | +import org.apache.spark.sql.internal.SQLConf |
29 | 29 | import org.apache.spark.sql.test.SharedSQLContext |
30 | 30 | import org.apache.spark.sql.types.{IntegerType, StringType, StructType} |
31 | 31 |
|
@@ -137,7 +137,6 @@ class WholeStageCodegenSuite extends SparkPlanTest with SharedSQLContext { |
137 | 137 | val df1 = Seq((1, 1), (2, 2), (3, 3)).toDF("key", "int") |
138 | 138 | val df2 = Seq((1, "1"), (2, "2"), (3, "3")).toDF("key", "str") |
139 | 139 |
|
140 | | - // join condition contains CodegenFallback expression (i.e., reflect) |
141 | 140 | val df = df1.join(df2, df1("key") === df2("key")) |
142 | 141 | .filter("int = 2 or reflect('java.lang.Integer', 'valueOf', str) = 1") |
143 | 142 | .select("int") |
|
0 commit comments