@@ -19,13 +19,12 @@ package org.apache.spark.sql
1919
2020import org .scalatest .BeforeAndAfter
2121
22- class SingleLevelAggregateHashMapSuite extends DataFrameAggregateSuite with BeforeAndAfter {
22+ import org . apache . spark . SparkConf
2323
24- protected override def beforeAll (): Unit = {
25- sparkConf.set(" spark.sql.codegen.fallback" , " false" )
26- sparkConf.set(" spark.sql.codegen.aggregate.map.twolevel.enable" , " false" )
27- super .beforeAll()
28- }
24+ class SingleLevelAggregateHashMapSuite extends DataFrameAggregateSuite with BeforeAndAfter {
25+ override protected def sparkConf : SparkConf = super .sparkConf
26+ .set(" spark.sql.codegen.fallback" , " false" )
27+ .set(" spark.sql.codegen.aggregate.map.twolevel.enable" , " false" )
2928
3029 // adding some checking after each test is run, assuring that the configs are not changed
3130 // in test code
@@ -38,12 +37,9 @@ class SingleLevelAggregateHashMapSuite extends DataFrameAggregateSuite with Befo
3837}
3938
4039class TwoLevelAggregateHashMapSuite extends DataFrameAggregateSuite with BeforeAndAfter {
41-
42- protected override def beforeAll (): Unit = {
43- sparkConf.set(" spark.sql.codegen.fallback" , " false" )
44- sparkConf.set(" spark.sql.codegen.aggregate.map.twolevel.enable" , " true" )
45- super .beforeAll()
46- }
40+ override protected def sparkConf : SparkConf = super .sparkConf
41+ .set(" spark.sql.codegen.fallback" , " false" )
42+ .set(" spark.sql.codegen.aggregate.map.twolevel.enable" , " true" )
4743
4844 // adding some checking after each test is run, assuring that the configs are not changed
4945 // in test code
@@ -55,15 +51,14 @@ class TwoLevelAggregateHashMapSuite extends DataFrameAggregateSuite with BeforeA
5551 }
5652}
5753
58- class TwoLevelAggregateHashMapWithVectorizedMapSuite extends DataFrameAggregateSuite with
59- BeforeAndAfter {
54+ class TwoLevelAggregateHashMapWithVectorizedMapSuite
55+ extends DataFrameAggregateSuite
56+ with BeforeAndAfter {
6057
61- protected override def beforeAll (): Unit = {
62- sparkConf.set(" spark.sql.codegen.fallback" , " false" )
63- sparkConf.set(" spark.sql.codegen.aggregate.map.twolevel.enable" , " true" )
64- sparkConf.set(" spark.sql.codegen.aggregate.map.vectorized.enable" , " true" )
65- super .beforeAll()
66- }
58+ override protected def sparkConf : SparkConf = super .sparkConf
59+ .set(" spark.sql.codegen.fallback" , " false" )
60+ .set(" spark.sql.codegen.aggregate.map.twolevel.enable" , " true" )
61+ .set(" spark.sql.codegen.aggregate.map.vectorized.enable" , " true" )
6762
6863 // adding some checking after each test is run, assuring that the configs are not changed
6964 // in test code
0 commit comments