11diff --git a/pom.xml b/pom.xml
2- index 68e2c422a24..d971894ffe6 100644
2+ index a0e25ce4d8d..7db86212507 100644
33--- a/pom.xml
44+++ b/pom.xml
55@@ -152,6 +152,8 @@
@@ -38,7 +38,7 @@ index 68e2c422a24..d971894ffe6 100644
3838 </dependencyManagement>
3939
4040diff --git a/sql/core/pom.xml b/sql/core/pom.xml
41- index f08b33575fc..424e0da32fd 100644
41+ index e3d324c8edb..22342150522 100644
4242--- a/sql/core/pom.xml
4343+++ b/sql/core/pom.xml
4444@@ -77,6 +77,10 @@
@@ -216,7 +216,7 @@ index 0efe0877e9b..423d3b3d76d 100644
216216 -- SELECT_HAVING
217217 -- https://github.com/postgres/postgres/blob/REL_12_BETA2/src/test/regress/sql/select_having.sql
218218diff --git a/sql/core/src/test/scala/org/apache/spark/sql/CachedTableSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/CachedTableSuite.scala
219- index 9815cb816c9..95b5f9992b0 100644
219+ index e5494726695..00937f025c2 100644
220220--- a/sql/core/src/test/scala/org/apache/spark/sql/CachedTableSuite.scala
221221+++ b/sql/core/src/test/scala/org/apache/spark/sql/CachedTableSuite.scala
222222@@ -38,7 +38,7 @@ import org.apache.spark.sql.catalyst.util.DateTimeConstants
@@ -239,7 +239,7 @@ index 9815cb816c9..95b5f9992b0 100644
239239
240240 test("A cached table preserves the partitioning and ordering of its cached SparkPlan") {
241241diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala
242- index 5a8681aed97..da9d25e2eb4 100644
242+ index 6f3090d8908..c08a60fb0c2 100644
243243--- a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala
244244+++ b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala
245245@@ -28,7 +28,7 @@ import org.apache.spark.sql.catalyst.plans.logical.Expand
@@ -336,7 +336,7 @@ index 7ee18df3756..d09f70e5d99 100644
336336 assert(exchanges.size == 2)
337337 }
338338diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameWindowFunctionsSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameWindowFunctionsSuite.scala
339- index 47a311c71d5..342e71cfdd4 100644
339+ index a1d5d579338..c201d39cc78 100644
340340--- a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameWindowFunctionsSuite.scala
341341+++ b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameWindowFunctionsSuite.scala
342342@@ -24,8 +24,9 @@ import org.apache.spark.sql.catalyst.expressions.{AttributeReference, Expression
@@ -482,7 +482,7 @@ index a206e97c353..fea1149b67d 100644
482482
483483 test("SPARK-35884: Explain Formatted") {
484484diff --git a/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala
485- index 93275487f29..01e5c601763 100644
485+ index 93275487f29..33b2e7ad3b1 100644
486486--- a/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala
487487+++ b/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala
488488@@ -23,6 +23,7 @@ import java.nio.file.{Files, StandardOpenOption}
@@ -522,15 +522,15 @@ index 93275487f29..01e5c601763 100644
522522
523523 val fileScan = df.queryExecution.executedPlan collectFirst {
524524 case BatchScanExec(_, f: FileScan, _, _, _, _) => f
525- + case CometBatchScanExec(BatchScanExec(_, f: FileScan, _, _, _, _), _) => f
525+ + case CometBatchScanExec(BatchScanExec(_, f: FileScan, _, _, _, _), _, _ ) => f
526526 }
527527 assert(fileScan.nonEmpty)
528528 assert(fileScan.get.partitionFilters.nonEmpty)
529529@@ -1056,6 +1062,7 @@ class FileBasedDataSourceSuite extends QueryTest
530530
531531 val fileScan = df.queryExecution.executedPlan collectFirst {
532532 case BatchScanExec(_, f: FileScan, _, _, _, _) => f
533- + case CometBatchScanExec(BatchScanExec(_, f: FileScan, _, _, _, _), _) => f
533+ + case CometBatchScanExec(BatchScanExec(_, f: FileScan, _, _, _, _), _, _ ) => f
534534 }
535535 assert(fileScan.nonEmpty)
536536 assert(fileScan.get.partitionFilters.isEmpty)
@@ -624,7 +624,7 @@ index 7af826583bd..3c3def1eb67 100644
624624 assert(shuffleMergeJoins.size == 1)
625625 }
626626diff --git a/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala
627- index 4d256154c85..66a5473852d 100644
627+ index 44c8cb92fc3..f098beeca26 100644
628628--- a/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala
629629+++ b/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala
630630@@ -31,7 +31,8 @@ import org.apache.spark.sql.catalyst.analysis.UnresolvedRelation
@@ -822,7 +822,7 @@ index 4d256154c85..66a5473852d 100644
822822 checkAnswer(fullJoinDF, Row(100))
823823 }
824824 }
825- @@ -1583 ,6 +1612 ,9 @@ class JoinSuite extends QueryTest with SharedSparkSession with AdaptiveSparkPlan
825+ @@ -1611 ,6 +1640 ,9 @@ class JoinSuite extends QueryTest with SharedSparkSession with AdaptiveSparkPlan
826826 Seq(semiJoinDF, antiJoinDF).foreach { df =>
827827 assert(collect(df.queryExecution.executedPlan) {
828828 case j: ShuffledHashJoinExec if j.ignoreDuplicatedKey == ignoreDuplicatedKey => true
@@ -832,7 +832,7 @@ index 4d256154c85..66a5473852d 100644
832832 }.size == 1)
833833 }
834834 }
835- @@ -1627 ,14 +1659 ,20 @@ class JoinSuite extends QueryTest with SharedSparkSession with AdaptiveSparkPlan
835+ @@ -1655 ,14 +1687 ,20 @@ class JoinSuite extends QueryTest with SharedSparkSession with AdaptiveSparkPlan
836836
837837 test("SPARK-43113: Full outer join with duplicate stream-side references in condition (SMJ)") {
838838 def check(plan: SparkPlan): Unit = {
@@ -855,7 +855,7 @@ index 4d256154c85..66a5473852d 100644
855855 }
856856 dupStreamSideColTest("SHUFFLE_HASH", check)
857857 }
858- @@ -1770 ,7 +1808 ,8 @@ class ThreadLeakInSortMergeJoinSuite
858+ @@ -1798 ,7 +1836 ,8 @@ class ThreadLeakInSortMergeJoinSuite
859859 sparkConf.set(SHUFFLE_SPILL_NUM_ELEMENTS_FORCE_SPILL_THRESHOLD, 20))
860860 }
861861
@@ -879,7 +879,7 @@ index c26757c9cff..d55775f09d7 100644
879879 protected val baseResourcePath = {
880880 // use the same way as `SQLQueryTestSuite` to get the resource path
881881diff --git a/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala
882- index 793a0da6a86..181bfc16e4b 100644
882+ index 3cf2bfd17ab..49728c35c42 100644
883883--- a/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala
884884+++ b/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala
885885@@ -1521,7 +1521,8 @@ class SQLQuerySuite extends QueryTest with SharedSparkSession with AdaptiveSpark
@@ -2050,10 +2050,10 @@ index 8e88049f51e..8f3cf8a0f80 100644
20502050 case _ =>
20512051 throw new AnalysisException("Can not match ParquetTable in the query.")
20522052diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetIOSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetIOSuite.scala
2053- index 4f8a9e39716..fb55ac7a955 100644
2053+ index 8ed9ef1630e..eed2a6f5ad5 100644
20542054--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetIOSuite.scala
20552055+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetIOSuite.scala
2056- @@ -1335 ,7 +1335 ,8 @@ class ParquetIOSuite extends QueryTest with ParquetTest with SharedSparkSession
2056+ @@ -1345 ,7 +1345 ,8 @@ class ParquetIOSuite extends QueryTest with ParquetTest with SharedSparkSession
20572057 }
20582058 }
20592059
0 commit comments