From dd0226116fcae1dc7c5685d9a1c44b7cd147c075 Mon Sep 17 00:00:00 2001 From: ashahid Date: Fri, 28 Feb 2025 10:57:06 -0800 Subject: [PATCH] SPARK-33152. fix test failures after merge --- .../approved-plans-v1_4/q47.sf100/explain.txt | 4 ++-- .../approved-plans-v1_4/q47.sf100/simplified.txt | 2 +- .../approved-plans-v1_4/q47/explain.txt | 4 ++-- .../approved-plans-v1_4/q47/simplified.txt | 2 +- .../approved-plans-v2_7/q47.sf100/explain.txt | 4 ++-- .../approved-plans-v2_7/q47.sf100/simplified.txt | 2 +- .../approved-plans-v2_7/q47/explain.txt | 8 ++++---- .../approved-plans-v2_7/q47/simplified.txt | 4 ++-- .../approved-plans-v2_7/q75/explain.txt | 4 ++-- .../approved-plans-v2_7/q75/simplified.txt | 2 +- 10 files changed, 18 insertions(+), 18 deletions(-) diff --git a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47.sf100/explain.txt b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47.sf100/explain.txt index 6877737154537..5c438d6772201 100644 --- a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47.sf100/explain.txt +++ b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47.sf100/explain.txt @@ -85,7 +85,7 @@ Input [7]: [ss_item_sk#1, ss_store_sk#2, ss_sales_price#3, ss_sold_date_sk#4, d_ Output [3]: [s_store_sk#9, s_store_name#10, s_company_name#11] Batched: true Location [not included in comparison]/{warehouse_dir}/store] -PushedFilters: [IsNotNull(s_store_sk), IsNotNull(s_company_name), IsNotNull(s_store_name)] +PushedFilters: [IsNotNull(s_store_sk), IsNotNull(s_store_name), IsNotNull(s_company_name)] ReadSchema: struct (8) ColumnarToRow [codegen id : 2] @@ -93,7 +93,7 @@ Input [3]: [s_store_sk#9, s_store_name#10, s_company_name#11] (9) Filter [codegen id : 2] Input [3]: [s_store_sk#9, s_store_name#10, s_company_name#11] -Condition : ((isnotnull(s_store_sk#9) AND isnotnull(s_company_name#11)) AND isnotnull(s_store_name#10)) +Condition : ((isnotnull(s_store_sk#9) AND isnotnull(s_store_name#10)) AND isnotnull(s_company_name#11)) (10) BroadcastExchange Input [3]: [s_store_sk#9, s_store_name#10, s_company_name#11] diff --git a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47.sf100/simplified.txt b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47.sf100/simplified.txt index d56b308bec2ec..4cc4834cd2c8d 100644 --- a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47.sf100/simplified.txt +++ b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47.sf100/simplified.txt @@ -58,7 +58,7 @@ TakeOrderedAndProject [sum_sales,avg_monthly_sales,s_store_name,i_category,i_bra InputAdapter BroadcastExchange #6 WholeStageCodegen (2) - Filter [s_store_sk,s_company_name,s_store_name] + Filter [s_store_sk,s_store_name,s_company_name] ColumnarToRow InputAdapter Scan parquet spark_catalog.default.store [s_store_sk,s_store_name,s_company_name] diff --git a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47/explain.txt b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47/explain.txt index e60eabd2b89ab..765d595542cbc 100644 --- a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47/explain.txt +++ b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47/explain.txt @@ -106,7 +106,7 @@ Input [8]: [i_brand#2, i_category#3, ss_store_sk#5, ss_sales_price#6, ss_sold_da Output [3]: [s_store_sk#12, s_store_name#13, s_company_name#14] Batched: true Location [not included in comparison]/{warehouse_dir}/store] -PushedFilters: [IsNotNull(s_store_sk), IsNotNull(s_company_name), IsNotNull(s_store_name)] +PushedFilters: [IsNotNull(s_store_sk), IsNotNull(s_store_name), IsNotNull(s_company_name)] ReadSchema: struct (14) ColumnarToRow [codegen id : 3] @@ -114,7 +114,7 @@ Input [3]: [s_store_sk#12, s_store_name#13, s_company_name#14] (15) Filter [codegen id : 3] Input [3]: [s_store_sk#12, s_store_name#13, s_company_name#14] -Condition : ((isnotnull(s_store_sk#12) AND isnotnull(s_company_name#14)) AND isnotnull(s_store_name#13)) +Condition : ((isnotnull(s_store_sk#12) AND isnotnull(s_store_name#13)) AND isnotnull(s_company_name#14)) (16) BroadcastExchange Input [3]: [s_store_sk#12, s_store_name#13, s_company_name#14] diff --git a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47/simplified.txt b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47/simplified.txt index 72d27af82f77f..6b3a03aedbf1b 100644 --- a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47/simplified.txt +++ b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v1_4/q47/simplified.txt @@ -51,7 +51,7 @@ TakeOrderedAndProject [sum_sales,avg_monthly_sales,s_store_name,i_category,i_bra InputAdapter BroadcastExchange #5 WholeStageCodegen (3) - Filter [s_store_sk,s_company_name,s_store_name] + Filter [s_store_sk,s_store_name,s_company_name] ColumnarToRow InputAdapter Scan parquet spark_catalog.default.store [s_store_sk,s_store_name,s_company_name] diff --git a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47.sf100/explain.txt b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47.sf100/explain.txt index 6c5a267ff93ca..cbd3815845580 100644 --- a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47.sf100/explain.txt +++ b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47.sf100/explain.txt @@ -121,7 +121,7 @@ Arguments: [ss_item_sk#1 ASC NULLS FIRST], false, 0 Output [3]: [i_item_sk#12, i_brand#13, i_category#14] Batched: true Location [not included in comparison]/{warehouse_dir}/item] -PushedFilters: [IsNotNull(i_item_sk), IsNotNull(i_brand), IsNotNull(i_category)] +PushedFilters: [IsNotNull(i_item_sk), IsNotNull(i_category), IsNotNull(i_brand)] ReadSchema: struct (16) ColumnarToRow [codegen id : 5] @@ -129,7 +129,7 @@ Input [3]: [i_item_sk#12, i_brand#13, i_category#14] (17) Filter [codegen id : 5] Input [3]: [i_item_sk#12, i_brand#13, i_category#14] -Condition : ((isnotnull(i_item_sk#12) AND isnotnull(i_brand#13)) AND isnotnull(i_category#14)) +Condition : ((isnotnull(i_item_sk#12) AND isnotnull(i_category#14)) AND isnotnull(i_brand#13)) (18) Exchange Input [3]: [i_item_sk#12, i_brand#13, i_category#14] diff --git a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47.sf100/simplified.txt b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47.sf100/simplified.txt index 0e3280bd69a3b..89900bbee37e1 100644 --- a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47.sf100/simplified.txt +++ b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47.sf100/simplified.txt @@ -68,7 +68,7 @@ TakeOrderedAndProject [sum_sales,avg_monthly_sales,d_moy,i_category,d_year,psum, InputAdapter Exchange [i_item_sk] #7 WholeStageCodegen (5) - Filter [i_item_sk,i_brand,i_category] + Filter [i_item_sk,i_category,i_brand] ColumnarToRow InputAdapter Scan parquet spark_catalog.default.item [i_item_sk,i_brand,i_category] diff --git a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47/explain.txt b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47/explain.txt index f7859846a83d4..4efc75b7fab07 100644 --- a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47/explain.txt +++ b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47/explain.txt @@ -50,7 +50,7 @@ TakeOrderedAndProject (45) Output [3]: [i_item_sk#1, i_brand#2, i_category#3] Batched: true Location [not included in comparison]/{warehouse_dir}/item] -PushedFilters: [IsNotNull(i_item_sk), IsNotNull(i_category), IsNotNull(i_brand)] +PushedFilters: [IsNotNull(i_item_sk), IsNotNull(i_brand), IsNotNull(i_category)] ReadSchema: struct (2) ColumnarToRow [codegen id : 4] @@ -58,7 +58,7 @@ Input [3]: [i_item_sk#1, i_brand#2, i_category#3] (3) Filter [codegen id : 4] Input [3]: [i_item_sk#1, i_brand#2, i_category#3] -Condition : ((isnotnull(i_item_sk#1) AND isnotnull(i_category#3)) AND isnotnull(i_brand#2)) +Condition : ((isnotnull(i_item_sk#1) AND isnotnull(i_brand#2)) AND isnotnull(i_category#3)) (4) Scan parquet spark_catalog.default.store_sales Output [4]: [ss_item_sk#4, ss_store_sk#5, ss_sales_price#6, ss_sold_date_sk#7] @@ -106,7 +106,7 @@ Input [8]: [i_brand#2, i_category#3, ss_store_sk#5, ss_sales_price#6, ss_sold_da Output [3]: [s_store_sk#12, s_store_name#13, s_company_name#14] Batched: true Location [not included in comparison]/{warehouse_dir}/store] -PushedFilters: [IsNotNull(s_store_sk), IsNotNull(s_store_name), IsNotNull(s_company_name)] +PushedFilters: [IsNotNull(s_store_sk), IsNotNull(s_company_name), IsNotNull(s_store_name)] ReadSchema: struct (14) ColumnarToRow [codegen id : 3] @@ -114,7 +114,7 @@ Input [3]: [s_store_sk#12, s_store_name#13, s_company_name#14] (15) Filter [codegen id : 3] Input [3]: [s_store_sk#12, s_store_name#13, s_company_name#14] -Condition : ((isnotnull(s_store_sk#12) AND isnotnull(s_store_name#13)) AND isnotnull(s_company_name#14)) +Condition : ((isnotnull(s_store_sk#12) AND isnotnull(s_company_name#14)) AND isnotnull(s_store_name#13)) (16) BroadcastExchange Input [3]: [s_store_sk#12, s_store_name#13, s_company_name#14] diff --git a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47/simplified.txt b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47/simplified.txt index 425aec10c8571..a83b2baac82fe 100644 --- a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47/simplified.txt +++ b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q47/simplified.txt @@ -28,7 +28,7 @@ TakeOrderedAndProject [sum_sales,avg_monthly_sales,d_moy,i_category,d_year,psum, BroadcastHashJoin [ss_sold_date_sk,d_date_sk] Project [i_brand,i_category,ss_store_sk,ss_sales_price,ss_sold_date_sk] BroadcastHashJoin [i_item_sk,ss_item_sk] - Filter [i_item_sk,i_category,i_brand] + Filter [i_item_sk,i_brand,i_category] ColumnarToRow InputAdapter Scan parquet spark_catalog.default.item [i_item_sk,i_brand,i_category] @@ -51,7 +51,7 @@ TakeOrderedAndProject [sum_sales,avg_monthly_sales,d_moy,i_category,d_year,psum, InputAdapter BroadcastExchange #5 WholeStageCodegen (3) - Filter [s_store_sk,s_store_name,s_company_name] + Filter [s_store_sk,s_company_name,s_store_name] ColumnarToRow InputAdapter Scan parquet spark_catalog.default.store [s_store_sk,s_store_name,s_company_name] diff --git a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q75/explain.txt b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q75/explain.txt index dd39bd4ca7e60..05fb0ee8a1ea1 100644 --- a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q75/explain.txt +++ b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q75/explain.txt @@ -149,7 +149,7 @@ Condition : isnotnull(cs_item_sk#1) Output [6]: [i_item_sk#7, i_brand_id#8, i_class_id#9, i_category_id#10, i_category#11, i_manufact_id#12] Batched: true Location [not included in comparison]/{warehouse_dir}/item] -PushedFilters: [IsNotNull(i_category), EqualTo(i_category,Books ), IsNotNull(i_item_sk), IsNotNull(i_category_id), IsNotNull(i_brand_id), IsNotNull(i_class_id), IsNotNull(i_manufact_id)] +PushedFilters: [IsNotNull(i_category), EqualTo(i_category,Books ), IsNotNull(i_item_sk), IsNotNull(i_class_id), IsNotNull(i_category_id), IsNotNull(i_brand_id), IsNotNull(i_manufact_id)] ReadSchema: struct (5) ColumnarToRow [codegen id : 1] @@ -157,7 +157,7 @@ Input [6]: [i_item_sk#7, i_brand_id#8, i_class_id#9, i_category_id#10, i_categor (6) Filter [codegen id : 1] Input [6]: [i_item_sk#7, i_brand_id#8, i_class_id#9, i_category_id#10, i_category#11, i_manufact_id#12] -Condition : ((((((isnotnull(i_category#11) AND (i_category#11 = Books )) AND isnotnull(i_item_sk#7)) AND isnotnull(i_category_id#10)) AND isnotnull(i_brand_id#8)) AND isnotnull(i_class_id#9)) AND isnotnull(i_manufact_id#12)) +Condition : ((((((isnotnull(i_category#11) AND (i_category#11 = Books )) AND isnotnull(i_item_sk#7)) AND isnotnull(i_class_id#9)) AND isnotnull(i_category_id#10)) AND isnotnull(i_brand_id#8)) AND isnotnull(i_manufact_id#12)) (7) Project [codegen id : 1] Output [5]: [i_item_sk#7, i_brand_id#8, i_class_id#9, i_category_id#10, i_manufact_id#12] diff --git a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q75/simplified.txt b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q75/simplified.txt index 123c112e03157..17e0f7f4cc438 100644 --- a/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q75/simplified.txt +++ b/sql/core/src/test/resources/tpcds-plan-stability/approved-plans-v2_7/q75/simplified.txt @@ -49,7 +49,7 @@ TakeOrderedAndProject [sales_cnt_diff,sales_amt_diff,prev_year,year,i_brand_id,i BroadcastExchange #6 WholeStageCodegen (1) Project [i_item_sk,i_brand_id,i_class_id,i_category_id,i_manufact_id] - Filter [i_category,i_item_sk,i_category_id,i_brand_id,i_class_id,i_manufact_id] + Filter [i_category,i_item_sk,i_class_id,i_category_id,i_brand_id,i_manufact_id] ColumnarToRow InputAdapter Scan parquet spark_catalog.default.item [i_item_sk,i_brand_id,i_class_id,i_category_id,i_category,i_manufact_id]