andygrove commented on issue #1839: URL: https://github.com/apache/datafusion-comet/issues/1839#issuecomment-2967466668
The "canocilization and exchange reuse" test is expected to fail and should be ignored until Comet supports DPP. The two exchanges are different. One contains a `CometScan` and one contains `FileScan` with a `SubqueryAdaptiveBroadcast` (due to fallback because of no DPP support). ``` Look for CometColumnarExchange hashpartitioning(store_id#3596, 5), ENSURE_REQUIREMENTS, [plan_id=543] +- Project [store_id#3596] +- Filter (isnotnull(units_sold#3595) AND (units_sold#3595 = 70)) +- FileScan parquet spark_catalog.default.fact_stats[units_sold#3595,store_id#3596] Batched: true, DataFilters: [isnotnull(units_sold#3595), (units_sold#3595 = 70)], Format: Parquet, Location: InMemoryFileIndex(25 paths)[file:/home/andy/git/apache/apache-spark-3.5.6/sql/core/spark-warehous..., PartitionFilters: [isnotnull(store_id#3596), dynamicpruningexpression(store_id#3596 IN dynamicpruning#3609)], PushedFilters: [IsNotNull(units_sold), EqualTo(units_sold,70)], ReadSchema: struct<units_sold:int> +- SubqueryAdaptiveBroadcast dynamicpruning#3609, 0, true, Project [store_id#3608], [store_id#3608] +- AdaptiveSparkPlan isFinalPlan=false +- CometProject [store_id#3608], [store_id#3608] +- CometFilter [units_sold#3607, store_id#3608], (isnotnull(units_sold#3607) AND (units_sold#3607 = 70)) +- CometScan parquet spark_catalog.default.fact_stats[units_sold#3607,store_id#3608] Batched: true, DataFilters: [isnotnull(units_sold#3607), (units_sold#3607 = 70)], Format: CometParquet, Location: InMemoryFileIndex(25 paths)[file:/home/andy/git/apache/apache-spark-3.5.6/sql/core/spark-warehous..., PartitionFilters: [isnotnull(store_id#3608)], PushedFilters: [IsNotNull(units_sold), EqualTo(units_sold,70)], ReadSchema: struct<units_sold:int> getOrElseUpdate with CometColumnarExchange hashpartitioning(store_id#3596, 5), ENSURE_REQUIREMENTS, [plan_id=568] +- CometProject [store_id#3596], [store_id#3596] +- CometFilter [units_sold#3595, store_id#3596], (isnotnull(units_sold#3595) AND (units_sold#3595 = 70)) +- CometScan parquet spark_catalog.default.fact_stats[units_sold#3595,store_id#3596] Batched: true, DataFilters: [isnotnull(units_sold#3595), (units_sold#3595 = 70)], Format: CometParquet, Location: InMemoryFileIndex(25 paths)[file:/home/andy/git/apache/apache-spark-3.5.6/sql/core/spark-warehous..., PartitionFilters: [isnotnull(store_id#3596), dynamicpruningexpression(true)], PushedFilters: [IsNotNull(units_sold), EqualTo(units_sold,70)], ReadSchema: struct<units_sold:int> NO REUSE Look for CometColumnarExchange hashpartitioning(store_id#3608, 5), ENSURE_REQUIREMENTS, [plan_id=570] +- CometProject [store_id#3608], [store_id#3608] +- CometFilter [units_sold#3607, store_id#3608], (isnotnull(units_sold#3607) AND (units_sold#3607 = 70)) +- CometScan parquet spark_catalog.default.fact_stats[units_sold#3607,store_id#3608] Batched: true, DataFilters: [isnotnull(units_sold#3607), (units_sold#3607 = 70)], Format: CometParquet, Location: InMemoryFileIndex(25 paths)[file:/home/andy/git/apache/apache-spark-3.5.6/sql/core/spark-warehous..., PartitionFilters: [isnotnull(store_id#3608)], PushedFilters: [IsNotNull(units_sold), EqualTo(units_sold,70)], ReadSchema: struct<units_sold:int> getOrElseUpdate with CometColumnarExchange hashpartitioning(store_id#3608, 5), ENSURE_REQUIREMENTS, [plan_id=580] +- CometProject [store_id#3608], [store_id#3608] +- CometFilter [units_sold#3607, store_id#3608], (isnotnull(units_sold#3607) AND (units_sold#3607 = 70)) +- CometScan parquet spark_catalog.default.fact_stats[units_sold#3607,store_id#3608] Batched: true, DataFilters: [isnotnull(units_sold#3607), (units_sold#3607 = 70)], Format: CometParquet, Location: InMemoryFileIndex(25 paths)[file:/home/andy/git/apache/apache-spark-3.5.6/sql/core/spark-warehous..., PartitionFilters: [isnotnull(store_id#3608)], PushedFilters: [IsNotNull(units_sold), EqualTo(units_sold,70)], ReadSchema: struct<units_sold:int> NO REUSE ``` -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: github-unsubscr...@datafusion.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: github-unsubscr...@datafusion.apache.org For additional commands, e-mail: github-h...@datafusion.apache.org