[ 
https://issues.apache.org/jira/browse/SPARK-25894?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dongjoon Hyun updated SPARK-25894:
----------------------------------
    Affects Version/s:     (was: 2.4.0)
                       3.0.0

> Include a count of the number of physical columns read for a columnar data 
> source in the metadata of FileSourceScanExec
> -----------------------------------------------------------------------------------------------------------------------
>
>                 Key: SPARK-25894
>                 URL: https://issues.apache.org/jira/browse/SPARK-25894
>             Project: Spark
>          Issue Type: Improvement
>          Components: SQL
>    Affects Versions: 3.0.0
>            Reporter: Michael Allman
>            Priority: Minor
>
> Knowing the number of physical columns Spark will read from a columnar file 
> format (such as Parquet) is extremely helpful (if not critical) in validating 
> an assumption about that number of columns based on a given query and schema 
> pruning functionality. For example, take a {{contacts}} table with a {{name}} 
> struct like {{name.first, name.last}}. Without schema pruning the following 
> query reads both columns in the {{name}} struct:
> {{select name.first from contacts}}
> With schema pruning, the same query reads only the {{name.first}} column.
> This issue (and related PR) proposes an additional metadata field for 
> {{FileSourceScanExec}} which identifies the number of columns Spark will read 
> from that file source. This metadata will be printed as part of a physical 
> plan explanation.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to