[ 
https://issues.apache.org/jira/browse/SPARK-26222?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16722561#comment-16722561
 ] 

ASF GitHub Bot commented on SPARK-26222:
----------------------------------------

xuanyuanking opened a new pull request #23327: [SPARK-26222][SQL] Track file 
listing time
URL: https://github.com/apache/spark/pull/23327
 
 
   ## What changes were proposed in this pull request?
   
   Tracking file listing time and add them into scan node's metrics, also add 
the start and end timestamp metrics.
   
![image](https://user-images.githubusercontent.com/4833765/50058268-cacfd200-01b0-11e9-95dc-b2093fc54178.png)
   
   
   ## How was this patch tested?
   
   Add new tests in SQLMetricsSuite.
   

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Scan: track file listing time
> -----------------------------
>
>                 Key: SPARK-26222
>                 URL: https://issues.apache.org/jira/browse/SPARK-26222
>             Project: Spark
>          Issue Type: Sub-task
>          Components: SQL
>    Affects Versions: 2.4.0
>            Reporter: Reynold Xin
>            Priority: Major
>
> We should track file listing time and add it to the scan node's SQL metric, 
> so we have visibility how much is spent in file listing. It'd be useful to 
> track not just duration, but also start and end time so we can construct a 
> timeline.
> This requires a little bit design to define what file listing time means, 
> when we are reading from cache, vs not cache.
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to