[ 
https://issues.apache.org/jira/browse/SPARK-27019?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16781670#comment-16781670
 ] 

shahid edited comment on SPARK-27019 at 3/2/19 9:44 PM:
--------------------------------------------------------

seems event reordering has happened. Job start event came after sql execution 
end event, when the query failed. Could you please share spark eventLog for the 
application, if possible.


was (Author: shahid):
seems event reordering has happened. Job start event came after sql execution 
end event, when the query failed. Could you please share spark eventLog for the 
application, if possible, as I'm unable to reproduce in our cluster.

> Spark UI's SQL tab shows inconsistent values
> --------------------------------------------
>
>                 Key: SPARK-27019
>                 URL: https://issues.apache.org/jira/browse/SPARK-27019
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL, Web UI
>    Affects Versions: 2.4.0
>            Reporter: peay
>            Priority: Major
>         Attachments: Screenshot from 2019-03-01 21-31-48.png, 
> application_1550040445209_4748, query-1-details.png, query-1-list.png, 
> query-job-1.png, screenshot-spark-ui-details.png, screenshot-spark-ui-list.png
>
>
> Since 2.4.0, I am frequently seeing broken outputs in the SQL tab of the 
> Spark UI, where submitted/duration make no sense, description has the ID 
> instead of the actual description.
> Clicking on the link to open a query, the SQL plan is missing as well.
> I have tried to increase `spark.scheduler.listenerbus.eventqueue.capacity` to 
> very large values like 30k out of paranoia that we may have too many events, 
> but to no avail. I have not identified anything particular that leads to 
> that: it doesn't occur in all my jobs, but it does occur in a lot of them 
> still.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to