[ https://issues.apache.org/jira/browse/HIVE-10458?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14539875#comment-14539875 ]
Xuefu Zhang commented on HIVE-10458: ------------------------------------ Hi [~lirui], it doesn't seem making sense to have parallel order and then reduce again with one reducer. Thus, disabling parallel order for order by + limit seems better. As a side question, I remember that you mentioned that parallel order doesn't help much on performance. Could you quantify that? If so, maybe we shouldn't consider parallel order at all. > Enable parallel order by for spark [Spark Branch] > ------------------------------------------------- > > Key: HIVE-10458 > URL: https://issues.apache.org/jira/browse/HIVE-10458 > Project: Hive > Issue Type: Sub-task > Components: Spark > Reporter: Rui Li > Assignee: Rui Li > Attachments: HIVE-10458.1-spark.patch > > > We don't have to force reducer# to 1 as spark supports parallel sorting. -- This message was sent by Atlassian JIRA (v6.3.4#6332)