[ https://issues.apache.org/jira/browse/SPARK-22751?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16348430#comment-16348430 ]
Apache Spark commented on SPARK-22751: -------------------------------------- User 'lucio-yz' has created a pull request for this issue: https://github.com/apache/spark/pull/20472 > Improve ML RandomForest shuffle performance > ------------------------------------------- > > Key: SPARK-22751 > URL: https://issues.apache.org/jira/browse/SPARK-22751 > Project: Spark > Issue Type: Improvement > Components: ML > Affects Versions: 2.2.0 > Reporter: lucio35 > Priority: Minor > Original Estimate: 48h > Remaining Estimate: 48h > > When I try to use ML Randomforest to train a classifier with dataset > news20.binary, which has 19,996 training examples and 1,355,191 features, i > found that shuffle write size( 51 GB ) of findSplitsBySorting is very large > compared with the small data size( 133.52 MB ). I think it is useful to > replace groupByKey by reduceByKey to improve shuffle performance. -- This message was sent by Atlassian JIRA (v7.6.3#76005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org