Hi guys, I also experienced a situation when Spark 1.6.2 ignored my hint to do a broadcast join (i.e. broadcast(df)) with a small dataset. However, this happened only in 1 of 3 cases. Setting the "spark.sql.autoBroadcastJoinThreshold" property did not have any impact as well. All 3 cases work fine in Spark 2.0.
Is there any chance that Spark can neglect manually specified broadcast operation? In other words, is Spark forced to perform a broadcast if one specifies "df1.join(broadcast(df2), ...)"? Best regards, Anton 2016-11-26 21:04 GMT+01:00 Swapnil Shinde <swapnilushi...@gmail.com>: > I am using Spark 1.6.3 and below is the real plan (a,b,c in above were > just for illustration purpose) > > == Physical Plan == > Project [ltt#3800 AS ltt#3814,CASE WHEN isnull(etv_demo_id#3813) THEN > mr_demo_id#3801 ELSE etv_demo_id#3813 AS etv_demo_id#3815] > +- SortMergeOuterJoin [mr_demoname#3802,mr_demo_id#3801], > [mr_demoname#3810,mr_demo_id#3811], LeftOuter, None > :- Sort [mr_demoname#3802 ASC,mr_demo_id#3801 ASC], false, 0 > : +- TungstenExchange > hashpartitioning(mr_demoname#3802,mr_demo_id#3801,200), > None > : +- Project [_1#3797 AS ltt#3800,_2#3798 AS > mr_demo_id#3801,_3#3799 AS mr_demoname#3802] > : +- Scan ExistingRDD[_1#3797,_2#3798,_3#3799] > +- Sort [mr_demoname#3810 ASC,mr_demo_id#3811 ASC], false, 0 > +- TungstenExchange > hashpartitioning(mr_demoname#3810,mr_demo_id#3811,200), > None > +- Project [mr_demoname#3810,mr_demo_id#3811,etv_demo_id#3813] > +- Project [demogroup#3803 AS mr_demoname#3810,demovalue#3804 > AS mr_demo_id#3811,demoname#3805 AS mr_demo_value#3812,demovalue_etv_map#3806 > AS etv_demo_id#3813] > +- Filter ((map_type#3809 = master_roster_to_etv) && NOT > (demogroup#3803 = gender_age_id)) > +- Scan ExistingRDD[demogroup#3803, > demovalue#3804,demoname#3805,demovalue_etv_map#3806,demoname_etv_map#3807, > demovalue_old_map#3808,map_type#3809] > > > Thanks > Swapnil > > On Sat, Nov 26, 2016 at 2:32 PM, Benyi Wang <bewang.t...@gmail.com> wrote: > >> Could you post the result of explain `c.explain`? If it is broadcast >> join, you will see it in explain. >> >> On Sat, Nov 26, 2016 at 10:51 AM, Swapnil Shinde < >> swapnilushi...@gmail.com> wrote: >> >>> Hello >>> I am trying a broadcast join on dataframes but it is still doing >>> SortMergeJoin. I even try setting spark.sql.autoBroadcastJoinThreshold >>> higher but still no luck. >>> >>> Related piece of code- >>> val c = a.join(braodcast(b), "id") >>> >>> On a side note, if I do SizeEstimator.estimate(b) and it is really >>> high(460956584 bytes) compared to data it contains. b has just 85 rows and >>> around 4964 bytes. >>> Help is very much appreciated!! >>> >>> Thanks >>> Swapnil >>> >>> >>> >> >