----------------------------------------------------------- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/28017/#review61373 -----------------------------------------------------------
ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SparkMapJoinResolver.java <https://reviews.apache.org/r/28017/#comment102946> currentTask seems to be the container for sparkWork. Do we need to pass in both of them? BTW, currentTask seems to be a misleading varaible name. ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SparkMapJoinResolver.java <https://reviews.apache.org/r/28017/#comment102947> Can we have a local variable for sparkWork.getAllWorkUnsorted()? It's called again later in the code. - Xuefu Zhang On Nov. 14, 2014, 12:03 a.m., Chao Sun wrote: > > ----------------------------------------------------------- > This is an automatically generated e-mail. To reply, visit: > https://reviews.apache.org/r/28017/ > ----------------------------------------------------------- > > (Updated Nov. 14, 2014, 12:03 a.m.) > > > Review request for hive, Jimmy Xiang, Szehon Ho, and Xuefu Zhang. > > > Bugs: HIVE-8776 > https://issues.apache.org/jira/browse/HIVE-8776 > > > Repository: hive-git > > > Description > ------- > > In SparkMapJoinResolver, we need to populate MapredLocalWork for all MapWorks > with MapJoinOperator. It is needed later in HashTableLoader, for example, to > retrieve small hash tables and direct fetch tables. > We need to set up information, such as aliasToWork, aliasToFetchWork, > directFetchOp, inputFileChangeSensitive, tmpPath, etc., for the new local > works. > > > Diffs > ----- > > ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HashTableLoader.java > d30ae51 > > ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SparkMapJoinResolver.java > 4b9a6cb > ql/src/java/org/apache/hadoop/hive/ql/plan/MapredLocalWork.java 785e4a0 > > Diff: https://reviews.apache.org/r/28017/diff/ > > > Testing > ------- > > > Thanks, > > Chao Sun > >