http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out 
b/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out
index 98789d7..ddb436b 100644
--- a/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out
+++ b/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out
@@ -187,25 +187,28 @@ STAGE PLANS:
             alias: test1_n5
             Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE Column 
stats: NONE
             Select Operator
-              expressions: key (type: int), value (type: int), col_1 (type: 
string)
-              outputColumnNames: _col0, _col1, _col2
+              expressions: key (type: int), value (type: int), col_1 (type: 
string), key BETWEEN 100 AND 102 (type: boolean)
+              outputColumnNames: _col0, _col1, _col2, _col3
               Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE 
Column stats: NONE
               Sorted Merge Bucket Map Join Operator
                 condition map:
                      Left Outer Join 0 to 1
                 filter predicates:
-                  0 {_col0 BETWEEN 100 AND 102}
+                  0 {_col3}
                   1 
                 keys:
                   0 _col1 (type: int)
                   1 _col1 (type: int)
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                File Output Operator
-                  compressed: false
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                outputColumnNames: _col0, _col1, _col2, _col4, _col5, _col6
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col2 
(type: string), _col4 (type: int), _col5 (type: int), _col6 (type: string)
+                  outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
+                  File Output Operator
+                    compressed: false
+                    table:
+                        input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
 
   Stage: Stage-0
     Fetch Operator
@@ -283,7 +286,7 @@ STAGE PLANS:
                 Statistics: Num rows: 4 Data size: 38 Basic stats: COMPLETE 
Column stats: NONE
                 HashTable Sink Operator
                   filter predicates:
-                    0 {_col0 BETWEEN 100 AND 102}
+                    0 {_col3}
                     1 
                   keys:
                     0 
@@ -296,27 +299,31 @@ STAGE PLANS:
             alias: test1_n5
             Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE Column 
stats: NONE
             Select Operator
-              expressions: key (type: int), value (type: int), col_1 (type: 
string)
-              outputColumnNames: _col0, _col1, _col2
+              expressions: key (type: int), value (type: int), col_1 (type: 
string), key BETWEEN 100 AND 102 (type: boolean)
+              outputColumnNames: _col0, _col1, _col2, _col3
               Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE 
Column stats: NONE
               Map Join Operator
                 condition map:
                      Left Outer Join 0 to 1
                 filter predicates:
-                  0 {_col0 BETWEEN 100 AND 102}
+                  0 {_col3}
                   1 
                 keys:
                   0 
                   1 
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
+                outputColumnNames: _col0, _col1, _col2, _col4, _col5, _col6
                 Statistics: Num rows: 24 Data size: 476 Basic stats: COMPLETE 
Column stats: NONE
-                File Output Operator
-                  compressed: false
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col2 
(type: string), _col4 (type: int), _col5 (type: int), _col6 (type: string)
+                  outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
                   Statistics: Num rows: 24 Data size: 476 Basic stats: 
COMPLETE Column stats: NONE
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 24 Data size: 476 Basic stats: 
COMPLETE Column stats: NONE
+                    table:
+                        input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
       Execution mode: vectorized
       Local Work:
         Map Reduce Local Work
@@ -461,7 +468,7 @@ STAGE PLANS:
               Statistics: Num rows: 4 Data size: 38 Basic stats: COMPLETE 
Column stats: NONE
               HashTable Sink Operator
                 filter predicates:
-                  0 {_col0 BETWEEN 100 AND 102}
+                  0 {_col3}
                   1 
                 keys:
                   0 
@@ -474,27 +481,31 @@ STAGE PLANS:
             alias: test1_n5
             Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE Column 
stats: NONE
             Select Operator
-              expressions: key (type: int), value (type: int), col_1 (type: 
string)
-              outputColumnNames: _col0, _col1, _col2
+              expressions: key (type: int), value (type: int), col_1 (type: 
string), key BETWEEN 100 AND 102 (type: boolean)
+              outputColumnNames: _col0, _col1, _col2, _col3
               Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE 
Column stats: NONE
               Map Join Operator
                 condition map:
                      Left Outer Join 0 to 1
                 filter predicates:
-                  0 {_col0 BETWEEN 100 AND 102}
+                  0 {_col3}
                   1 
                 keys:
                   0 
                   1 
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
+                outputColumnNames: _col0, _col1, _col2, _col4, _col5, _col6
                 Statistics: Num rows: 24 Data size: 476 Basic stats: COMPLETE 
Column stats: NONE
-                File Output Operator
-                  compressed: false
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col2 
(type: string), _col4 (type: int), _col5 (type: int), _col6 (type: string)
+                  outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
                   Statistics: Num rows: 24 Data size: 476 Basic stats: 
COMPLETE Column stats: NONE
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 24 Data size: 476 Basic stats: 
COMPLETE Column stats: NONE
+                    table:
+                        input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
       Execution mode: vectorized
       Local Work:
         Map Reduce Local Work
@@ -586,8 +597,8 @@ STAGE PLANS:
             alias: test1_n5
             Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE Column 
stats: NONE
             Select Operator
-              expressions: key (type: int), value (type: int), col_1 (type: 
string)
-              outputColumnNames: _col0, _col1, _col2
+              expressions: key (type: int), value (type: int), col_1 (type: 
string), key BETWEEN 100 AND 102 (type: boolean)
+              outputColumnNames: _col0, _col1, _col2, _col3
               Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE 
Column stats: NONE
               Map Join Operator
                 condition map:
@@ -595,16 +606,20 @@ STAGE PLANS:
                 keys:
                   0 
                   1 
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                residual filter predicates: {((_col1 = _col4) or _col0 BETWEEN 
100 AND 102 or _col3 BETWEEN 100 AND 102)}
+                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6
+                residual filter predicates: {((_col1 = _col5) or _col3 or 
_col4 BETWEEN 100 AND 102)}
                 Statistics: Num rows: 24 Data size: 476 Basic stats: COMPLETE 
Column stats: NONE
-                File Output Operator
-                  compressed: false
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col2 
(type: string), _col4 (type: int), _col5 (type: int), _col6 (type: string)
+                  outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
                   Statistics: Num rows: 24 Data size: 476 Basic stats: 
COMPLETE Column stats: NONE
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 24 Data size: 476 Basic stats: 
COMPLETE Column stats: NONE
+                    table:
+                        input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
       Local Work:
         Map Reduce Local Work
 
@@ -698,8 +713,8 @@ STAGE PLANS:
             alias: test1_n5
             Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE Column 
stats: NONE
             Select Operator
-              expressions: key (type: int), value (type: int), col_1 (type: 
string)
-              outputColumnNames: _col0, _col1, _col2
+              expressions: key (type: int), value (type: int), col_1 (type: 
string), key BETWEEN 100 AND 102 (type: boolean)
+              outputColumnNames: _col0, _col1, _col2, _col3
               Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE 
Column stats: NONE
               Map Join Operator
                 condition map:
@@ -707,16 +722,20 @@ STAGE PLANS:
                 keys:
                   0 
                   1 
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                residual filter predicates: {((_col1 = _col4) or _col0 BETWEEN 
100 AND 102)}
+                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6
+                residual filter predicates: {((_col1 = _col5) or _col3)}
                 Statistics: Num rows: 24 Data size: 476 Basic stats: COMPLETE 
Column stats: NONE
-                File Output Operator
-                  compressed: false
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col2 
(type: string), _col4 (type: int), _col5 (type: int), _col6 (type: string)
+                  outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
                   Statistics: Num rows: 24 Data size: 476 Basic stats: 
COMPLETE Column stats: NONE
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 24 Data size: 476 Basic stats: 
COMPLETE Column stats: NONE
+                    table:
+                        input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
       Local Work:
         Map Reduce Local Work
 
@@ -893,8 +912,8 @@ STAGE PLANS:
             alias: test1_n5
             Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE Column 
stats: NONE
             Select Operator
-              expressions: key (type: int), value (type: int), col_1 (type: 
string)
-              outputColumnNames: _col0, _col1, _col2
+              expressions: key (type: int), value (type: int), col_1 (type: 
string), key BETWEEN 100 AND 102 (type: boolean)
+              outputColumnNames: _col0, _col1, _col2, _col3
               Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE 
Column stats: NONE
               Sorted Merge Bucket Map Join Operator
                 condition map:
@@ -902,14 +921,17 @@ STAGE PLANS:
                 keys:
                   0 _col1 (type: int)
                   1 _col1 (type: int)
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                residual filter predicates: {(_col0 BETWEEN 100 AND 102 or 
_col3 BETWEEN 100 AND 102)}
-                File Output Operator
-                  compressed: false
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6
+                residual filter predicates: {(_col3 or _col4 BETWEEN 100 AND 
102)}
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col2 
(type: string), _col4 (type: int), _col5 (type: int), _col6 (type: string)
+                  outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
+                  File Output Operator
+                    compressed: false
+                    table:
+                        input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
 
   Stage: Stage-0
     Fetch Operator
@@ -996,8 +1018,8 @@ STAGE PLANS:
             alias: test2_n3
             Statistics: Num rows: 4 Data size: 38 Basic stats: COMPLETE Column 
stats: NONE
             Select Operator
-              expressions: key (type: int), value (type: int), col_2 (type: 
string)
-              outputColumnNames: _col0, _col1, _col2
+              expressions: key (type: int), value (type: int), col_2 (type: 
string), key BETWEEN 100 AND 102 (type: boolean)
+              outputColumnNames: _col0, _col1, _col2, _col3
               Statistics: Num rows: 4 Data size: 38 Basic stats: COMPLETE 
Column stats: NONE
               Map Join Operator
                 condition map:
@@ -1005,16 +1027,20 @@ STAGE PLANS:
                 keys:
                   0 
                   1 
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                residual filter predicates: {((_col1 = _col4) or _col0 BETWEEN 
100 AND 102 or _col3 BETWEEN 100 AND 102)}
+                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6
+                residual filter predicates: {((_col1 = _col4) or _col0 BETWEEN 
100 AND 102 or _col6)}
                 Statistics: Num rows: 24 Data size: 476 Basic stats: COMPLETE 
Column stats: NONE
-                File Output Operator
-                  compressed: false
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col2 
(type: string), _col3 (type: int), _col4 (type: int), _col5 (type: string)
+                  outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
                   Statistics: Num rows: 24 Data size: 476 Basic stats: 
COMPLETE Column stats: NONE
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 24 Data size: 476 Basic stats: 
COMPLETE Column stats: NONE
+                    table:
+                        input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
       Local Work:
         Map Reduce Local Work
 
@@ -1215,8 +1241,8 @@ STAGE PLANS:
             alias: test2_n3
             Statistics: Num rows: 4 Data size: 38 Basic stats: COMPLETE Column 
stats: NONE
             Select Operator
-              expressions: key (type: int), value (type: int), col_2 (type: 
string)
-              outputColumnNames: _col0, _col1, _col2
+              expressions: key (type: int), value (type: int), col_2 (type: 
string), key BETWEEN 100 AND 102 (type: boolean)
+              outputColumnNames: _col0, _col1, _col2, _col3
               Statistics: Num rows: 4 Data size: 38 Basic stats: COMPLETE 
Column stats: NONE
               Map Join Operator
                 condition map:
@@ -1224,16 +1250,20 @@ STAGE PLANS:
                 keys:
                   0 
                   1 
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                residual filter predicates: {((_col1 = _col4) or _col3 BETWEEN 
100 AND 102)}
+                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6
+                residual filter predicates: {((_col1 = _col4) or _col6)}
                 Statistics: Num rows: 24 Data size: 476 Basic stats: COMPLETE 
Column stats: NONE
-                File Output Operator
-                  compressed: false
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col2 
(type: string), _col3 (type: int), _col4 (type: int), _col5 (type: string)
+                  outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
                   Statistics: Num rows: 24 Data size: 476 Basic stats: 
COMPLETE Column stats: NONE
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 24 Data size: 476 Basic stats: 
COMPLETE Column stats: NONE
+                    table:
+                        input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
       Local Work:
         Map Reduce Local Work
 
@@ -1302,8 +1332,8 @@ STAGE PLANS:
             alias: test2_n3
             Statistics: Num rows: 4 Data size: 38 Basic stats: COMPLETE Column 
stats: NONE
             Select Operator
-              expressions: key (type: int), value (type: int), col_2 (type: 
string)
-              outputColumnNames: _col0, _col1, _col2
+              expressions: key (type: int), value (type: int), col_2 (type: 
string), key BETWEEN 100 AND 102 (type: boolean)
+              outputColumnNames: _col0, _col1, _col2, _col3
               Statistics: Num rows: 4 Data size: 38 Basic stats: COMPLETE 
Column stats: NONE
               Sorted Merge Bucket Map Join Operator
                 condition map:
@@ -1311,14 +1341,17 @@ STAGE PLANS:
                 keys:
                   0 _col1 (type: int)
                   1 _col1 (type: int)
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                residual filter predicates: {(_col0 BETWEEN 100 AND 102 or 
_col3 BETWEEN 100 AND 102)}
-                File Output Operator
-                  compressed: false
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6
+                residual filter predicates: {(_col0 BETWEEN 100 AND 102 or 
_col6)}
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col2 
(type: string), _col3 (type: int), _col4 (type: int), _col5 (type: string)
+                  outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
+                  File Output Operator
+                    compressed: false
+                    table:
+                        input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
 
   Stage: Stage-0
     Fetch Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/smb_mapjoin_47.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/smb_mapjoin_47.q.out 
b/ql/src/test/results/clientpositive/smb_mapjoin_47.q.out
index 42940d9..0276595 100644
--- a/ql/src/test/results/clientpositive/smb_mapjoin_47.q.out
+++ b/ql/src/test/results/clientpositive/smb_mapjoin_47.q.out
@@ -404,8 +404,8 @@ STAGE PLANS:
             alias: test2_n5
             Statistics: Num rows: 4 Data size: 38 Basic stats: COMPLETE Column 
stats: NONE
             Select Operator
-              expressions: key (type: int), value (type: int), col_2 (type: 
string)
-              outputColumnNames: _col0, _col1, _col2
+              expressions: key (type: int), value (type: int), col_2 (type: 
string), key BETWEEN 100 AND 102 (type: boolean)
+              outputColumnNames: _col0, _col1, _col2, _col3
               Statistics: Num rows: 4 Data size: 38 Basic stats: COMPLETE 
Column stats: NONE
               HashTable Sink Operator
                 keys:
@@ -419,8 +419,8 @@ STAGE PLANS:
             alias: test1_n8
             Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE Column 
stats: NONE
             Select Operator
-              expressions: key (type: int), value (type: int), col_1 (type: 
string)
-              outputColumnNames: _col0, _col1, _col2
+              expressions: key (type: int), value (type: int), col_1 (type: 
string), key BETWEEN 100 AND 102 (type: boolean)
+              outputColumnNames: _col0, _col1, _col2, _col3
               Statistics: Num rows: 6 Data size: 56 Basic stats: COMPLETE 
Column stats: NONE
               Map Join Operator
                 condition map:
@@ -428,19 +428,23 @@ STAGE PLANS:
                 keys:
                   0 
                   1 
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                residual filter predicates: {((_col1 = _col4) or _col0 BETWEEN 
100 AND 102 or _col3 BETWEEN 100 AND 102)}
+                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7
+                residual filter predicates: {((_col1 = _col5) or _col3 or 
_col7)}
                 Statistics: Num rows: 24 Data size: 476 Basic stats: COMPLETE 
Column stats: NONE
-                Limit
-                  Number of rows: 10
-                  Statistics: Num rows: 10 Data size: 190 Basic stats: 
COMPLETE Column stats: NONE
-                  File Output Operator
-                    compressed: false
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col2 
(type: string), _col4 (type: int), _col5 (type: int), _col6 (type: string)
+                  outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
+                  Statistics: Num rows: 24 Data size: 476 Basic stats: 
COMPLETE Column stats: NONE
+                  Limit
+                    Number of rows: 10
                     Statistics: Num rows: 10 Data size: 190 Basic stats: 
COMPLETE Column stats: NONE
-                    table:
-                        input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                        output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                    File Output Operator
+                      compressed: false
+                      Statistics: Num rows: 10 Data size: 190 Basic stats: 
COMPLETE Column stats: NONE
+                      table:
+                          input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
+                          output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                          serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
       Local Work:
         Map Reduce Local Work
 

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/spark/auto_join13.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/auto_join13.q.out 
b/ql/src/test/results/clientpositive/spark/auto_join13.q.out
index bade284..0a0bdd3 100644
--- a/ql/src/test/results/clientpositive/spark/auto_join13.q.out
+++ b/ql/src/test/results/clientpositive/spark/auto_join13.q.out
@@ -64,13 +64,13 @@ STAGE PLANS:
                     predicate: (UDFToDouble(key) < 200.0D) (type: boolean)
                     Statistics: Num rows: 166 Data size: 1763 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
+                      expressions: UDFToDouble(key) (type: double)
                       outputColumnNames: _col0
                       Statistics: Num rows: 166 Data size: 1763 Basic stats: 
COMPLETE Column stats: NONE
                       Spark HashTable Sink Operator
                         keys:
                           0 (UDFToDouble(_col2) + UDFToDouble(_col0)) (type: 
double)
-                          1 UDFToDouble(_col0) (type: double)
+                          1 _col0 (type: double)
             Execution mode: vectorized
             Local Work:
               Map Reduce Local Work
@@ -109,7 +109,7 @@ STAGE PLANS:
                                Inner Join 0 to 1
                           keys:
                             0 (UDFToDouble(_col2) + UDFToDouble(_col0)) (type: 
double)
-                            1 UDFToDouble(_col0) (type: double)
+                            1 _col0 (type: double)
                           outputColumnNames: _col1, _col2
                           input vertices:
                             1 Map 4

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/spark/auto_join19.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/auto_join19.q.out 
b/ql/src/test/results/clientpositive/spark/auto_join19.q.out
index b0d2552..9ab1b8b 100644
--- a/ql/src/test/results/clientpositive/spark/auto_join19.q.out
+++ b/ql/src/test/results/clientpositive/spark/auto_join19.q.out
@@ -85,12 +85,12 @@ STAGE PLANS:
                         keys:
                           0 _col0 (type: string)
                           1 _col0 (type: string)
-                        outputColumnNames: _col0, _col4
+                        outputColumnNames: _col0, _col2
                         input vertices:
                           1 Map 2
                         Statistics: Num rows: 2200 Data size: 23372 Basic 
stats: COMPLETE Column stats: NONE
                         Select Operator
-                          expressions: UDFToInteger(_col0) (type: int), _col4 
(type: string)
+                          expressions: UDFToInteger(_col0) (type: int), _col2 
(type: string)
                           outputColumnNames: _col0, _col1
                           Statistics: Num rows: 2200 Data size: 23372 Basic 
stats: COMPLETE Column stats: NONE
                           File Output Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/spark/auto_join2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/auto_join2.q.out 
b/ql/src/test/results/clientpositive/spark/auto_join2.q.out
index 32026e4..6aeadb5 100644
--- a/ql/src/test/results/clientpositive/spark/auto_join2.q.out
+++ b/ql/src/test/results/clientpositive/spark/auto_join2.q.out
@@ -39,8 +39,8 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Spark HashTable Sink Operator
                         keys:
@@ -59,13 +59,13 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string), value (type: string)
+                      expressions: value (type: string), UDFToDouble(key) 
(type: double)
                       outputColumnNames: _col0, _col1
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Spark HashTable Sink Operator
                         keys:
-                          0 (UDFToDouble(_col0) + UDFToDouble(_col1)) (type: 
double)
-                          1 UDFToDouble(_col0) (type: double)
+                          0 (_col1 + _col3) (type: double)
+                          1 _col1 (type: double)
             Execution mode: vectorized
             Local Work:
               Map Reduce Local Work
@@ -84,8 +84,8 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Map Join Operator
                         condition map:
@@ -93,7 +93,7 @@ STAGE PLANS:
                         keys:
                           0 _col0 (type: string)
                           1 _col0 (type: string)
-                        outputColumnNames: _col0, _col1
+                        outputColumnNames: _col0, _col1, _col3
                         input vertices:
                           1 Map 2
                         Statistics: Num rows: 550 Data size: 5843 Basic stats: 
COMPLETE Column stats: NONE
@@ -101,14 +101,14 @@ STAGE PLANS:
                           condition map:
                                Inner Join 0 to 1
                           keys:
-                            0 (UDFToDouble(_col0) + UDFToDouble(_col1)) (type: 
double)
-                            1 UDFToDouble(_col0) (type: double)
-                          outputColumnNames: _col0, _col3
+                            0 (_col1 + _col3) (type: double)
+                            1 _col1 (type: double)
+                          outputColumnNames: _col0, _col4
                           input vertices:
                             1 Map 3
                           Statistics: Num rows: 605 Data size: 6427 Basic 
stats: COMPLETE Column stats: NONE
                           Select Operator
-                            expressions: UDFToInteger(_col0) (type: int), 
_col3 (type: string)
+                            expressions: UDFToInteger(_col0) (type: int), 
_col4 (type: string)
                             outputColumnNames: _col0, _col1
                             Statistics: Num rows: 605 Data size: 6427 Basic 
stats: COMPLETE Column stats: NONE
                             File Output Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/spark/auto_join32.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/auto_join32.q.out 
b/ql/src/test/results/clientpositive/spark/auto_join32.q.out
index ed3dfbe..7f7b5e9 100644
--- a/ql/src/test/results/clientpositive/spark/auto_join32.q.out
+++ b/ql/src/test/results/clientpositive/spark/auto_join32.q.out
@@ -499,10 +499,10 @@ STAGE PLANS:
                         keys:
                           0 _col0 (type: string)
                           1 _col0 (type: string)
-                        outputColumnNames: _col0, _col3
+                        outputColumnNames: _col0, _col2
                         Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE
                         Group By Operator
-                          keys: _col0 (type: string), _col3 (type: string)
+                          keys: _col0 (type: string), _col2 (type: string)
                           mode: hash
                           outputColumnNames: _col0, _col1
                           Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/spark/auto_join9.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/auto_join9.q.out 
b/ql/src/test/results/clientpositive/spark/auto_join9.q.out
index ce0e92d..e14fb06 100644
--- a/ql/src/test/results/clientpositive/spark/auto_join9.q.out
+++ b/ql/src/test/results/clientpositive/spark/auto_join9.q.out
@@ -77,12 +77,12 @@ STAGE PLANS:
                         keys:
                           0 _col0 (type: string)
                           1 _col0 (type: string)
-                        outputColumnNames: _col0, _col4
+                        outputColumnNames: _col0, _col2
                         input vertices:
                           1 Map 2
                         Statistics: Num rows: 550 Data size: 5843 Basic stats: 
COMPLETE Column stats: NONE
                         Select Operator
-                          expressions: UDFToInteger(_col0) (type: int), _col4 
(type: string)
+                          expressions: UDFToInteger(_col0) (type: int), _col2 
(type: string)
                           outputColumnNames: _col0, _col1
                           Statistics: Num rows: 550 Data size: 5843 Basic 
stats: COMPLETE Column stats: NONE
                           File Output Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/spark/auto_join_filters.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/auto_join_filters.q.out 
b/ql/src/test/results/clientpositive/spark/auto_join_filters.q.out
index 59cd394..b64241c 100644
--- a/ql/src/test/results/clientpositive/spark/auto_join_filters.q.out
+++ b/ql/src/test/results/clientpositive/spark/auto_join_filters.q.out
@@ -81,13 +81,13 @@ STAGE PLANS:
                   alias: a
                   Statistics: Num rows: 1 Data size: 220 Basic stats: COMPLETE 
Column stats: NONE
                   Select Operator
-                    expressions: key (type: int), value (type: int)
-                    outputColumnNames: _col0, _col1
+                    expressions: key (type: int), value (type: int), (key > 
40) (type: boolean), (value > 50) (type: boolean), (key = value) (type: boolean)
+                    outputColumnNames: _col0, _col1, _col2, _col3, _col4
                     Statistics: Num rows: 1 Data size: 220 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       sort order: 
                       Statistics: Num rows: 1 Data size: 220 Basic stats: 
COMPLETE Column stats: NONE
-                      value expressions: _col0 (type: int), _col1 (type: int)
+                      value expressions: _col0 (type: int), _col1 (type: int), 
_col2 (type: boolean), _col3 (type: boolean), _col4 (type: boolean)
             Execution mode: vectorized
         Map 4 
             Map Operator Tree:
@@ -95,13 +95,13 @@ STAGE PLANS:
                   alias: b
                   Statistics: Num rows: 1 Data size: 220 Basic stats: COMPLETE 
Column stats: NONE
                   Select Operator
-                    expressions: key (type: int), value (type: int)
-                    outputColumnNames: _col0, _col1
+                    expressions: key (type: int), value (type: int), (key > 
40) (type: boolean), (value > 50) (type: boolean), (key = value) (type: boolean)
+                    outputColumnNames: _col0, _col1, _col2, _col3, _col4
                     Statistics: Num rows: 1 Data size: 220 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       sort order: 
                       Statistics: Num rows: 1 Data size: 220 Basic stats: 
COMPLETE Column stats: NONE
-                      value expressions: _col0 (type: int), _col1 (type: int)
+                      value expressions: _col0 (type: int), _col1 (type: int), 
_col2 (type: boolean), _col3 (type: boolean), _col4 (type: boolean)
             Execution mode: vectorized
         Reducer 2 
             Reduce Operator Tree:
@@ -109,15 +109,15 @@ STAGE PLANS:
                 condition map:
                      Full Outer Join 0 to 1
                 filter predicates:
-                  0 {(VALUE._col0 > 40)} {(VALUE._col1 > 50)} {(VALUE._col0 = 
VALUE._col1)}
-                  1 {(VALUE._col0 > 40)} {(VALUE._col1 > 50)} {(VALUE._col0 = 
VALUE._col1)}
+                  0 {VALUE._col2} {VALUE._col3} {VALUE._col4}
+                  1 {VALUE._col2} {VALUE._col3} {VALUE._col4}
                 keys:
                   0 
                   1 
-                outputColumnNames: _col0, _col1, _col2, _col3
+                outputColumnNames: _col0, _col1, _col5, _col6
                 Statistics: Num rows: 1 Data size: 441 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
-                  expressions: hash(_col0,_col1,_col2,_col3) (type: int)
+                  expressions: hash(_col0,_col1,_col5,_col6) (type: int)
                   outputColumnNames: _col0
                   Statistics: Num rows: 1 Data size: 441 Basic stats: COMPLETE 
Column stats: NONE
                   Group By Operator
@@ -331,15 +331,15 @@ STAGE PLANS:
                   alias: a
                   Statistics: Num rows: 1 Data size: 220 Basic stats: COMPLETE 
Column stats: NONE
                   Select Operator
-                    expressions: key (type: int), value (type: int)
-                    outputColumnNames: _col0, _col1
+                    expressions: key (type: int), value (type: int), (key > 
40) (type: boolean), (value > 50) (type: boolean), (key = value) (type: boolean)
+                    outputColumnNames: _col0, _col1, _col2, _col3, _col4
                     Statistics: Num rows: 1 Data size: 220 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       key expressions: _col0 (type: int)
                       sort order: +
                       Map-reduce partition columns: _col0 (type: int)
                       Statistics: Num rows: 1 Data size: 220 Basic stats: 
COMPLETE Column stats: NONE
-                      value expressions: _col1 (type: int)
+                      value expressions: _col1 (type: int), _col2 (type: 
boolean), _col3 (type: boolean), _col4 (type: boolean)
             Execution mode: vectorized
         Map 4 
             Map Operator Tree:
@@ -347,15 +347,15 @@ STAGE PLANS:
                   alias: b
                   Statistics: Num rows: 1 Data size: 220 Basic stats: COMPLETE 
Column stats: NONE
                   Select Operator
-                    expressions: key (type: int), value (type: int)
-                    outputColumnNames: _col0, _col1
+                    expressions: key (type: int), value (type: int), (key > 
40) (type: boolean), (value > 50) (type: boolean), (key = value) (type: boolean)
+                    outputColumnNames: _col0, _col1, _col2, _col3, _col4
                     Statistics: Num rows: 1 Data size: 220 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       key expressions: _col1 (type: int)
                       sort order: +
                       Map-reduce partition columns: _col1 (type: int)
                       Statistics: Num rows: 1 Data size: 220 Basic stats: 
COMPLETE Column stats: NONE
-                      value expressions: _col0 (type: int)
+                      value expressions: _col0 (type: int), _col2 (type: 
boolean), _col3 (type: boolean), _col4 (type: boolean)
             Execution mode: vectorized
         Reducer 2 
             Reduce Operator Tree:
@@ -363,15 +363,15 @@ STAGE PLANS:
                 condition map:
                      Full Outer Join 0 to 1
                 filter predicates:
-                  0 {(KEY.reducesinkkey0 > 40)} {(VALUE._col0 > 50)} 
{(KEY.reducesinkkey0 = VALUE._col0)}
-                  1 {(VALUE._col0 > 40)} {(KEY.reducesinkkey0 > 50)} 
{(VALUE._col0 = KEY.reducesinkkey0)}
+                  0 {VALUE._col1} {VALUE._col2} {VALUE._col3}
+                  1 {VALUE._col1} {VALUE._col2} {VALUE._col3}
                 keys:
                   0 _col0 (type: int)
                   1 _col1 (type: int)
-                outputColumnNames: _col0, _col1, _col2, _col3
+                outputColumnNames: _col0, _col1, _col5, _col6
                 Statistics: Num rows: 1 Data size: 242 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
-                  expressions: hash(_col0,_col1,_col2,_col3) (type: int)
+                  expressions: hash(_col0,_col1,_col5,_col6) (type: int)
                   outputColumnNames: _col0
                   Statistics: Num rows: 1 Data size: 242 Basic stats: COMPLETE 
Column stats: NONE
                   Group By Operator
@@ -763,15 +763,15 @@ STAGE PLANS:
                   alias: a
                   Statistics: Num rows: 1 Data size: 220 Basic stats: COMPLETE 
Column stats: NONE
                   Select Operator
-                    expressions: key (type: int), value (type: int)
-                    outputColumnNames: _col0, _col1
+                    expressions: key (type: int), value (type: int), (key > 
40) (type: boolean), (value > 50) (type: boolean), (key = value) (type: boolean)
+                    outputColumnNames: _col0, _col1, _col2, _col3, _col4
                     Statistics: Num rows: 1 Data size: 220 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       key expressions: _col0 (type: int)
                       sort order: +
                       Map-reduce partition columns: _col0 (type: int)
                       Statistics: Num rows: 1 Data size: 220 Basic stats: 
COMPLETE Column stats: NONE
-                      value expressions: _col1 (type: int)
+                      value expressions: _col1 (type: int), _col2 (type: 
boolean), _col3 (type: boolean), _col4 (type: boolean)
             Execution mode: vectorized
         Map 4 
             Map Operator Tree:
@@ -779,15 +779,15 @@ STAGE PLANS:
                   alias: b
                   Statistics: Num rows: 1 Data size: 220 Basic stats: COMPLETE 
Column stats: NONE
                   Select Operator
-                    expressions: key (type: int), value (type: int)
-                    outputColumnNames: _col0, _col1
+                    expressions: key (type: int), value (type: int), (key > 
40) (type: boolean), (value > 50) (type: boolean), (key = value) (type: boolean)
+                    outputColumnNames: _col0, _col1, _col2, _col3, _col4
                     Statistics: Num rows: 1 Data size: 220 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       key expressions: _col1 (type: int)
                       sort order: +
                       Map-reduce partition columns: _col1 (type: int)
                       Statistics: Num rows: 1 Data size: 220 Basic stats: 
COMPLETE Column stats: NONE
-                      value expressions: _col0 (type: int)
+                      value expressions: _col0 (type: int), _col2 (type: 
boolean), _col3 (type: boolean), _col4 (type: boolean)
             Execution mode: vectorized
         Reducer 2 
             Reduce Operator Tree:
@@ -795,15 +795,15 @@ STAGE PLANS:
                 condition map:
                      Full Outer Join 0 to 1
                 filter predicates:
-                  0 {(KEY.reducesinkkey0 > 40)} {(VALUE._col0 > 50)} 
{(KEY.reducesinkkey0 = VALUE._col0)}
-                  1 {(VALUE._col0 > 40)} {(KEY.reducesinkkey0 > 50)} 
{(VALUE._col0 = KEY.reducesinkkey0)}
+                  0 {VALUE._col1} {VALUE._col2} {VALUE._col3}
+                  1 {VALUE._col1} {VALUE._col2} {VALUE._col3}
                 keys:
                   0 _col0 (type: int)
                   1 _col1 (type: int)
-                outputColumnNames: _col0, _col1, _col2, _col3
+                outputColumnNames: _col0, _col1, _col5, _col6
                 Statistics: Num rows: 1 Data size: 242 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
-                  expressions: hash(_col0,_col1,_col2,_col3) (type: int)
+                  expressions: hash(_col0,_col1,_col5,_col6) (type: int)
                   outputColumnNames: _col0
                   Statistics: Num rows: 1 Data size: 242 Basic stats: COMPLETE 
Column stats: NONE
                   Group By Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/spark/auto_join_stats.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/auto_join_stats.q.out 
b/ql/src/test/results/clientpositive/spark/auto_join_stats.q.out
index e181ff2..a6d0922 100644
--- a/ql/src/test/results/clientpositive/spark/auto_join_stats.q.out
+++ b/ql/src/test/results/clientpositive/spark/auto_join_stats.q.out
@@ -52,13 +52,13 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 6 Data size: 24 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 6 Data size: 24 Basic stats: 
COMPLETE Column stats: NONE
                       Spark HashTable Sink Operator
                         keys:
-                          0 (UDFToDouble(_col0) + UDFToDouble(_col1)) (type: 
double)
-                          1 UDFToDouble(_col0) (type: double)
+                          0 (_col1 + _col3) (type: double)
+                          1 _col1 (type: double)
             Execution mode: vectorized
             Local Work:
               Map Reduce Local Work
@@ -79,14 +79,15 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
                         key expressions: _col0 (type: string)
                         sort order: +
                         Map-reduce partition columns: _col0 (type: string)
                         Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
+                        value expressions: _col1 (type: double)
             Execution mode: vectorized
         Map 3 
             Map Operator Tree:
@@ -98,14 +99,15 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
                         key expressions: _col0 (type: string)
                         sort order: +
                         Map-reduce partition columns: _col0 (type: string)
                         Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
+                        value expressions: _col1 (type: double)
             Execution mode: vectorized
         Reducer 2 
             Local Work:
@@ -117,25 +119,29 @@ STAGE PLANS:
                 keys:
                   0 _col0 (type: string)
                   1 _col0 (type: string)
-                outputColumnNames: _col0, _col1
+                outputColumnNames: _col0, _col1, _col2, _col3
                 Statistics: Num rows: 550 Data size: 5843 Basic stats: 
COMPLETE Column stats: NONE
                 Map Join Operator
                   condition map:
                        Inner Join 0 to 1
                   keys:
-                    0 (UDFToDouble(_col0) + UDFToDouble(_col1)) (type: double)
-                    1 UDFToDouble(_col0) (type: double)
-                  outputColumnNames: _col0, _col1, _col2
+                    0 (_col1 + _col3) (type: double)
+                    1 _col1 (type: double)
+                  outputColumnNames: _col0, _col2, _col4
                   input vertices:
                     1 Map 4
                   Statistics: Num rows: 605 Data size: 6427 Basic stats: 
COMPLETE Column stats: NONE
-                  File Output Operator
-                    compressed: false
+                  Select Operator
+                    expressions: _col0 (type: string), _col2 (type: string), 
_col4 (type: string)
+                    outputColumnNames: _col0, _col1, _col2
                     Statistics: Num rows: 605 Data size: 6427 Basic stats: 
COMPLETE Column stats: NONE
-                    table:
-                        input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                        output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                    File Output Operator
+                      compressed: false
+                      Statistics: Num rows: 605 Data size: 6427 Basic stats: 
COMPLETE Column stats: NONE
+                      table:
+                          input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
+                          output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                          serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
 
   Stage: Stage-0
     Fetch Operator
@@ -211,14 +217,14 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 1 Data size: 300 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
+                      expressions: UDFToDouble(key) (type: double)
                       outputColumnNames: _col0
                       Statistics: Num rows: 1 Data size: 300 Basic stats: 
COMPLETE Column stats: NONE
                       Spark HashTable Sink Operator
                         keys:
-                          0 (UDFToDouble(_col0) + UDFToDouble(_col1)) (type: 
double)
-                          1 UDFToDouble(_col0) (type: double)
-                          2 UDFToDouble(_col0) (type: double)
+                          0 (_col1 + _col3) (type: double)
+                          1 _col0 (type: double)
+                          2 _col1 (type: double)
             Execution mode: vectorized
             Local Work:
               Map Reduce Local Work
@@ -232,14 +238,14 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 6 Data size: 24 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 6 Data size: 24 Basic stats: 
COMPLETE Column stats: NONE
                       Spark HashTable Sink Operator
                         keys:
-                          0 (UDFToDouble(_col0) + UDFToDouble(_col1)) (type: 
double)
-                          1 UDFToDouble(_col0) (type: double)
-                          2 UDFToDouble(_col0) (type: double)
+                          0 (_col1 + _col3) (type: double)
+                          1 _col0 (type: double)
+                          2 _col1 (type: double)
             Execution mode: vectorized
             Local Work:
               Map Reduce Local Work
@@ -260,14 +266,15 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
                         key expressions: _col0 (type: string)
                         sort order: +
                         Map-reduce partition columns: _col0 (type: string)
                         Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
+                        value expressions: _col1 (type: double)
             Execution mode: vectorized
         Map 3 
             Map Operator Tree:
@@ -279,14 +286,15 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
                         key expressions: _col0 (type: string)
                         sort order: +
                         Map-reduce partition columns: _col0 (type: string)
                         Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
+                        value expressions: _col1 (type: double)
             Execution mode: vectorized
         Reducer 2 
             Local Work:
@@ -298,23 +306,23 @@ STAGE PLANS:
                 keys:
                   0 _col0 (type: string)
                   1 _col0 (type: string)
-                outputColumnNames: _col0, _col1
+                outputColumnNames: _col0, _col1, _col2, _col3
                 Statistics: Num rows: 550 Data size: 5843 Basic stats: 
COMPLETE Column stats: NONE
                 Map Join Operator
                   condition map:
                        Inner Join 0 to 1
                        Inner Join 0 to 2
                   keys:
-                    0 (UDFToDouble(_col0) + UDFToDouble(_col1)) (type: double)
-                    1 UDFToDouble(_col0) (type: double)
-                    2 UDFToDouble(_col0) (type: double)
-                  outputColumnNames: _col0, _col1, _col3
+                    0 (_col1 + _col3) (type: double)
+                    1 _col0 (type: double)
+                    2 _col1 (type: double)
+                  outputColumnNames: _col0, _col2, _col5
                   input vertices:
                     1 Map 4
                     2 Map 5
                   Statistics: Num rows: 1210 Data size: 12854 Basic stats: 
COMPLETE Column stats: NONE
                   Select Operator
-                    expressions: _col0 (type: string), _col1 (type: string), 
_col3 (type: string)
+                    expressions: _col0 (type: string), _col2 (type: string), 
_col5 (type: string)
                     outputColumnNames: _col0, _col1, _col2
                     Statistics: Num rows: 1210 Data size: 12854 Basic stats: 
COMPLETE Column stats: NONE
                     File Output Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/spark/auto_join_stats2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/auto_join_stats2.q.out 
b/ql/src/test/results/clientpositive/spark/auto_join_stats2.q.out
index 6fe4b10..fbcc489 100644
--- a/ql/src/test/results/clientpositive/spark/auto_join_stats2.q.out
+++ b/ql/src/test/results/clientpositive/spark/auto_join_stats2.q.out
@@ -45,8 +45,8 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 1 Data size: 300 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 1 Data size: 300 Basic stats: 
COMPLETE Column stats: NONE
                       Spark HashTable Sink Operator
                         keys:
@@ -65,8 +65,8 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Spark HashTable Sink Operator
                         keys:
@@ -90,8 +90,8 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Map Join Operator
                         condition map:
@@ -99,7 +99,7 @@ STAGE PLANS:
                         keys:
                           0 
                           1 
-                        outputColumnNames: _col0, _col1
+                        outputColumnNames: _col0, _col1, _col2, _col3
                         input vertices:
                           1 Map 2
                         Statistics: Num rows: 500 Data size: 155812 Basic 
stats: COMPLETE Column stats: NONE
@@ -109,15 +109,15 @@ STAGE PLANS:
                           keys:
                             0 _col0 (type: string)
                             1 _col0 (type: string)
-                          outputColumnNames: _col0, _col1, _col2
+                          outputColumnNames: _col0, _col1, _col2, _col3, 
_col4, _col5
                           input vertices:
                             1 Map 3
                           Statistics: Num rows: 550 Data size: 171393 Basic 
stats: COMPLETE Column stats: NONE
                           Filter Operator
-                            predicate: ((UDFToDouble(_col2) + 
UDFToDouble(_col0)) = UDFToDouble(_col1)) (type: boolean)
+                            predicate: ((_col5 + _col1) = _col3) (type: 
boolean)
                             Statistics: Num rows: 275 Data size: 85696 Basic 
stats: COMPLETE Column stats: NONE
                             Select Operator
-                              expressions: _col2 (type: string), _col0 (type: 
string), _col1 (type: string)
+                              expressions: _col4 (type: string), _col0 (type: 
string), _col2 (type: string)
                               outputColumnNames: _col0, _col1, _col2
                               Statistics: Num rows: 275 Data size: 85696 Basic 
stats: COMPLETE Column stats: NONE
                               File Output Operator
@@ -206,8 +206,8 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Spark HashTable Sink Operator
                         keys:
@@ -226,14 +226,14 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 1 Data size: 300 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
+                      expressions: UDFToDouble(key) (type: double)
                       outputColumnNames: _col0
                       Statistics: Num rows: 1 Data size: 300 Basic stats: 
COMPLETE Column stats: NONE
                       Spark HashTable Sink Operator
                         keys:
-                          0 (UDFToDouble(_col0) + UDFToDouble(_col1)) (type: 
double)
-                          1 UDFToDouble(_col0) (type: double)
-                          2 UDFToDouble(_col0) (type: double)
+                          0 (_col1 + _col3) (type: double)
+                          1 _col0 (type: double)
+                          2 _col1 (type: double)
             Execution mode: vectorized
             Local Work:
               Map Reduce Local Work
@@ -247,14 +247,14 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 6 Data size: 24 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 6 Data size: 24 Basic stats: 
COMPLETE Column stats: NONE
                       Spark HashTable Sink Operator
                         keys:
-                          0 (UDFToDouble(_col0) + UDFToDouble(_col1)) (type: 
double)
-                          1 UDFToDouble(_col0) (type: double)
-                          2 UDFToDouble(_col0) (type: double)
+                          0 (_col1 + _col3) (type: double)
+                          1 _col0 (type: double)
+                          2 _col1 (type: double)
             Execution mode: vectorized
             Local Work:
               Map Reduce Local Work
@@ -273,8 +273,8 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
-                      outputColumnNames: _col0
+                      expressions: key (type: string), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Map Join Operator
                         condition map:
@@ -282,7 +282,7 @@ STAGE PLANS:
                         keys:
                           0 _col0 (type: string)
                           1 _col0 (type: string)
-                        outputColumnNames: _col0, _col1
+                        outputColumnNames: _col0, _col1, _col2, _col3
                         input vertices:
                           1 Map 2
                         Statistics: Num rows: 550 Data size: 5843 Basic stats: 
COMPLETE Column stats: NONE
@@ -291,16 +291,16 @@ STAGE PLANS:
                                Inner Join 0 to 1
                                Inner Join 0 to 2
                           keys:
-                            0 (UDFToDouble(_col0) + UDFToDouble(_col1)) (type: 
double)
-                            1 UDFToDouble(_col0) (type: double)
-                            2 UDFToDouble(_col0) (type: double)
-                          outputColumnNames: _col0, _col1, _col3
+                            0 (_col1 + _col3) (type: double)
+                            1 _col0 (type: double)
+                            2 _col1 (type: double)
+                          outputColumnNames: _col0, _col2, _col5
                           input vertices:
                             1 Map 3
                             2 Map 4
                           Statistics: Num rows: 1210 Data size: 12854 Basic 
stats: COMPLETE Column stats: NONE
                           Select Operator
-                            expressions: _col0 (type: string), _col1 (type: 
string), _col3 (type: string)
+                            expressions: _col0 (type: string), _col2 (type: 
string), _col5 (type: string)
                             outputColumnNames: _col0, _col1, _col2
                             Statistics: Num rows: 1210 Data size: 12854 Basic 
stats: COMPLETE Column stats: NONE
                             File Output Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_6.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_6.q.out 
b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_6.q.out
index 7491c7b..6fe72d7 100644
--- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_6.q.out
+++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_6.q.out
@@ -558,13 +558,13 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
+                      expressions: UDFToDouble(key) (type: double)
                       outputColumnNames: _col0
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
-                        key expressions: UDFToDouble(_col0) (type: double)
+                        key expressions: _col0 (type: double)
                         sort order: +
-                        Map-reduce partition columns: UDFToDouble(_col0) 
(type: double)
+                        Map-reduce partition columns: _col0 (type: double)
                         Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
             Execution mode: vectorized
         Map 5 
@@ -577,13 +577,13 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: int)
-                      outputColumnNames: _col0
+                      expressions: key (type: int), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
-                        key expressions: UDFToDouble(_col0) (type: double)
+                        key expressions: _col1 (type: double)
                         sort order: +
-                        Map-reduce partition columns: UDFToDouble(_col0) 
(type: double)
+                        Map-reduce partition columns: _col1 (type: double)
                         Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                         value expressions: _col0 (type: int)
             Execution mode: vectorized
@@ -612,8 +612,8 @@ STAGE PLANS:
                 condition map:
                      Inner Join 0 to 1
                 keys:
-                  0 UDFToDouble(_col0) (type: double)
-                  1 UDFToDouble(_col0) (type: double)
+                  0 _col0 (type: double)
+                  1 _col1 (type: double)
                 outputColumnNames: _col1
                 Statistics: Num rows: 550 Data size: 5843 Basic stats: 
COMPLETE Column stats: NONE
                 Reduce Output Operator
@@ -1071,13 +1071,13 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: string)
+                      expressions: UDFToDouble(key) (type: double)
                       outputColumnNames: _col0
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
-                        key expressions: UDFToDouble(_col0) (type: double)
+                        key expressions: _col0 (type: double)
                         sort order: +
-                        Map-reduce partition columns: UDFToDouble(_col0) 
(type: double)
+                        Map-reduce partition columns: _col0 (type: double)
                         Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
             Execution mode: vectorized
         Map 5 
@@ -1090,13 +1090,13 @@ STAGE PLANS:
                     predicate: key is not null (type: boolean)
                     Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
-                      expressions: key (type: int)
-                      outputColumnNames: _col0
+                      expressions: key (type: int), UDFToDouble(key) (type: 
double)
+                      outputColumnNames: _col0, _col1
                       Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
-                        key expressions: UDFToDouble(_col0) (type: double)
+                        key expressions: _col1 (type: double)
                         sort order: +
-                        Map-reduce partition columns: UDFToDouble(_col0) 
(type: double)
+                        Map-reduce partition columns: _col1 (type: double)
                         Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
                         value expressions: _col0 (type: int)
             Execution mode: vectorized
@@ -1125,8 +1125,8 @@ STAGE PLANS:
                 condition map:
                      Inner Join 0 to 1
                 keys:
-                  0 UDFToDouble(_col0) (type: double)
-                  1 UDFToDouble(_col0) (type: double)
+                  0 _col0 (type: double)
+                  1 _col1 (type: double)
                 outputColumnNames: _col1
                 Statistics: Num rows: 550 Data size: 5843 Basic stats: 
COMPLETE Column stats: NONE
                 Reduce Output Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/spark/bucket_map_join_spark1.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark1.q.out 
b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark1.q.out
index 80eefef..b951591 100644
--- a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark1.q.out
+++ b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark1.q.out
@@ -130,7 +130,7 @@ OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, 
`t2`.`value` AS `value1`
 FROM (SELECT `key`, `value`
 FROM `default`.`srcbucket_mapjoin_part_n19`
 WHERE `key` IS NOT NULL) AS `t0`
-INNER JOIN (SELECT `key`, `value`, CAST('2008-04-08' AS STRING) AS `ds`
+INNER JOIN (SELECT `key`, `value`
 FROM `default`.`srcbucket_mapjoin_part_2_n16`
 WHERE `ds` = '2008-04-08' AND `key` IS NOT NULL) AS `t2` ON `t0`.`key` = 
`t2`.`key`
 STAGE DEPENDENCIES:
@@ -461,7 +461,7 @@ OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, 
`t2`.`value` AS `value1`
 FROM (SELECT `key`, `value`
 FROM `default`.`srcbucket_mapjoin_part_n19`
 WHERE `key` IS NOT NULL) AS `t0`
-INNER JOIN (SELECT `key`, `value`, CAST('2008-04-08' AS STRING) AS `ds`
+INNER JOIN (SELECT `key`, `value`
 FROM `default`.`srcbucket_mapjoin_part_2_n16`
 WHERE `ds` = '2008-04-08' AND `key` IS NOT NULL) AS `t2` ON `t0`.`key` = 
`t2`.`key`
 STAGE DEPENDENCIES:

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/spark/bucket_map_join_spark2.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark2.q.out 
b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark2.q.out
index 6f25359..7752b4c 100644
--- a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark2.q.out
+++ b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark2.q.out
@@ -114,7 +114,7 @@ OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, 
`t2`.`value` AS `value1`
 FROM (SELECT `key`, `value`
 FROM `default`.`srcbucket_mapjoin_part_n12`
 WHERE `key` IS NOT NULL) AS `t0`
-INNER JOIN (SELECT `key`, `value`, CAST('2008-04-08' AS STRING) AS `ds`
+INNER JOIN (SELECT `key`, `value`
 FROM `default`.`srcbucket_mapjoin_part_2_n10`
 WHERE `ds` = '2008-04-08' AND `key` IS NOT NULL) AS `t2` ON `t0`.`key` = 
`t2`.`key`
 STAGE DEPENDENCIES:
@@ -445,7 +445,7 @@ OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, 
`t2`.`value` AS `value1`
 FROM (SELECT `key`, `value`
 FROM `default`.`srcbucket_mapjoin_part_n12`
 WHERE `key` IS NOT NULL) AS `t0`
-INNER JOIN (SELECT `key`, `value`, CAST('2008-04-08' AS STRING) AS `ds`
+INNER JOIN (SELECT `key`, `value`
 FROM `default`.`srcbucket_mapjoin_part_2_n10`
 WHERE `ds` = '2008-04-08' AND `key` IS NOT NULL) AS `t2` ON `t0`.`key` = 
`t2`.`key`
 STAGE DEPENDENCIES:

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/spark/bucket_map_join_spark3.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark3.q.out 
b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark3.q.out
index e6bc741..fe81ef5 100644
--- a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark3.q.out
+++ b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark3.q.out
@@ -114,7 +114,7 @@ OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, 
`t2`.`value` AS `value1`
 FROM (SELECT `key`, `value`
 FROM `default`.`srcbucket_mapjoin_part_n4`
 WHERE `key` IS NOT NULL) AS `t0`
-INNER JOIN (SELECT `key`, `value`, CAST('2008-04-08' AS STRING) AS `ds`
+INNER JOIN (SELECT `key`, `value`
 FROM `default`.`srcbucket_mapjoin_part_2_n3`
 WHERE `ds` = '2008-04-08' AND `key` IS NOT NULL) AS `t2` ON `t0`.`key` = 
`t2`.`key`
 STAGE DEPENDENCIES:
@@ -445,7 +445,7 @@ OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, 
`t2`.`value` AS `value1`
 FROM (SELECT `key`, `value`
 FROM `default`.`srcbucket_mapjoin_part_n4`
 WHERE `key` IS NOT NULL) AS `t0`
-INNER JOIN (SELECT `key`, `value`, CAST('2008-04-08' AS STRING) AS `ds`
+INNER JOIN (SELECT `key`, `value`
 FROM `default`.`srcbucket_mapjoin_part_2_n3`
 WHERE `ds` = '2008-04-08' AND `key` IS NOT NULL) AS `t2` ON `t0`.`key` = 
`t2`.`key`
 STAGE DEPENDENCIES:

http://git-wip-us.apache.org/repos/asf/hive/blob/55887646/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out 
b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out
index f072ab0..339fe44 100644
--- a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out
+++ b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out
@@ -74,7 +74,7 @@ POSTHOOK: Input: default@tbl1_n0
 POSTHOOK: Input: default@tbl2_n0
 POSTHOOK: Input: default@tbl3
 #### A masked pattern was here ####
-OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value` AS `val1`, `t4`.`value` AS 
`val2`, `t2`.`value` AS `val3`
+OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, `t4`.`value` AS `value0`, 
`t2`.`value` AS `value1`
 FROM (SELECT `key`, `value`
 FROM `default`.`tbl1_n0`
 WHERE `key` IS NOT NULL AND `value` IS NOT NULL) AS `t0`
@@ -474,7 +474,7 @@ POSTHOOK: Input: default@tbl1_n0
 POSTHOOK: Input: default@tbl2_n0
 POSTHOOK: Input: default@tbl3
 #### A masked pattern was here ####
-OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value` AS `val1`, `t4`.`value` AS 
`val2`, `t2`.`value` AS `val3`
+OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, `t4`.`value` AS `value0`, 
`t2`.`value` AS `value1`
 FROM (SELECT `key`, `value`
 FROM `default`.`tbl1_n0`
 WHERE `key` IS NOT NULL AND `value` IS NOT NULL) AS `t0`

Reply via email to