Modified: hive/branches/spark/ql/src/test/results/clientpositive/union26.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/union26.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/union26.q.out 
(original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/union26.q.out Fri 
Nov  7 20:41:34 2014
@@ -103,91 +103,88 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: srcpart
-            Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE 
Column stats: NONE
-            Filter Operator
-              predicate: ((ds = '2008-04-08') and (hr = '11')) (type: boolean)
-              Statistics: Num rows: 2000 Data size: 21248 Basic stats: 
COMPLETE Column stats: NONE
-              Lateral View Forward
-                Statistics: Num rows: 2000 Data size: 21248 Basic stats: 
COMPLETE Column stats: NONE
-                Select Operator
-                  expressions: key (type: string), value (type: string)
-                  outputColumnNames: key, value
-                  Statistics: Num rows: 2000 Data size: 21248 Basic stats: 
COMPLETE Column stats: NONE
+            Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE 
Column stats: NONE
+            Lateral View Forward
+              Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE 
Column stats: NONE
+              Select Operator
+                expressions: key (type: string), value (type: string)
+                outputColumnNames: key, value
+                Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
+                Lateral View Join Operator
+                  outputColumnNames: _col0, _col1, _col7
+                  Statistics: Num rows: 1000 Data size: 10624 Basic stats: 
COMPLETE Column stats: NONE
+                  Select Operator
+                    expressions: _col0 (type: string), _col1 (type: string)
+                    outputColumnNames: _col0, _col1
+                    Statistics: Num rows: 1000 Data size: 10624 Basic stats: 
COMPLETE Column stats: NONE
+                    Union
+                      Statistics: Num rows: 1275 Data size: 13545 Basic stats: 
COMPLETE Column stats: NONE
+                      Select Operator
+                        expressions: _col0 (type: string), _col1 (type: string)
+                        outputColumnNames: _col0, _col1
+                        Statistics: Num rows: 1275 Data size: 13545 Basic 
stats: COMPLETE Column stats: NONE
+                        Group By Operator
+                          aggregations: count(1)
+                          keys: _col0 (type: string), _col1 (type: string)
+                          mode: hash
+                          outputColumnNames: _col0, _col1, _col2
+                          Statistics: Num rows: 1275 Data size: 13545 Basic 
stats: COMPLETE Column stats: NONE
+                          Reduce Output Operator
+                            key expressions: _col0 (type: string), _col1 
(type: string)
+                            sort order: ++
+                            Map-reduce partition columns: _col0 (type: 
string), _col1 (type: string)
+                            Statistics: Num rows: 1275 Data size: 13545 Basic 
stats: COMPLETE Column stats: NONE
+                            value expressions: _col2 (type: bigint)
+              Select Operator
+                expressions: array(1,2,3) (type: array<int>)
+                outputColumnNames: _col0
+                Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
+                UDTF Operator
+                  Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
+                  function name: explode
                   Lateral View Join Operator
                     outputColumnNames: _col0, _col1, _col7
-                    Statistics: Num rows: 4000 Data size: 42496 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1000 Data size: 10624 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
                       expressions: _col0 (type: string), _col1 (type: string)
                       outputColumnNames: _col0, _col1
-                      Statistics: Num rows: 4000 Data size: 42496 Basic stats: 
COMPLETE Column stats: NONE
+                      Statistics: Num rows: 1000 Data size: 10624 Basic stats: 
COMPLETE Column stats: NONE
                       Union
-                        Statistics: Num rows: 4275 Data size: 45417 Basic 
stats: COMPLETE Column stats: NONE
+                        Statistics: Num rows: 1275 Data size: 13545 Basic 
stats: COMPLETE Column stats: NONE
                         Select Operator
                           expressions: _col0 (type: string), _col1 (type: 
string)
                           outputColumnNames: _col0, _col1
-                          Statistics: Num rows: 4275 Data size: 45417 Basic 
stats: COMPLETE Column stats: NONE
+                          Statistics: Num rows: 1275 Data size: 13545 Basic 
stats: COMPLETE Column stats: NONE
                           Group By Operator
                             aggregations: count(1)
                             keys: _col0 (type: string), _col1 (type: string)
                             mode: hash
                             outputColumnNames: _col0, _col1, _col2
-                            Statistics: Num rows: 4275 Data size: 45417 Basic 
stats: COMPLETE Column stats: NONE
+                            Statistics: Num rows: 1275 Data size: 13545 Basic 
stats: COMPLETE Column stats: NONE
                             Reduce Output Operator
                               key expressions: _col0 (type: string), _col1 
(type: string)
                               sort order: ++
                               Map-reduce partition columns: _col0 (type: 
string), _col1 (type: string)
-                              Statistics: Num rows: 4275 Data size: 45417 
Basic stats: COMPLETE Column stats: NONE
+                              Statistics: Num rows: 1275 Data size: 13545 
Basic stats: COMPLETE Column stats: NONE
                               value expressions: _col2 (type: bigint)
-                Select Operator
-                  expressions: array(1,2,3) (type: array<int>)
-                  outputColumnNames: _col0
-                  Statistics: Num rows: 2000 Data size: 21248 Basic stats: 
COMPLETE Column stats: NONE
-                  UDTF Operator
-                    Statistics: Num rows: 2000 Data size: 21248 Basic stats: 
COMPLETE Column stats: NONE
-                    function name: explode
-                    Lateral View Join Operator
-                      outputColumnNames: _col0, _col1, _col7
-                      Statistics: Num rows: 4000 Data size: 42496 Basic stats: 
COMPLETE Column stats: NONE
-                      Select Operator
-                        expressions: _col0 (type: string), _col1 (type: string)
-                        outputColumnNames: _col0, _col1
-                        Statistics: Num rows: 4000 Data size: 42496 Basic 
stats: COMPLETE Column stats: NONE
-                        Union
-                          Statistics: Num rows: 4275 Data size: 45417 Basic 
stats: COMPLETE Column stats: NONE
-                          Select Operator
-                            expressions: _col0 (type: string), _col1 (type: 
string)
-                            outputColumnNames: _col0, _col1
-                            Statistics: Num rows: 4275 Data size: 45417 Basic 
stats: COMPLETE Column stats: NONE
-                            Group By Operator
-                              aggregations: count(1)
-                              keys: _col0 (type: string), _col1 (type: string)
-                              mode: hash
-                              outputColumnNames: _col0, _col1, _col2
-                              Statistics: Num rows: 4275 Data size: 45417 
Basic stats: COMPLETE Column stats: NONE
-                              Reduce Output Operator
-                                key expressions: _col0 (type: string), _col1 
(type: string)
-                                sort order: ++
-                                Map-reduce partition columns: _col0 (type: 
string), _col1 (type: string)
-                                Statistics: Num rows: 4275 Data size: 45417 
Basic stats: COMPLETE Column stats: NONE
-                                value expressions: _col2 (type: bigint)
           TableScan
             Union
-              Statistics: Num rows: 4275 Data size: 45417 Basic stats: 
COMPLETE Column stats: NONE
+              Statistics: Num rows: 1275 Data size: 13545 Basic stats: 
COMPLETE Column stats: NONE
               Select Operator
                 expressions: _col0 (type: string), _col1 (type: string)
                 outputColumnNames: _col0, _col1
-                Statistics: Num rows: 4275 Data size: 45417 Basic stats: 
COMPLETE Column stats: NONE
+                Statistics: Num rows: 1275 Data size: 13545 Basic stats: 
COMPLETE Column stats: NONE
                 Group By Operator
                   aggregations: count(1)
                   keys: _col0 (type: string), _col1 (type: string)
                   mode: hash
                   outputColumnNames: _col0, _col1, _col2
-                  Statistics: Num rows: 4275 Data size: 45417 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 1275 Data size: 13545 Basic stats: 
COMPLETE Column stats: NONE
                   Reduce Output Operator
                     key expressions: _col0 (type: string), _col1 (type: string)
                     sort order: ++
                     Map-reduce partition columns: _col0 (type: string), _col1 
(type: string)
-                    Statistics: Num rows: 4275 Data size: 45417 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1275 Data size: 13545 Basic stats: 
COMPLETE Column stats: NONE
                     value expressions: _col2 (type: bigint)
       Reduce Operator Tree:
         Group By Operator
@@ -195,14 +192,14 @@ STAGE PLANS:
           keys: KEY._col0 (type: string), KEY._col1 (type: string)
           mode: mergepartial
           outputColumnNames: _col0, _col1, _col2
-          Statistics: Num rows: 2137 Data size: 22703 Basic stats: COMPLETE 
Column stats: NONE
+          Statistics: Num rows: 637 Data size: 6767 Basic stats: COMPLETE 
Column stats: NONE
           Select Operator
             expressions: _col2 (type: bigint), _col0 (type: string), _col1 
(type: string)
             outputColumnNames: _col0, _col1, _col2
-            Statistics: Num rows: 2137 Data size: 22703 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 637 Data size: 6767 Basic stats: COMPLETE 
Column stats: NONE
             File Output Operator
               compressed: false
-              Statistics: Num rows: 2137 Data size: 22703 Basic stats: 
COMPLETE Column stats: NONE
+              Statistics: Num rows: 637 Data size: 6767 Basic stats: COMPLETE 
Column stats: NONE
               table:
                   input format: org.apache.hadoop.mapred.TextInputFormat
                   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -238,8 +235,6 @@ PREHOOK: type: QUERY
 PREHOOK: Input: default@srcpart
 PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
 PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
-PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
 #### A masked pattern was here ####
 POSTHOOK: query: SELECT 
 count(1) as counts,
@@ -265,8 +260,6 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@srcpart
 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
-POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
 #### A masked pattern was here ####
 10     100     val_100
 10     103     val_103
@@ -601,8 +594,6 @@ PREHOOK: type: QUERY
 PREHOOK: Input: default@srcpart
 PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
 PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
-PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
 #### A masked pattern was here ####
 POSTHOOK: query: SELECT 
 count(1) as counts,
@@ -628,8 +619,6 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@srcpart
 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
-POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
 #### A masked pattern was here ####
 10     100     val_100
 10     103     val_103
@@ -964,8 +953,6 @@ PREHOOK: type: QUERY
 PREHOOK: Input: default@srcpart
 PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
 PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
-PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
 #### A masked pattern was here ####
 POSTHOOK: query: SELECT 
 count(1) as counts,
@@ -991,8 +978,6 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@srcpart
 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
-POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
 #### A masked pattern was here ####
 10     100     val_100
 10     103     val_103

Modified: hive/branches/spark/ql/src/test/results/clientpositive/union5.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/union5.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/union5.q.out 
(original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/union5.q.out Fri Nov 
 7 20:41:34 2014
@@ -102,14 +102,14 @@ STAGE PLANS:
           keys: KEY._col0 (type: string)
           mode: mergepartial
           outputColumnNames: _col0, _col1
-          Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column 
stats: COMPLETE
+          Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column 
stats: COMPLETE
           Select Operator
             expressions: _col0 (type: string), _col1 (type: bigint)
             outputColumnNames: _col0, _col1
-            Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column 
stats: COMPLETE
             File Output Operator
               compressed: false
-              Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE 
Column stats: COMPLETE
+              Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE 
Column stats: COMPLETE
               table:
                   input format: org.apache.hadoop.mapred.TextInputFormat
                   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

Modified: hive/branches/spark/ql/src/test/results/clientpositive/union7.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/union7.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/union7.q.out 
(original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/union7.q.out Fri Nov 
 7 20:41:34 2014
@@ -107,14 +107,14 @@ STAGE PLANS:
           keys: KEY._col0 (type: string)
           mode: mergepartial
           outputColumnNames: _col0, _col1
-          Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column 
stats: PARTIAL
+          Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column 
stats: PARTIAL
           Select Operator
             expressions: _col0 (type: string), _col1 (type: bigint)
             outputColumnNames: _col0, _col1
-            Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE 
Column stats: PARTIAL
+            Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column 
stats: PARTIAL
             File Output Operator
               compressed: false
-              Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE 
Column stats: PARTIAL
+              Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE 
Column stats: PARTIAL
               table:
                   input format: org.apache.hadoop.mapred.TextInputFormat
                   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

Modified: 
hive/branches/spark/ql/src/test/results/clientpositive/vector_between_in.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/vector_between_in.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/vector_between_in.q.out 
(original)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/vector_between_in.q.out 
Fri Nov  7 20:41:34 2014
@@ -662,15 +662,15 @@ POSTHOOK: Input: default@decimal_date_te
 14.9324324324
 19.1135135135
 20.3081081081
-22.1000000000
+22.1
 24.4891891892
 33.4486486486
 34.6432432432
 40.0189189189
 42.4081081081
 43.0054054054
-44.2000000000
-44.2000000000
+44.2
+44.2
 44.7972972973
 45.9918918919
 PREHOOK: query: SELECT COUNT(*) FROM decimal_date_test WHERE cdecimal1 NOT 
BETWEEN -2000 AND 4390.1351351351

Modified: 
hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_aggregate.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_aggregate.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_aggregate.q.out
 (original)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_aggregate.q.out
 Fri Nov  7 20:41:34 2014
@@ -106,14 +106,14 @@ POSTHOOK: query: SELECT cint,
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@decimal_vgby
 #### A masked pattern was here ####
-NULL   3072    9318.4351351351 -4298.1513513514        5018444.1081079808      
3072    11160.71538461538500    -5147.90769230769300    6010604.30769230735360
--3728  6       5831542.2692483780      -3367.6517567568        
5817556.0411483778      6       6984454.21109769200000  -4033.445769230769      
6967702.86724384584710
--563   2       -515.6210729730 -3367.6517567568        -3883.2728297298        
2       -617.56077692307690     -4033.445769230769      -4651.00654615384590
-762    2       5831542.2692483780      1531.2194054054 5833073.4886537834      
2       6984454.21109769200000  1833.9456923076925      6986288.15678999969250
-6981   3       5831542.269248378       -515.6210729730 5830511.0271024320      
3       6984454.211097692       -617.56077692307690     6983219.08954384584620
-253665376      1024    9767.0054054054 -9779.5486486487        
-347484.0818378374      1024    11697.96923076923100    -11712.99230769231000   
-416182.64030769233089
-528534767      1024    5831542.2692483780      -9777.1594594595        
11646372.8607481068     1024    6984454.21109769200000  -11710.13076923077100   
13948892.79980307629003
-626923679      1024    9723.4027027027 -9778.9513513514        
10541.0525297287        1024    11645.74615384615400    -11712.27692307692300   
12625.04759999997746
+NULL   3072    9318.4351351351 -4298.1513513514        5018444.1081079808      
3072    11160.715384615385      -5147.907692307693      6010604.3076923073536
+-3728  6       5831542.269248378       -3367.6517567568        
5817556.0411483778      6       6984454.211097692       -4033.445769230769      
6967702.8672438458471
+-563   2       -515.6210729730 -3367.6517567568        -3883.2728297298        
2       -617.5607769230769      -4033.445769230769      -4651.0065461538459
+762    2       5831542.269248378       1531.2194054054 5833073.4886537834      
2       6984454.211097692       1833.9456923076925      6986288.1567899996925
+6981   3       5831542.269248378       -515.6210729730 5830511.0271024320      
3       6984454.211097692       -617.5607769230769      6983219.0895438458462
+253665376      1024    9767.0054054054 -9779.5486486487        
-347484.0818378374      1024    11697.969230769231      -11712.99230769231      
-416182.64030769233089
+528534767      1024    5831542.269248378       -9777.1594594595        
11646372.8607481068     1024    6984454.211097692       -11710.130769230771     
13948892.79980307629003
+626923679      1024    9723.4027027027 -9778.9513513514        
10541.0525297287        1024    11645.746153846154      -11712.276923076923     
12625.04759999997746
 PREHOOK: query: -- Now add the others...
 EXPLAIN SELECT cint,
     COUNT(cdecimal1), MAX(cdecimal1), MIN(cdecimal1), SUM(cdecimal1), 
AVG(cdecimal1), STDDEV_POP(cdecimal1), STDDEV_SAMP(cdecimal1),
@@ -204,11 +204,11 @@ POSTHOOK: query: SELECT cint,
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@decimal_vgby
 #### A masked pattern was here ####
-NULL   3072    9318.4351351351 -4298.1513513514        5018444.1081079808      
1633.60810810806667     5695.483082135364       5696.4103077145055      3072    
11160.71538461538500    -5147.90769230769300    6010604.30769230735360  
1956.576923076922966667 6821.495748565159       6822.606289190924
--3728  6       5831542.2692483780      -3367.6517567568        
5817556.0411483778      969592.67352472963333   2174330.2092403853      
2381859.406131774       6       6984454.21109769200000  -4033.445769230769      
6967702.86724384584710  1161283.811207307641183333      2604201.2704476737      
2852759.5602156054
--563   2       -515.6210729730 -3367.6517567568        -3883.2728297298        
-1941.6364148649        1426.0153418918999      2016.6902366556308      2       
-617.56077692307690     -4033.445769230769      -4651.00654615384590    
-2325.50327307692295    1707.9424961538462      2415.395441814127
-762    2       5831542.2692483780      1531.2194054054 5833073.4886537834      
2916536.7443268917      2915005.5249214866      4122440.3477364695      2       
6984454.21109769200000  1833.9456923076925      6986288.15678999969250  
3493144.07839499984625  3491310.1327026924      4937458.140118758
-6981   3       5831542.269248378       -515.6210729730 5830511.0271024320      
1943503.67570081066667  2749258.455012492       3367140.1929065133      3       
6984454.211097692       -617.56077692307690     6983219.08954384584620  
2327739.696514615282066667      3292794.4113115156      4032833.0678006653
-253665376      1024    9767.0054054054 -9779.5486486487        
-347484.0818378374      -339.33992366976309     5708.9563478862 
5711.745967572779       1024    11697.96923076923100    -11712.99230769231000   
-416182.64030769233089  -406.428359675480791885 6837.632716002934       
6840.973851172274
-528534767      1024    5831542.2692483780      -9777.1594594595        
11646372.8607481068     11373.41099682432305    257528.92988206653      
257654.7686043977       1024    6984454.21109769200000  -11710.13076923077100   
13948892.79980307629003 13621.965624807691689482        308443.1074570801       
308593.82484083984
-626923679      1024    9723.4027027027 -9778.9513513514        
10541.0525297287        10.29399661106318       5742.09145323734        
5744.897264034267       1024    11645.74615384615400    -11712.27692307692300   
12625.04759999997746    12.329148046874977988   6877.318722794877       
6880.679250101603
+NULL   3072    9318.4351351351 -4298.1513513514        5018444.1081079808      
1633.60810810806667     5695.483082135364       5696.4103077145055      3072    
11160.715384615385      -5147.907692307693      6010604.3076923073536   
1956.576923076922966667 6821.495748565159       6822.606289190924
+-3728  6       5831542.269248378       -3367.6517567568        
5817556.0411483778      969592.67352472963333   2174330.2092403853      
2381859.406131774       6       6984454.211097692       -4033.445769230769      
6967702.8672438458471   1161283.811207307641183333      2604201.2704476737      
2852759.5602156054
+-563   2       -515.6210729730 -3367.6517567568        -3883.2728297298        
-1941.6364148649        1426.0153418918999      2016.6902366556308      2       
-617.5607769230769      -4033.445769230769      -4651.0065461538459     
-2325.50327307692295    1707.9424961538462      2415.395441814127
+762    2       5831542.269248378       1531.2194054054 5833073.4886537834      
2916536.7443268917      2915005.5249214866      4122440.3477364695      2       
6984454.211097692       1833.9456923076925      6986288.1567899996925   
3493144.07839499984625  3491310.1327026924      4937458.140118758
+6981   3       5831542.269248378       -515.6210729730 5830511.0271024320      
1943503.67570081066667  2749258.455012492       3367140.1929065133      3       
6984454.211097692       -617.5607769230769      6983219.0895438458462   
2327739.696514615282066667      3292794.4113115156      4032833.0678006653
+253665376      1024    9767.0054054054 -9779.5486486487        
-347484.0818378374      -339.33992366976309     5708.9563478862 
5711.745967572779       1024    11697.969230769231      -11712.99230769231      
-416182.64030769233089  -406.428359675480791885 6837.632716002934       
6840.973851172274
+528534767      1024    5831542.269248378       -9777.1594594595        
11646372.8607481068     11373.41099682432305    257528.92988206653      
257654.7686043977       1024    6984454.211097692       -11710.130769230771     
13948892.79980307629003 13621.965624807691689482        308443.1074570801       
308593.82484083984
+626923679      1024    9723.4027027027 -9778.9513513514        
10541.0525297287        10.29399661106318       5742.09145323734        
5744.897264034267       1024    11645.746153846154      -11712.276923076923     
12625.04759999997746    12.329148046874977988   6877.318722794877       
6880.679250101603

Modified: 
hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_cast.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_cast.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_cast.q.out
 (original)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_cast.q.out
 Fri Nov  7 20:41:34 2014
@@ -46,13 +46,13 @@ POSTHOOK: query: SELECT cdouble, cint, c
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@alltypesorc
 #### A masked pattern was here ####
--13326.0       528534767       true    1969-12-31 15:59:46.674 
-13326.0000000000       528534767.00000000000000        1.00    -13
--15813.0       528534767       true    1969-12-31 15:59:55.787 
-15813.0000000000       528534767.00000000000000        1.00    -4
--9566.0        528534767       true    1969-12-31 15:59:44.187 
-9566.0000000000        528534767.00000000000000        1.00    -16
-15007.0        528534767       true    1969-12-31 15:59:50.434 
15007.0000000000        528534767.00000000000000        1.00    -10
-7021.0 528534767       true    1969-12-31 16:00:15.007 7021.0000000000 
528534767.00000000000000        1.00    15
-4963.0 528534767       true    1969-12-31 16:00:07.021 4963.0000000000 
528534767.00000000000000        1.00    7
--7824.0        528534767       true    1969-12-31 16:00:04.963 
-7824.0000000000        528534767.00000000000000        1.00    5
--15431.0       528534767       true    1969-12-31 15:59:52.176 
-15431.0000000000       528534767.00000000000000        1.00    -8
--15549.0       528534767       true    1969-12-31 15:59:44.569 
-15549.0000000000       528534767.00000000000000        1.00    -15
-5780.0 528534767       true    1969-12-31 15:59:44.451 5780.0000000000 
528534767.00000000000000        1.00    -16
+-13326.0       528534767       true    1969-12-31 15:59:46.674 -13326.0        
528534767       1       -13
+-15813.0       528534767       true    1969-12-31 15:59:55.787 -15813.0        
528534767       1       -4
+-9566.0        528534767       true    1969-12-31 15:59:44.187 -9566.0 
528534767       1       -16
+15007.0        528534767       true    1969-12-31 15:59:50.434 15007.0 
528534767       1       -10
+7021.0 528534767       true    1969-12-31 16:00:15.007 7021.0  528534767       
1       15
+4963.0 528534767       true    1969-12-31 16:00:07.021 4963.0  528534767       
1       7
+-7824.0        528534767       true    1969-12-31 16:00:04.963 -7824.0 
528534767       1       5
+-15431.0       528534767       true    1969-12-31 15:59:52.176 -15431.0        
528534767       1       -8
+-15549.0       528534767       true    1969-12-31 15:59:44.569 -15549.0        
528534767       1       -15
+5780.0 528534767       true    1969-12-31 15:59:44.451 5780.0  528534767       
1       -16

Modified: 
hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_expressions.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_expressions.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_expressions.q.out
 (original)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_expressions.q.out
 Fri Nov  7 20:41:34 2014
@@ -56,13 +56,13 @@ POSTHOOK: query: SELECT cdecimal1 + cdec
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@decimal_test
 #### A masked pattern was here ####
-19699.41746361742300   -12507.91330561334600   0.8351496686995997      
2.8303425077026896E7    3.6405405405    8963    10735   -17     8963    true    
10735.776923076923      8963.641        10735.776923076923      1969-12-31 
18:29:23.64054054
-9216.33970893968500    -5851.80644490647000    0.8353975893550668      
6195112.1797296945      3.6243243243    4193    5022    -98     4193    true    
5022.715384615385       4193.6245       5022.715384615385       1969-12-31 
17:09:53.624324324
-6514.84033264034640    -4136.52120582119280    0.8355907765708067      
3095563.9418919063      4.3864864865    2964    3550    -34     2964    true    
3550.4538461538464      2964.3865       3550.4538461538464      1969-12-31 
16:49:24.386486486
-7587.30145530147700    -4817.46777546775400    0.8354976172734904      
4198623.24324327        2.3783783784    3452    4134    38      3452    true    
4134.923076923077       3452.3784       4134.923076923077       1969-12-31 
16:57:32.378378378
-19197.97297297300000   -12189.52702702700000   0.835155361813429       
2.6880848817567654E7    5.4729729730    8735    10462   -34     8735    true    
10462.5 8735.473        10462.5 1969-12-31 18:25:35.472972973
-17098.99459459460000   -10856.80540540540000   0.8351828165813104      
2.132423090270272E7     0.3945945946    7780    9318    102     7780    true    
9318.6  7780.3945       9318.6  1969-12-31 18:09:40.394594594
-12433.72307692307700   -7894.64615384615400    0.8352770361086894      
1.12754688E7    7.6000000000    5657    6776    120     5657    true    
6776.123076923077       5657.6  6776.123076923077       1969-12-31 17:34:17.6
-7247.31683991686200    -4601.59854469852400    0.8355241651897876      
3830775.6932432684      7.6783783784    3297    3949    109     3297    true    
3949.638461538462       3297.6785       3949.638461538462       1969-12-31 
16:54:57.678378378
-14757.17006237004650   -9369.89147609149300    0.8352226654922171      
1.5883214124324286E7    4.8162162162    6714    8042    106     6714    true    
8042.3538461538465      6714.8164       8042.3538461538465      1969-12-31 
17:51:54.816216216
-10964.83201663199300   -6961.99106029108600    0.8353232978714221      
8768719.779729689       9.2243243243    4989    5975    87      4989    true    
5975.607692307693       4989.224        5975.607692307693       1969-12-31 
17:23:09.224324324
+19699.417463617423     -12507.913305613346     0.8351496686995997      
2.8303425077026896E7    3.6405405405    8963    10735   -17     8963    true    
10735.776923076923      8963.641        10735.776923076923      1969-12-31 
18:29:23.64054054
+9216.339708939685      -5851.806444906470      0.8353975893550668      
6195112.1797296945      3.6243243243    4193    5022    -98     4193    true    
5022.715384615385       4193.6245       5022.715384615385       1969-12-31 
17:09:53.624324324
+6514.8403326403464     -4136.5212058211928     0.8355907765708067      
3095563.9418919063      4.3864864865    2964    3550    -34     2964    true    
3550.4538461538464      2964.3865       3550.4538461538464      1969-12-31 
16:49:24.386486486
+7587.301455301477      -4817.467775467754      0.8354976172734904      
4198623.24324327        2.3783783784    3452    4134    38      3452    true    
4134.923076923077       3452.3784       4134.923076923077       1969-12-31 
16:57:32.378378378
+19197.9729729730       -12189.5270270270       0.835155361813429       
2.6880848817567654E7    5.4729729730    8735    10462   -34     8735    true    
10462.5 8735.473        10462.5 1969-12-31 18:25:35.472972973
+17098.9945945946       -10856.8054054054       0.8351828165813104      
2.132423090270272E7     0.3945945946    7780    9318    102     7780    true    
9318.6  7780.3945       9318.6  1969-12-31 18:09:40.394594594
+12433.723076923077     -7894.646153846154      0.8352770361086894      
1.12754688E7    7.6     5657    6776    120     5657    true    
6776.123076923077       5657.6  6776.123076923077       1969-12-31 17:34:17.6
+7247.316839916862      -4601.598544698524      0.8355241651897876      
3830775.6932432684      7.6783783784    3297    3949    109     3297    true    
3949.638461538462       3297.6785       3949.638461538462       1969-12-31 
16:54:57.678378378
+14757.1700623700465    -9369.8914760914930     0.8352226654922171      
1.5883214124324286E7    4.8162162162    6714    8042    106     6714    true    
8042.3538461538465      6714.8164       8042.3538461538465      1969-12-31 
17:51:54.816216216
+10964.832016631993     -6961.991060291086      0.8353232978714221      
8768719.779729689       9.2243243243    4989    5975    87      4989    true    
5975.607692307693       4989.224        5975.607692307693       1969-12-31 
17:23:09.224324324

Modified: 
hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_math_funcs.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_math_funcs.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_math_funcs.q.out
 (original)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/vector_decimal_math_funcs.q.out
 Fri Nov  7 20:41:34 2014
@@ -195,14 +195,14 @@ and sin(cdecimal1) >= -1.0
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@decimal_test
 #### A masked pattern was here ####
--119.4594594595        -119.46 -119    -120    -119    1.316485E-52    NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    119.459459459500000000  
-0.07885666683797002    NaN     0.9968859644388647      NaN     
-1.5624254815943668     -6844.522849943508      -2.0849608902209606     
-119.4594594595 119.4594594595  -1      NULL
-9318.4351351351        9318.44 9318    9318    9319    Infinity        
9.13974998962673        3.969342986470191       13.185871984999437      NULL    
13.185871984999437      173.867220004793        173.867220004793        
96.53204201266593       9318.435135135100000000 0.4540668481851705      NaN     
0.8909676185918236      NaN     1.5706890126394983      533907.0049096602       
162.63737424163023      9318.4351351351 -9318.4351351351        1       
-0.9607267417229353
-9318.4351351351        9318.44 9318    9318    9319    Infinity        
9.13974998962673        3.969342986470191       13.185871984999437      NULL    
13.185871984999437      173.867220004793        173.867220004793        
96.53204201266593       9318.435135135100000000 0.4540668481851705      NaN     
0.8909676185918236      NaN     1.5706890126394983      533907.0049096602       
162.63737424163023      9318.4351351351 -9318.4351351351        1       
-0.9607267417229353
-9318.4351351351        9318.44 9318    9318    9319    Infinity        
9.13974998962673        3.969342986470191       13.185871984999437      NULL    
13.185871984999437      173.867220004793        173.867220004793        
96.53204201266593       9318.435135135100000000 0.4540668481851705      NaN     
0.8909676185918236      NaN     1.5706890126394983      533907.0049096602       
162.63737424163023      9318.4351351351 -9318.4351351351        1       
-0.9607267417229353
--4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.151351351400000000 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL
--4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.151351351400000000 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL
--4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.151351351400000000 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL
--4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.151351351400000000 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL
--4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.151351351400000000 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL
--4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.151351351400000000 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL
--4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.151351351400000000 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL
+-119.4594594595        -119.46 -119    -120    -119    1.316485E-52    NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    119.4594594595  
-0.07885666683797002    NaN     0.9968859644388647      NaN     
-1.5624254815943668     -6844.522849943508      -2.0849608902209606     
-119.4594594595 119.4594594595  -1      NULL
+9318.4351351351        9318.44 9318    9318    9319    Infinity        
9.13974998962673        3.969342986470191       13.185871984999437      NULL    
13.185871984999437      173.867220004793        173.867220004793        
96.53204201266593       9318.4351351351 0.4540668481851705      NaN     
0.8909676185918236      NaN     1.5706890126394983      533907.0049096602       
162.63737424163023      9318.4351351351 -9318.4351351351        1       
-0.9607267417229353
+9318.4351351351        9318.44 9318    9318    9319    Infinity        
9.13974998962673        3.969342986470191       13.185871984999437      NULL    
13.185871984999437      173.867220004793        173.867220004793        
96.53204201266593       9318.4351351351 0.4540668481851705      NaN     
0.8909676185918236      NaN     1.5706890126394983      533907.0049096602       
162.63737424163023      9318.4351351351 -9318.4351351351        1       
-0.9607267417229353
+9318.4351351351        9318.44 9318    9318    9319    Infinity        
9.13974998962673        3.969342986470191       13.185871984999437      NULL    
13.185871984999437      173.867220004793        173.867220004793        
96.53204201266593       9318.4351351351 0.4540668481851705      NaN     
0.8909676185918236      NaN     1.5706890126394983      533907.0049096602       
162.63737424163023      9318.4351351351 -9318.4351351351        1       
-0.9607267417229353
+-4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.1513513514 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL
+-4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.1513513514 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL
+-4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.1513513514 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL
+-4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.1513513514 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL
+-4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.1513513514 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL
+-4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.1513513514 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL
+-4298.1513513514       -4298.15        -4298   -4299   -4298   0.0     NULL    
NULL    NULL    NULL    NULL    NULL    NULL    NULL    4298.1513513514 
-0.43730633941118113    NaN     0.899312607223313       NaN     
-1.5705636686355597     -246265.93214088667     -75.01689283012556      
-4298.1513513514        4298.1513513514 -1      NULL

Modified: 
hive/branches/spark/ql/src/test/results/clientpositive/vector_mapjoin_reduce.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/vector_mapjoin_reduce.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/vector_mapjoin_reduce.q.out
 (original)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/vector_mapjoin_reduce.q.out
 Fri Nov  7 20:41:34 2014
@@ -1,104 +1,6 @@
 PREHOOK: query: -- SORT_QUERY_RESULTS
 
-DROP TABLE part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: -- SORT_QUERY_RESULTS
-
-DROP TABLE part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@part
-PREHOOK: query: DROP TABLE lineitem
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE lineitem
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE lineitem (L_ORDERKEY      INT,
-                                L_PARTKEY       INT,
-                                L_SUPPKEY       INT,
-                                L_LINENUMBER    INT,
-                                L_QUANTITY      DOUBLE,
-                                L_EXTENDEDPRICE DOUBLE,
-                                L_DISCOUNT      DOUBLE,
-                                L_TAX           DOUBLE,
-                                L_RETURNFLAG    STRING,
-                                L_LINESTATUS    STRING,
-                                l_shipdate      STRING,
-                                L_COMMITDATE    STRING,
-                                L_RECEIPTDATE   STRING,
-                                L_SHIPINSTRUCT  STRING,
-                                L_SHIPMODE      STRING,
-                                L_COMMENT       STRING)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY '|'
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@lineitem
-POSTHOOK: query: CREATE TABLE lineitem (L_ORDERKEY      INT,
-                                L_PARTKEY       INT,
-                                L_SUPPKEY       INT,
-                                L_LINENUMBER    INT,
-                                L_QUANTITY      DOUBLE,
-                                L_EXTENDEDPRICE DOUBLE,
-                                L_DISCOUNT      DOUBLE,
-                                L_TAX           DOUBLE,
-                                L_RETURNFLAG    STRING,
-                                L_LINESTATUS    STRING,
-                                l_shipdate      STRING,
-                                L_COMMITDATE    STRING,
-                                L_RECEIPTDATE   STRING,
-                                L_SHIPINSTRUCT  STRING,
-                                L_SHIPMODE      STRING,
-                                L_COMMENT       STRING)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY '|'
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@lineitem
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' 
OVERWRITE INTO TABLE lineitem
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@lineitem
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' 
OVERWRITE INTO TABLE lineitem
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@lineitem
-PREHOOK: query: -- Verify HIVE-8097 with a query that has a Vectorized MapJoin 
in the Reducer.
+-- Verify HIVE-8097 with a query that has a Vectorized MapJoin in the Reducer.
 -- Query copied from subquery_in.q
 
 -- non agg, non corr, with join in Parent Query
@@ -108,7 +10,9 @@ from (select distinct l_partkey as p_par
 where li.l_linenumber = 1 and
  li.l_orderkey in (select l_orderkey from lineitem where l_shipmode = 'AIR')
 PREHOOK: type: QUERY
-POSTHOOK: query: -- Verify HIVE-8097 with a query that has a Vectorized 
MapJoin in the Reducer.
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+-- Verify HIVE-8097 with a query that has a Vectorized MapJoin in the Reducer.
 -- Query copied from subquery_in.q
 
 -- non agg, non corr, with join in Parent Query
@@ -130,34 +34,34 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: lineitem
-            Statistics: Num rows: 3024 Data size: 12099 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE 
Column stats: NONE
             Filter Operator
               predicate: l_partkey is not null (type: boolean)
-              Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE 
Column stats: NONE
               Select Operator
                 expressions: l_partkey (type: int)
                 outputColumnNames: l_partkey
-                Statistics: Num rows: 1512 Data size: 6049 Basic stats: 
COMPLETE Column stats: NONE
+                Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE 
Column stats: NONE
                 Group By Operator
                   keys: l_partkey (type: int)
                   mode: hash
                   outputColumnNames: _col0
-                  Statistics: Num rows: 1512 Data size: 6049 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 50 Data size: 5999 Basic stats: 
COMPLETE Column stats: NONE
                   Reduce Output Operator
                     key expressions: _col0 (type: int)
                     sort order: +
                     Map-reduce partition columns: _col0 (type: int)
-                    Statistics: Num rows: 1512 Data size: 6049 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 50 Data size: 5999 Basic stats: 
COMPLETE Column stats: NONE
       Reduce Operator Tree:
         Group By Operator
           keys: KEY._col0 (type: int)
           mode: mergepartial
           outputColumnNames: _col0
-          Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE 
Column stats: NONE
+          Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE 
Column stats: NONE
           Select Operator
             expressions: _col0 (type: int)
             outputColumnNames: _col0
-            Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE 
Column stats: NONE
             File Output Operator
               compressed: false
               table:
@@ -178,10 +82,10 @@ STAGE PLANS:
         li 
           TableScan
             alias: li
-            Statistics: Num rows: 756 Data size: 12099 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE 
Column stats: NONE
             Filter Operator
               predicate: ((l_partkey is not null and l_orderkey is not null) 
and (l_linenumber = 1)) (type: boolean)
-              Statistics: Num rows: 94 Data size: 1504 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 12 Data size: 1439 Basic stats: COMPLETE 
Column stats: NONE
               HashTable Sink Operator
                 condition expressions:
                   0 {_col0}
@@ -192,19 +96,19 @@ STAGE PLANS:
         sq_1:lineitem 
           TableScan
             alias: lineitem
-            Statistics: Num rows: 1728 Data size: 12099 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE 
Column stats: NONE
             Filter Operator
               predicate: ((l_shipmode = 'AIR') and l_orderkey is not null) 
(type: boolean)
-              Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE 
Column stats: NONE
               Select Operator
                 expressions: l_orderkey (type: int)
                 outputColumnNames: _col0
-                Statistics: Num rows: 432 Data size: 3024 Basic stats: 
COMPLETE Column stats: NONE
+                Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE 
Column stats: NONE
                 Group By Operator
                   keys: _col0 (type: int)
                   mode: hash
                   outputColumnNames: _col0
-                  Statistics: Num rows: 432 Data size: 3024 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 25 Data size: 2999 Basic stats: 
COMPLETE Column stats: NONE
                   HashTable Sink Operator
                     condition expressions:
                       0 {_col0} {_col3}
@@ -227,7 +131,7 @@ STAGE PLANS:
                 0 _col0 (type: int)
                 1 l_partkey (type: int)
               outputColumnNames: _col0, _col1, _col3
-              Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE 
Column stats: NONE
               Map Join Operator
                 condition map:
                      Left Semi Join 0 to 1
@@ -238,14 +142,14 @@ STAGE PLANS:
                   0 _col1 (type: int)
                   1 _col0 (type: int)
                 outputColumnNames: _col0, _col3
-                Statistics: Num rows: 914 Data size: 3658 Basic stats: 
COMPLETE Column stats: NONE
+                Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: _col0 (type: int), _col3 (type: int)
                   outputColumnNames: _col0, _col1
-                  Statistics: Num rows: 914 Data size: 3658 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 29 Data size: 3627 Basic stats: 
COMPLETE Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 914 Data size: 3658 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 29 Data size: 3627 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: org.apache.hadoop.mapred.TextInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -309,34 +213,34 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: lineitem
-            Statistics: Num rows: 3024 Data size: 12099 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE 
Column stats: NONE
             Filter Operator
               predicate: l_partkey is not null (type: boolean)
-              Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE 
Column stats: NONE
               Select Operator
                 expressions: l_partkey (type: int)
                 outputColumnNames: l_partkey
-                Statistics: Num rows: 1512 Data size: 6049 Basic stats: 
COMPLETE Column stats: NONE
+                Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE 
Column stats: NONE
                 Group By Operator
                   keys: l_partkey (type: int)
                   mode: hash
                   outputColumnNames: _col0
-                  Statistics: Num rows: 1512 Data size: 6049 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 50 Data size: 5999 Basic stats: 
COMPLETE Column stats: NONE
                   Reduce Output Operator
                     key expressions: _col0 (type: int)
                     sort order: +
                     Map-reduce partition columns: _col0 (type: int)
-                    Statistics: Num rows: 1512 Data size: 6049 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 50 Data size: 5999 Basic stats: 
COMPLETE Column stats: NONE
       Reduce Operator Tree:
         Group By Operator
           keys: KEY._col0 (type: int)
           mode: mergepartial
           outputColumnNames: _col0
-          Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE 
Column stats: NONE
+          Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE 
Column stats: NONE
           Select Operator
             expressions: _col0 (type: int)
             outputColumnNames: _col0
-            Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE 
Column stats: NONE
             File Output Operator
               compressed: false
               table:
@@ -357,10 +261,10 @@ STAGE PLANS:
         li 
           TableScan
             alias: li
-            Statistics: Num rows: 756 Data size: 12099 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE 
Column stats: NONE
             Filter Operator
               predicate: (((l_partkey is not null and l_orderkey is not null) 
and l_linenumber is not null) and (l_linenumber = 1)) (type: boolean)
-              Statistics: Num rows: 47 Data size: 752 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE 
Column stats: NONE
               HashTable Sink Operator
                 condition expressions:
                   0 {_col0}
@@ -371,19 +275,19 @@ STAGE PLANS:
         sq_1:lineitem 
           TableScan
             alias: lineitem
-            Statistics: Num rows: 1099 Data size: 12099 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE 
Column stats: NONE
             Filter Operator
-              predicate: (((l_shipmode = 'AIR') and l_orderkey is not null) 
and l_linenumber is not null) (type: boolean)
-              Statistics: Num rows: 138 Data size: 1519 Basic stats: COMPLETE 
Column stats: NONE
+              predicate: ((((l_shipmode = 'AIR') and l_orderkey is not null) 
and l_linenumber is not null) and (l_linenumber = 1)) (type: boolean)
+              Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE 
Column stats: NONE
               Select Operator
-                expressions: l_orderkey (type: int), l_linenumber (type: int)
+                expressions: l_orderkey (type: int), 1 (type: int)
                 outputColumnNames: _col0, _col1
-                Statistics: Num rows: 138 Data size: 1519 Basic stats: 
COMPLETE Column stats: NONE
+                Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE 
Column stats: NONE
                 Group By Operator
                   keys: _col0 (type: int), _col1 (type: int)
                   mode: hash
                   outputColumnNames: _col0, _col1
-                  Statistics: Num rows: 138 Data size: 1519 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE 
Column stats: NONE
                   HashTable Sink Operator
                     condition expressions:
                       0 {_col0} {_col3}
@@ -406,7 +310,7 @@ STAGE PLANS:
                 0 _col0 (type: int)
                 1 l_partkey (type: int)
               outputColumnNames: _col0, _col1, _col3
-              Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE 
Column stats: NONE
               Map Join Operator
                 condition map:
                      Left Semi Join 0 to 1
@@ -417,14 +321,14 @@ STAGE PLANS:
                   0 _col1 (type: int), 1 (type: int)
                   1 _col0 (type: int), _col1 (type: int)
                 outputColumnNames: _col0, _col3
-                Statistics: Num rows: 914 Data size: 3658 Basic stats: 
COMPLETE Column stats: NONE
+                Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: _col0 (type: int), _col3 (type: int)
                   outputColumnNames: _col0, _col1
-                  Statistics: Num rows: 914 Data size: 3658 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 29 Data size: 3627 Basic stats: 
COMPLETE Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 914 Data size: 3658 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 29 Data size: 3627 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: org.apache.hadoop.mapred.TextInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

Modified: 
hive/branches/spark/ql/src/test/results/clientpositive/vectorized_ptf.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/vectorized_ptf.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
Files 
hive/branches/spark/ql/src/test/results/clientpositive/vectorized_ptf.q.out 
(original) and 
hive/branches/spark/ql/src/test/results/clientpositive/vectorized_ptf.q.out Fri 
Nov  7 20:41:34 2014 differ

Modified: hive/branches/spark/ql/src/test/results/clientpositive/windowing.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/windowing.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/windowing.q.out 
(original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/windowing.q.out Fri 
Nov  7 20:41:34 2014
@@ -1,45 +1,3 @@
-PREHOOK: query: DROP TABLE part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@part
 PREHOOK: query: -- 1. testWindowing
 select p_mfgr, p_name, p_size,
 rank() over(distribute by p_mfgr sort by p_name) as r,

Modified: 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_adjust_rowcontainer_sz.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/windowing_adjust_rowcontainer_sz.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_adjust_rowcontainer_sz.q.out
 (original)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_adjust_rowcontainer_sz.q.out
 Fri Nov  7 20:41:34 2014
@@ -1,45 +1,3 @@
-PREHOOK: query: DROP TABLE part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@part
 PREHOOK: query: select p_mfgr, p_name, p_size,
 rank() over(distribute by p_mfgr sort by p_name) as r,
 dense_rank() over(distribute by p_mfgr sort by p_name) as dr,

Modified: 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_columnPruning.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/windowing_columnPruning.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_columnPruning.q.out
 (original)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_columnPruning.q.out
 Fri Nov  7 20:41:34 2014
@@ -1,45 +1,3 @@
-PREHOOK: query: DROP TABLE part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@part
 PREHOOK: query: -- 1. testQueryLevelPartitionColsNotInSelect
 select p_size,
 sum(p_retailprice) over (distribute by p_mfgr sort by p_name rows between 
unbounded preceding and current row) as s1

Modified: 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_decimal.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/windowing_decimal.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_decimal.q.out 
(original)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_decimal.q.out 
Fri Nov  7 20:41:34 2014
@@ -1,45 +1,3 @@
-PREHOOK: query: DROP TABLE part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@part
 PREHOOK: query: create table part_dec(
     p_partkey INT,
     p_name STRING,

Modified: 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_expressions.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/windowing_expressions.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_expressions.q.out
 (original)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_expressions.q.out
 Fri Nov  7 20:41:34 2014
@@ -1,45 +1,3 @@
-PREHOOK: query: DROP TABLE part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@part
 PREHOOK: query: drop table over10k
 PREHOOK: type: DROPTABLE
 POSTHOOK: query: drop table over10k

Modified: 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_streaming.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/windowing_streaming.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_streaming.q.out
 (original)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/windowing_streaming.q.out
 Fri Nov  7 20:41:34 2014
@@ -1,45 +1,3 @@
-PREHOOK: query: DROP TABLE if exists part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE if exists part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part( 
-    p_partkey INT,
-    p_name STRING,
-    p_mfgr STRING,
-    p_brand STRING,
-    p_type STRING,
-    p_size INT,
-    p_container STRING,
-    p_retailprice DOUBLE,
-    p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' 
overwrite into table part
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@part
 PREHOOK: query: drop table over10k
 PREHOOK: type: DROPTABLE
 POSTHOOK: query: drop table over10k
@@ -106,25 +64,25 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: part
-            Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE 
Column stats: NONE
             Reduce Output Operator
               key expressions: p_mfgr (type: string), p_name (type: string)
               sort order: ++
               Map-reduce partition columns: p_mfgr (type: string)
-              Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE 
Column stats: NONE
               value expressions: p_name (type: string), p_mfgr (type: string)
       Reduce Operator Tree:
         Extract
-          Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE 
Column stats: NONE
+          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE 
Column stats: NONE
           PTF Operator
-            Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE 
Column stats: NONE
             Select Operator
               expressions: _col2 (type: string), _wcol0 (type: int)
               outputColumnNames: _col0, _col1
-              Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE 
Column stats: NONE
               File Output Operator
                 compressed: false
-                Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE 
Column stats: NONE
                 table:
                     input format: org.apache.hadoop.mapred.TextInputFormat
                     output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -156,29 +114,29 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: part
-            Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE 
Column stats: NONE
             Reduce Output Operator
               key expressions: p_mfgr (type: string), p_name (type: string)
               sort order: ++
               Map-reduce partition columns: p_mfgr (type: string)
-              Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE 
Column stats: NONE
               TopN Hash Memory Usage: 0.8
               value expressions: p_name (type: string), p_mfgr (type: string)
       Reduce Operator Tree:
         Extract
-          Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE 
Column stats: NONE
+          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE 
Column stats: NONE
           PTF Operator
-            Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE 
Column stats: NONE
             Filter Operator
               predicate: (_wcol0 < 4) (type: boolean)
-              Statistics: Num rows: 5 Data size: 1057 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 8 Data size: 968 Basic stats: COMPLETE 
Column stats: NONE
               Select Operator
                 expressions: _col2 (type: string), _wcol0 (type: int)
                 outputColumnNames: _col0, _col1
-                Statistics: Num rows: 5 Data size: 1057 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 8 Data size: 968 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  Statistics: Num rows: 5 Data size: 1057 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 8 Data size: 968 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: org.apache.hadoop.mapred.TextInputFormat
                       output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

Modified: 
hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerDe.java
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerDe.java?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerDe.java
 (original)
+++ 
hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerDe.java
 Fri Nov  7 20:41:34 2014
@@ -92,10 +92,10 @@ public class AvroSerDe extends AbstractS
 
     final String columnNameProperty = 
properties.getProperty(serdeConstants.LIST_COLUMNS);
     final String columnTypeProperty = 
properties.getProperty(serdeConstants.LIST_COLUMN_TYPES);
-    final String columnCommentProperty = 
properties.getProperty(LIST_COLUMN_COMMENTS);
+    final String columnCommentProperty = 
properties.getProperty(LIST_COLUMN_COMMENTS,"");
 
-    if (properties.getProperty(AvroSerdeUtils.SCHEMA_LITERAL) != null
-        || properties.getProperty(AvroSerdeUtils.SCHEMA_URL) != null
+    if 
(properties.getProperty(AvroSerdeUtils.AvroTableProperties.SCHEMA_LITERAL.getPropName())
 != null
+        || 
properties.getProperty(AvroSerdeUtils.AvroTableProperties.SCHEMA_URL.getPropName())
 != null
         || columnNameProperty == null || columnNameProperty.isEmpty()
         || columnTypeProperty == null || columnTypeProperty.isEmpty()) {
       schema = determineSchemaOrReturnErrorSchema(properties);
@@ -104,28 +104,8 @@ public class AvroSerDe extends AbstractS
       columnNames = Arrays.asList(columnNameProperty.split(","));
       columnTypes = 
TypeInfoUtils.getTypeInfosFromTypeString(columnTypeProperty);
 
-      List<String> columnComments;
-      if (columnCommentProperty.isEmpty()) {
-        columnComments = new ArrayList<String>();
-      } else {
-        columnComments = Arrays.asList(columnCommentProperty.split(","));
-        LOG.info("columnComments is " + columnCommentProperty);
-      }
-      if (columnNames.size() != columnTypes.size()) {
-        throw new IllegalArgumentException("AvroSerde initialization failed. 
Number of column " +
-            "name and column type differs. columnNames = " + columnNames + ", 
columnTypes = " +
-            columnTypes);
-      }
-
-      final String tableName = properties.getProperty(TABLE_NAME);
-      final String tableComment = properties.getProperty(TABLE_COMMENT);
-      TypeInfoToSchema typeInfoToSchema = new TypeInfoToSchema();
-      schema = typeInfoToSchema.convert(columnNames, columnTypes, 
columnComments,
-          properties.getProperty(AvroSerdeUtils.SCHEMA_NAMESPACE),
-          properties.getProperty(AvroSerdeUtils.SCHEMA_NAME, tableName),
-          properties.getProperty(AvroSerdeUtils.SCHEMA_DOC, tableComment));
-
-      properties.setProperty(AvroSerdeUtils.SCHEMA_LITERAL, schema.toString());
+      schema = getSchemaFromCols(properties, columnNames, columnTypes, 
columnCommentProperty);
+      
properties.setProperty(AvroSerdeUtils.AvroTableProperties.SCHEMA_LITERAL.getPropName(),
 schema.toString());
     }
 
     LOG.info("Avro schema is " + schema);
@@ -133,7 +113,8 @@ public class AvroSerDe extends AbstractS
     if (configuration == null) {
       LOG.info("Configuration null, not inserting schema");
     } else {
-      configuration.set(AvroSerdeUtils.AVRO_SERDE_SCHEMA, 
schema.toString(false));
+      configuration.set(
+          AvroSerdeUtils.AvroTableProperties.AVRO_SERDE_SCHEMA.getPropName(), 
schema.toString(false));
     }
 
     badSchema = schema.equals(SchemaResolutionProblem.SIGNAL_BAD_SCHEMA);
@@ -144,6 +125,31 @@ public class AvroSerDe extends AbstractS
     this.oi = aoig.getObjectInspector();
   }
 
+  public static Schema getSchemaFromCols(Properties properties,
+          List<String> columnNames, List<TypeInfo> columnTypes, String 
columnCommentProperty) {
+    List<String> columnComments;
+    if (columnCommentProperty == null || columnCommentProperty.isEmpty()) {
+      columnComments = new ArrayList<String>();
+    } else {
+      columnComments = Arrays.asList(columnCommentProperty.split(","));
+      LOG.info("columnComments is " + columnCommentProperty);
+    }
+    if (columnNames.size() != columnTypes.size()) {
+      throw new IllegalArgumentException("AvroSerde initialization failed. 
Number of column " +
+          "name and column type differs. columnNames = " + columnNames + ", 
columnTypes = " +
+          columnTypes);
+    }
+
+    final String tableName = properties.getProperty(TABLE_NAME);
+    final String tableComment = properties.getProperty(TABLE_COMMENT);
+    TypeInfoToSchema typeInfoToSchema = new TypeInfoToSchema();
+    return typeInfoToSchema.convert(columnNames, columnTypes, columnComments,
+        
properties.getProperty(AvroSerdeUtils.AvroTableProperties.SCHEMA_NAMESPACE.getPropName()),
+        
properties.getProperty(AvroSerdeUtils.AvroTableProperties.SCHEMA_NAME.getPropName(),
 tableName),
+        
properties.getProperty(AvroSerdeUtils.AvroTableProperties.SCHEMA_DOC.getPropName(),
 tableComment));
+
+  }
+
   /**
    * Attempt to determine the schema via the usual means, but do not throw
    * an exception if we fail.  Instead, signal failure via a special

Modified: 
hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerdeUtils.java
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerdeUtils.java?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerdeUtils.java
 (original)
+++ 
hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerdeUtils.java
 Fri Nov  7 20:41:34 2014
@@ -48,16 +48,47 @@ import java.util.Properties;
 public class AvroSerdeUtils {
   private static final Log LOG = LogFactory.getLog(AvroSerdeUtils.class);
 
-  public static final String SCHEMA_LITERAL = "avro.schema.literal";
-  public static final String SCHEMA_URL = "avro.schema.url";
+  /**
+   * Enum container for all avro table properties.
+   * If introducing a new avro-specific table property,
+   * add it here. Putting them in an enum rather than separate strings
+   * allows them to be programmatically grouped and referenced together.
+   */
+  public static enum AvroTableProperties {
+    SCHEMA_LITERAL("avro.schema.literal"),
+    SCHEMA_URL("avro.schema.url"),
+    SCHEMA_NAMESPACE("avro.schema.namespace"),
+    SCHEMA_NAME("avro.schema.name"),
+    SCHEMA_DOC("avro.schema.doc"),
+    AVRO_SERDE_SCHEMA("avro.serde.schema"),
+    SCHEMA_RETRIEVER("avro.schema.retriever");
+
+    private final String propName;
+
+    AvroTableProperties(String propName) {
+      this.propName = propName;
+    }
+
+    public String getPropName(){
+      return this.propName;
+    }
+  }
+
+  // Following parameters slated for removal, prefer usage of enum above, that 
allows programmatic access.
+  @Deprecated public static final String SCHEMA_LITERAL = 
"avro.schema.literal";
+  @Deprecated public static final String SCHEMA_URL = "avro.schema.url";
+  @Deprecated public static final String SCHEMA_NAMESPACE = 
"avro.schema.namespace";
+  @Deprecated public static final String SCHEMA_NAME = "avro.schema.name";
+  @Deprecated public static final String SCHEMA_DOC = "avro.schema.doc";
+  @Deprecated public static final String AVRO_SERDE_SCHEMA = 
AvroTableProperties.AVRO_SERDE_SCHEMA.getPropName();
+  @Deprecated public static final String SCHEMA_RETRIEVER = 
AvroTableProperties.SCHEMA_RETRIEVER.getPropName();
+
   public static final String SCHEMA_NONE = "none";
-  public static final String SCHEMA_NAMESPACE = "avro.schema.namespace";
-  public static final String SCHEMA_NAME = "avro.schema.name";
-  public static final String SCHEMA_DOC = "avro.schema.doc";
-  public static final String EXCEPTION_MESSAGE = "Neither " + SCHEMA_LITERAL + 
" nor "
-          + SCHEMA_URL + " specified, can't determine table schema";
-  public static final String AVRO_SERDE_SCHEMA = "avro.serde.schema";
-  public static final String SCHEMA_RETRIEVER = "avro.schema.retriever";
+  public static final String EXCEPTION_MESSAGE = "Neither "
+      + AvroTableProperties.SCHEMA_LITERAL.getPropName() + " nor "
+      + AvroTableProperties.SCHEMA_URL.getPropName() + " specified, can't 
determine table schema";
+
+
 
   /**
    * Determine the schema to that's been provided for Avro serde work.
@@ -68,12 +99,12 @@ public class AvroSerdeUtils {
    */
   public static Schema determineSchemaOrThrowException(Properties properties)
           throws IOException, AvroSerdeException {
-    String schemaString = properties.getProperty(SCHEMA_LITERAL);
+    String schemaString = 
properties.getProperty(AvroTableProperties.SCHEMA_LITERAL.getPropName());
     if(schemaString != null && !schemaString.equals(SCHEMA_NONE))
       return AvroSerdeUtils.getSchemaFor(schemaString);
 
     // Try pulling directly from URL
-    schemaString = properties.getProperty(SCHEMA_URL);
+    schemaString = 
properties.getProperty(AvroTableProperties.SCHEMA_URL.getPropName());
     if(schemaString == null || schemaString.equals(SCHEMA_NONE))
       throw new AvroSerdeException(EXCEPTION_MESSAGE);
 

Modified: 
hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/io/HiveDecimalWritable.java
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/io/HiveDecimalWritable.java?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/io/HiveDecimalWritable.java
 (original)
+++ 
hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/io/HiveDecimalWritable.java
 Fri Nov  7 20:41:34 2014
@@ -62,6 +62,10 @@ public class HiveDecimalWritable impleme
     set(value.unscaledValue().toByteArray(), value.scale());
   }
 
+  public void set(HiveDecimal value, int maxPrecision, int maxScale) {
+    set(HiveDecimal.enforcePrecisionScale(value, maxPrecision, maxScale));
+  }
+
   public void set(HiveDecimalWritable writable) {
     set(writable.getHiveDecimal());
   }

Modified: 
hive/branches/spark/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestTypeInfoToSchema.java
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestTypeInfoToSchema.java?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestTypeInfoToSchema.java
 (original)
+++ 
hive/branches/spark/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestTypeInfoToSchema.java
 Fri Nov  7 20:41:34 2014
@@ -80,6 +80,7 @@ public class TestTypeInfoToSchema {
 
   private TypeInfoToSchema typeInfoToSchema;
 
+  private final String lineSeparator = System.getProperty("line.separator");
 
   private String getAvroSchemaString(TypeInfo columnType) {
     return typeInfoToSchema.convert(
@@ -383,7 +384,7 @@ public class TestTypeInfoToSchema {
     LOGGER.info("structTypeInfo is " + structTypeInfo);
 
     final String specificSchema = 
IOUtils.toString(Resources.getResource("avro-struct.avsc")
-        .openStream()).replace("\n", "");
+        .openStream()).replace(lineSeparator, "");
     String expectedSchema = genSchema(
         specificSchema);
 
@@ -414,7 +415,7 @@ public class TestTypeInfoToSchema {
     superStructTypeInfo.setAllStructFieldTypeInfos(superTypeInfos);
 
     final String specificSchema = 
IOUtils.toString(Resources.getResource("avro-nested-struct.avsc")
-        .openStream()).replace("\n", "");
+        .openStream()).replace(lineSeparator, "");
     String expectedSchema = genSchema(
         specificSchema);
     Assert.assertEquals("Test for nested struct's avro schema failed",

Modified: hive/branches/spark/service/pom.xml
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/service/pom.xml?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- hive/branches/spark/service/pom.xml (original)
+++ hive/branches/spark/service/pom.xml Fri Nov  7 20:41:34 2014
@@ -86,6 +86,11 @@
       <artifactId>libthrift</artifactId>
       <version>${libthrift.version}</version>
     </dependency>
+    <dependency>
+      <groupId>org.apache.curator</groupId>
+      <artifactId>curator-framework</artifactId>
+      <version>${curator.version}</version>
+    </dependency>
     <!-- intra-project -->
     <dependency>
       <groupId>org.apache.hive</groupId>

Modified: 
hive/branches/spark/service/src/java/org/apache/hive/service/auth/HiveAuthFactory.java
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/service/src/java/org/apache/hive/service/auth/HiveAuthFactory.java?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- 
hive/branches/spark/service/src/java/org/apache/hive/service/auth/HiveAuthFactory.java
 (original)
+++ 
hive/branches/spark/service/src/java/org/apache/hive/service/auth/HiveAuthFactory.java
 Fri Nov  7 20:41:34 2014
@@ -21,9 +21,13 @@ import java.io.IOException;
 import java.net.InetAddress;
 import java.net.InetSocketAddress;
 import java.net.UnknownHostException;
+import java.util.ArrayList;
+import java.util.Arrays;
 import java.util.HashMap;
+import java.util.List;
 import java.util.Map;
 
+import javax.net.ssl.SSLServerSocket;
 import javax.security.auth.login.LoginException;
 import javax.security.sasl.Sasl;
 
@@ -43,12 +47,16 @@ import org.apache.thrift.transport.TSock
 import org.apache.thrift.transport.TTransport;
 import org.apache.thrift.transport.TTransportException;
 import org.apache.thrift.transport.TTransportFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 
 /**
  * This class helps in some aspects of authentication. It creates the proper 
Thrift classes for the
  * given configuration as well as helps with authenticating requests.
  */
 public class HiveAuthFactory {
+  private static final Logger LOG = 
LoggerFactory.getLogger(HiveAuthFactory.class);
+
 
   public enum AuthTypes {
     NOSASL("NOSASL"),
@@ -218,7 +226,8 @@ public class HiveAuthFactory {
   }
 
   public static TServerSocket getServerSSLSocket(String hiveHost, int portNum, 
String keyStorePath,
-    String keyStorePassWord) throws TTransportException, UnknownHostException {
+    String keyStorePassWord,  List<String> sslVersionBlacklist)
+      throws TTransportException, UnknownHostException {
     TSSLTransportFactory.TSSLTransportParameters params =
       new TSSLTransportFactory.TSSLTransportParameters();
     params.setKeyStore(keyStorePath, keyStorePassWord);
@@ -229,7 +238,25 @@ public class HiveAuthFactory {
     } else {
       serverAddress = InetAddress.getByName(hiveHost);
     }
-    return TSSLTransportFactory.getServerSocket(portNum, 0, serverAddress, 
params);
+    TServerSocket thriftServerSocket = 
TSSLTransportFactory.getServerSocket(portNum, 0, serverAddress, params);
+    if (thriftServerSocket.getServerSocket() instanceof SSLServerSocket) {
+      List<String> sslVersionBlacklistLocal = new ArrayList<String>();
+      for (String sslVersion : sslVersionBlacklist) {
+        sslVersionBlacklistLocal.add(sslVersion.trim().toLowerCase());
+      }
+      SSLServerSocket sslServerSocket = 
(SSLServerSocket)thriftServerSocket.getServerSocket();
+      List<String> enabledProtocols = new ArrayList<String>();
+      for (String protocol : sslServerSocket.getEnabledProtocols()) {
+        if (sslVersionBlacklistLocal.contains(protocol.toLowerCase())) {
+          LOG.debug("Disabling SSL Protocol: " + protocol);
+        } else {
+          enabledProtocols.add(protocol);
+        }
+      }
+      sslServerSocket.setEnabledProtocols(enabledProtocols.toArray(new 
String[0]));
+      LOG.info("SSL Server Socket Enabled Protocols: " + 
Arrays.toString(sslServerSocket.getEnabledProtocols()));
+    }
+    return thriftServerSocket;
   }
 
   // retrieve delegation token for the given user


Reply via email to