http://git-wip-us.apache.org/repos/asf/hive/blob/dd512593/ql/src/test/results/clientpositive/llap/vector_decimal_aggregate.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_decimal_aggregate.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_aggregate.q.out
index 6cd1e8d..30a6770 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_aggregate.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_aggregate.q.out
@@ -111,8 +111,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false
@@ -291,8 +291,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false
@@ -465,7 +465,7 @@ STAGE PLANS:
                   Statistics: Num rows: 12289 Data size: 2662128 Basic stats: 
COMPLETE Column stats: NONE
                   TableScan Vectorization:
                       native: true
-                      vectorizationSchemaColumns: [0:cdouble:double, 
1:cdecimal1:decimal(11,5), 2:cdecimal2:decimal(16,0), 3:cint:int, 
4:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
+                      vectorizationSchemaColumns: [0:cdouble:double, 
1:cdecimal1:decimal(11,5)/DECIMAL_64, 2:cdecimal2:decimal(16,0)/DECIMAL_64, 
3:cint:int, 4:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
                   Select Operator
                     expressions: cdecimal1 (type: decimal(11,5)), cdecimal2 
(type: decimal(16,0)), cint (type: int)
                     outputColumnNames: cdecimal1, cdecimal2, cint
@@ -477,7 +477,7 @@ STAGE PLANS:
                     Group By Operator
                       aggregations: count(cdecimal1), max(cdecimal1), 
min(cdecimal1), sum(cdecimal1), count(cdecimal2), max(cdecimal2), 
min(cdecimal2), sum(cdecimal2), count()
                       Group By Vectorization:
-                          aggregators: VectorUDAFCount(col 1:decimal(11,5)) -> 
bigint, VectorUDAFMaxDecimal(col 1:decimal(11,5)) -> decimal(11,5), 
VectorUDAFMinDecimal(col 1:decimal(11,5)) -> decimal(11,5), 
VectorUDAFSumDecimal(col 1:decimal(11,5)) -> decimal(21,5), VectorUDAFCount(col 
2:decimal(16,0)) -> bigint, VectorUDAFMaxDecimal(col 2:decimal(16,0)) -> 
decimal(16,0), VectorUDAFMinDecimal(col 2:decimal(16,0)) -> decimal(16,0), 
VectorUDAFSumDecimal(col 2:decimal(16,0)) -> decimal(26,0), 
VectorUDAFCountStar(*) -> bigint
+                          aggregators: VectorUDAFCount(col 
1:decimal(11,5)/DECIMAL_64) -> bigint, VectorUDAFMaxDecimal64(col 
1:decimal(11,5)/DECIMAL_64) -> decimal(11,5)/DECIMAL_64, 
VectorUDAFMinDecimal64(col 1:decimal(11,5)/DECIMAL_64) -> 
decimal(11,5)/DECIMAL_64, VectorUDAFSumDecimal64ToDecimal(col 
1:decimal(11,5)/DECIMAL_64) -> decimal(21,5), VectorUDAFCount(col 
2:decimal(16,0)/DECIMAL_64) -> bigint, VectorUDAFMaxDecimal64(col 
2:decimal(16,0)/DECIMAL_64) -> decimal(16,0)/DECIMAL_64, 
VectorUDAFMinDecimal64(col 2:decimal(16,0)/DECIMAL_64) -> 
decimal(16,0)/DECIMAL_64, VectorUDAFSumDecimal64ToDecimal(col 
2:decimal(16,0)/DECIMAL_64) -> decimal(26,0), VectorUDAFCountStar(*) -> bigint
                           className: VectorGroupByOperator
                           groupByMode: HASH
                           keyExpressions: col 3:int
@@ -506,8 +506,7 @@ STAGE PLANS:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vector.serde.deserialize IS true
                 inputFormatFeatureSupport: [DECIMAL_64]
-                vectorizationSupportRemovedReasons: [DECIMAL_64 disabled 
because LLAP is enabled]
-                featureSupportInUse: []
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: org.apache.hadoop.mapred.TextInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false
@@ -515,7 +514,7 @@ STAGE PLANS:
                 rowBatchContext:
                     dataColumnCount: 4
                     includeColumns: [1, 2, 3]
-                    dataColumns: cdouble:double, cdecimal1:decimal(11,5), 
cdecimal2:decimal(16,0), cint:int
+                    dataColumns: cdouble:double, 
cdecimal1:decimal(11,5)/DECIMAL_64, cdecimal2:decimal(16,0)/DECIMAL_64, cint:int
                     partitionColumnCount: 0
                     scratchColumnTypeNames: []
         Reducer 2 
@@ -664,20 +663,20 @@ STAGE PLANS:
                   Statistics: Num rows: 12289 Data size: 2662128 Basic stats: 
COMPLETE Column stats: NONE
                   TableScan Vectorization:
                       native: true
-                      vectorizationSchemaColumns: [0:cdouble:double, 
1:cdecimal1:decimal(11,5), 2:cdecimal2:decimal(16,0), 3:cint:int, 
4:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
+                      vectorizationSchemaColumns: [0:cdouble:double, 
1:cdecimal1:decimal(11,5)/DECIMAL_64, 2:cdecimal2:decimal(16,0)/DECIMAL_64, 
3:cint:int, 4:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
                   Select Operator
                     expressions: cint (type: int), cdecimal1 (type: 
decimal(11,5)), cdecimal2 (type: decimal(16,0)), UDFToDouble(cdecimal1) (type: 
double), (UDFToDouble(cdecimal1) * UDFToDouble(cdecimal1)) (type: double), 
UDFToDouble(cdecimal2) (type: double), (UDFToDouble(cdecimal2) * 
UDFToDouble(cdecimal2)) (type: double)
                     outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6
                     Select Vectorization:
                         className: VectorSelectOperator
                         native: true
-                        projectedOutputColumnNums: [3, 1, 2, 5, 8, 6, 10]
-                        selectExpressions: CastDecimalToDouble(col 
1:decimal(11,5)) -> 5:double, DoubleColMultiplyDoubleColumn(col 6:double, col 
7:double)(children: CastDecimalToDouble(col 1:decimal(11,5)) -> 6:double, 
CastDecimalToDouble(col 1:decimal(11,5)) -> 7:double) -> 8:double, 
CastDecimalToDouble(col 2:decimal(16,0)) -> 6:double, 
DoubleColMultiplyDoubleColumn(col 7:double, col 9:double)(children: 
CastDecimalToDouble(col 2:decimal(16,0)) -> 7:double, CastDecimalToDouble(col 
2:decimal(16,0)) -> 9:double) -> 10:double
+                        projectedOutputColumnNums: [3, 1, 2, 6, 9, 7, 12]
+                        selectExpressions: CastDecimalToDouble(col 
5:decimal(11,5))(children: ConvertDecimal64ToDecimal(col 
1:decimal(11,5)/DECIMAL_64) -> 5:decimal(11,5)) -> 6:double, 
DoubleColMultiplyDoubleColumn(col 7:double, col 8:double)(children: 
CastDecimalToDouble(col 5:decimal(11,5))(children: 
ConvertDecimal64ToDecimal(col 1:decimal(11,5)/DECIMAL_64) -> 5:decimal(11,5)) 
-> 7:double, CastDecimalToDouble(col 5:decimal(11,5))(children: 
ConvertDecimal64ToDecimal(col 1:decimal(11,5)/DECIMAL_64) -> 5:decimal(11,5)) 
-> 8:double) -> 9:double, CastDecimalToDouble(col 10:decimal(16,0))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(16,0)/DECIMAL_64) -> 10:decimal(16,0)) 
-> 7:double, DoubleColMultiplyDoubleColumn(col 8:double, col 
11:double)(children: CastDecimalToDouble(col 10:decimal(16,0))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(16,0)/DECIMAL_64) -> 10:decimal(16,0)) 
-> 8:double, CastDecimalToDouble(col 10:decimal(16,0))(children: 
ConvertDecimal64ToDecimal(col 2:decima
 l(16,0)/DECIMAL_64) -> 10:decimal(16,0)) -> 11:double) -> 12:double
                     Statistics: Num rows: 12289 Data size: 2662128 Basic 
stats: COMPLETE Column stats: NONE
                     Group By Operator
                       aggregations: count(_col1), max(_col1), min(_col1), 
sum(_col1), sum(_col4), sum(_col3), count(_col2), max(_col2), min(_col2), 
sum(_col2), sum(_col6), sum(_col5), count()
                       Group By Vectorization:
-                          aggregators: VectorUDAFCount(col 1:decimal(11,5)) -> 
bigint, VectorUDAFMaxDecimal(col 1:decimal(11,5)) -> decimal(11,5), 
VectorUDAFMinDecimal(col 1:decimal(11,5)) -> decimal(11,5), 
VectorUDAFSumDecimal(col 1:decimal(11,5)) -> decimal(21,5), 
VectorUDAFSumDouble(col 8:double) -> double, VectorUDAFSumDouble(col 5:double) 
-> double, VectorUDAFCount(col 2:decimal(16,0)) -> bigint, 
VectorUDAFMaxDecimal(col 2:decimal(16,0)) -> decimal(16,0), 
VectorUDAFMinDecimal(col 2:decimal(16,0)) -> decimal(16,0), 
VectorUDAFSumDecimal(col 2:decimal(16,0)) -> decimal(26,0), 
VectorUDAFSumDouble(col 10:double) -> double, VectorUDAFSumDouble(col 6:double) 
-> double, VectorUDAFCountStar(*) -> bigint
+                          aggregators: VectorUDAFCount(col 
1:decimal(11,5)/DECIMAL_64) -> bigint, VectorUDAFMaxDecimal64(col 
1:decimal(11,5)/DECIMAL_64) -> decimal(11,5)/DECIMAL_64, 
VectorUDAFMinDecimal64(col 1:decimal(11,5)/DECIMAL_64) -> 
decimal(11,5)/DECIMAL_64, VectorUDAFSumDecimal64ToDecimal(col 
1:decimal(11,5)/DECIMAL_64) -> decimal(21,5), VectorUDAFSumDouble(col 9:double) 
-> double, VectorUDAFSumDouble(col 6:double) -> double, VectorUDAFCount(col 
2:decimal(16,0)/DECIMAL_64) -> bigint, VectorUDAFMaxDecimal64(col 
2:decimal(16,0)/DECIMAL_64) -> decimal(16,0)/DECIMAL_64, 
VectorUDAFMinDecimal64(col 2:decimal(16,0)/DECIMAL_64) -> 
decimal(16,0)/DECIMAL_64, VectorUDAFSumDecimal64ToDecimal(col 
2:decimal(16,0)/DECIMAL_64) -> decimal(26,0), VectorUDAFSumDouble(col 
12:double) -> double, VectorUDAFSumDouble(col 7:double) -> double, 
VectorUDAFCountStar(*) -> bigint
                           className: VectorGroupByOperator
                           groupByMode: HASH
                           keyExpressions: col 3:int
@@ -706,8 +705,7 @@ STAGE PLANS:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vector.serde.deserialize IS true
                 inputFormatFeatureSupport: [DECIMAL_64]
-                vectorizationSupportRemovedReasons: [DECIMAL_64 disabled 
because LLAP is enabled]
-                featureSupportInUse: []
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: org.apache.hadoop.mapred.TextInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false
@@ -715,9 +713,9 @@ STAGE PLANS:
                 rowBatchContext:
                     dataColumnCount: 4
                     includeColumns: [1, 2, 3]
-                    dataColumns: cdouble:double, cdecimal1:decimal(11,5), 
cdecimal2:decimal(16,0), cint:int
+                    dataColumns: cdouble:double, 
cdecimal1:decimal(11,5)/DECIMAL_64, cdecimal2:decimal(16,0)/DECIMAL_64, cint:int
                     partitionColumnCount: 0
-                    scratchColumnTypeNames: [double, double, double, double, 
double, double]
+                    scratchColumnTypeNames: [decimal(11,5), double, double, 
double, double, decimal(16,0), double, double]
         Reducer 2 
             Execution mode: vectorized, llap
             Reduce Vectorization:

http://git-wip-us.apache.org/repos/asf/hive/blob/dd512593/ql/src/test/results/clientpositive/llap/vector_decimal_cast.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_decimal_cast.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_cast.q.out
index 67630b4..2414907 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_cast.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_cast.q.out
@@ -60,8 +60,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false
@@ -184,8 +184,7 @@ STAGE PLANS:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vector.serde.deserialize IS true
                 inputFormatFeatureSupport: [DECIMAL_64]
-                vectorizationSupportRemovedReasons: [DECIMAL_64 disabled 
because LLAP is enabled]
-                featureSupportInUse: []
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: org.apache.hadoop.mapred.TextInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false

http://git-wip-us.apache.org/repos/asf/hive/blob/dd512593/ql/src/test/results/clientpositive/llap/vector_decimal_expressions.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_decimal_expressions.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_expressions.q.out
index c01637e..024ce07 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_expressions.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_expressions.q.out
@@ -94,8 +94,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -234,12 +234,12 @@ STAGE PLANS:
                   Statistics: Num rows: 12288 Data size: 2708600 Basic stats: 
COMPLETE Column stats: NONE
                   TableScan Vectorization:
                       native: true
-                      vectorizationSchemaColumns: [0:cdouble:double, 
1:cdecimal1:decimal(10,3), 2:cdecimal2:decimal(7,2), 
3:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
+                      vectorizationSchemaColumns: [0:cdouble:double, 
1:cdecimal1:decimal(10,3)/DECIMAL_64, 2:cdecimal2:decimal(7,2)/DECIMAL_64, 
3:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
                   Filter Operator
                     Filter Vectorization:
                         className: VectorFilterOperator
                         native: true
-                        predicateExpression: FilterExprAndExpr(children: 
FilterDecimalColGreaterDecimalScalar(col 1:decimal(10,3), val 0), 
FilterDecimalColLessDecimalScalar(col 1:decimal(10,3), val 12345.5678), 
FilterDecimalColNotEqualDecimalScalar(col 2:decimal(7,2), val 0), 
FilterDecimalColGreaterDecimalScalar(col 2:decimal(7,2), val 1000), 
SelectColumnIsNotNull(col 0:double))
+                        predicateExpression: FilterExprAndExpr(children: 
FilterDecimal64ColGreaterDecimal64Scalar(col 1:decimal(10,3)/DECIMAL_64, val 
0), FilterDecimalColLessDecimalScalar(col 4:decimal(10,3), val 
12345.5678)(children: ConvertDecimal64ToDecimal(col 1:decimal(10,3)/DECIMAL_64) 
-> 4:decimal(10,3)), FilterDecimal64ColNotEqualDecimal64Scalar(col 
2:decimal(7,2)/DECIMAL_64, val 0), FilterDecimal64ColGreaterDecimal64Scalar(col 
2:decimal(7,2)/DECIMAL_64, val 100000), SelectColumnIsNotNull(col 0:double))
                     predicate: ((cdecimal1 < 12345.5678) and (cdecimal1 > 0) 
and (cdecimal2 <> 0) and (cdecimal2 > 1000) and cdouble is not null) (type: 
boolean)
                     Statistics: Num rows: 455 Data size: 100294 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
@@ -248,15 +248,15 @@ STAGE PLANS:
                       Select Vectorization:
                           className: VectorSelectOperator
                           native: true
-                          projectedOutputColumnNums: [4, 6, 8, 10, 11, 12, 13, 
14, 15, 16, 17, 18, 19, 20]
-                          selectExpressions: DecimalColAddDecimalColumn(col 
1:decimal(10,3), col 2:decimal(7,2)) -> 4:decimal(11,3), 
DecimalColSubtractDecimalColumn(col 1:decimal(10,3), col 
5:decimal(9,2))(children: DecimalScalarMultiplyDecimalColumn(val 2, col 
2:decimal(7,2)) -> 5:decimal(9,2)) -> 6:decimal(11,3), 
DecimalColDivideDecimalColumn(col 7:decimal(11,3), col 
2:decimal(7,2))(children: DecimalColAddDecimalScalar(col 1:decimal(10,3), val 
2.34) -> 7:decimal(11,3)) -> 8:decimal(21,11), 
DecimalColMultiplyDecimalColumn(col 1:decimal(10,3), col 
9:decimal(12,6))(children: DecimalColDivideDecimalScalar(col 2:decimal(7,2), 
val 3.4) -> 9:decimal(12,6)) -> 10:decimal(23,9), 
DecimalColModuloDecimalScalar(col 1:decimal(10,3), val 10) -> 11:decimal(5,3), 
CastDecimalToLong(col 1:decimal(10,3)) -> 12:int, CastDecimalToLong(col 
2:decimal(7,2)) -> 13:smallint, CastDecimalToLong(col 2:decimal(7,2)) -> 
14:tinyint, CastDecimalToLong(col 1:decimal(10,3)) -> 15:bigint, 
CastDecimalToBoolean(col 1:
 decimal(10,3)) -> 16:boolean, CastDecimalToDouble(col 2:decimal(7,2)) -> 
17:double, CastDecimalToFloat(col 1:decimal(10,3)) -> 18:float, 
CastDecimalToString(col 2:decimal(7,2)) -> 19:string, 
CastDecimalToTimestamp(col 1:decimal(10,3)) -> 20:timestamp
+                          projectedOutputColumnNums: [6, 8, 10, 12, 13, 14, 
15, 16, 17, 18, 19, 20, 21, 22]
+                          selectExpressions: DecimalColAddDecimalColumn(col 
4:decimal(10,3), col 5:decimal(7,2))(children: ConvertDecimal64ToDecimal(col 
1:decimal(10,3)/DECIMAL_64) -> 4:decimal(10,3), ConvertDecimal64ToDecimal(col 
2:decimal(7,2)/DECIMAL_64) -> 5:decimal(7,2)) -> 6:decimal(11,3), 
DecimalColSubtractDecimalColumn(col 4:decimal(10,3), col 
7:decimal(9,2))(children: ConvertDecimal64ToDecimal(col 
1:decimal(10,3)/DECIMAL_64) -> 4:decimal(10,3), 
DecimalScalarMultiplyDecimalColumn(val 2, col 5:decimal(7,2))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(7,2)/DECIMAL_64) -> 5:decimal(7,2)) -> 
7:decimal(9,2)) -> 8:decimal(11,3), DecimalColDivideDecimalColumn(col 
23:decimal(11,3), col 5:decimal(7,2))(children: ConvertDecimal64ToDecimal(col 
9:decimal(11,3)/DECIMAL_64)(children: Decimal64ColAddDecimal64Scalar(col 
1:decimal(10,3)/DECIMAL_64, decimal64Val 2340, decimalVal 2.34) -> 
9:decimal(11,3)/DECIMAL_64) -> 23:decimal(11,3), ConvertDecimal64ToDecimal(col 
2:decimal(7,2)/DECIMA
 L_64) -> 5:decimal(7,2)) -> 10:decimal(21,11), 
DecimalColMultiplyDecimalColumn(col 4:decimal(10,3), col 
11:decimal(12,6))(children: ConvertDecimal64ToDecimal(col 
1:decimal(10,3)/DECIMAL_64) -> 4:decimal(10,3), 
DecimalColDivideDecimalScalar(col 5:decimal(7,2), val 3.4)(children: 
ConvertDecimal64ToDecimal(col 2:decimal(7,2)/DECIMAL_64) -> 5:decimal(7,2)) -> 
11:decimal(12,6)) -> 12:decimal(23,9), DecimalColModuloDecimalScalar(col 
4:decimal(10,3), val 10)(children: ConvertDecimal64ToDecimal(col 
1:decimal(10,3)/DECIMAL_64) -> 4:decimal(10,3)) -> 13:decimal(5,3), 
CastDecimalToLong(col 4:decimal(10,3))(children: ConvertDecimal64ToDecimal(col 
1:decimal(10,3)/DECIMAL_64) -> 4:decimal(10,3)) -> 14:int, 
CastDecimalToLong(col 5:decimal(7,2))(children: ConvertDecimal64ToDecimal(col 
2:decimal(7,2)/DECIMAL_64) -> 5:decimal(7,2)) -> 15:smallint, 
CastDecimalToLong(col 5:decimal(7,2))(children: ConvertDecimal64ToDecimal(col 
2:decimal(7,2)/DECIMAL_64) -> 5:decimal(7,2)) -> 16:tinyint, CastDecimalToLon
 g(col 4:decimal(10,3))(children: ConvertDecimal64ToDecimal(col 
1:decimal(10,3)/DECIMAL_64) -> 4:decimal(10,3)) -> 17:bigint, 
CastDecimalToBoolean(col 4:decimal(10,3))(children: 
ConvertDecimal64ToDecimal(col 1:decimal(10,3)/DECIMAL_64) -> 4:decimal(10,3)) 
-> 18:boolean, CastDecimalToDouble(col 5:decimal(7,2))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(7,2)/DECIMAL_64) -> 5:decimal(7,2)) -> 
19:double, CastDecimalToFloat(col 4:decimal(10,3))(children: 
ConvertDecimal64ToDecimal(col 1:decimal(10,3)/DECIMAL_64) -> 4:decimal(10,3)) 
-> 20:float, CastDecimalToString(col 5:decimal(7,2))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(7,2)/DECIMAL_64) -> 5:decimal(7,2)) -> 
21:string, CastDecimalToTimestamp(col 4:decimal(10,3))(children: 
ConvertDecimal64ToDecimal(col 1:decimal(10,3)/DECIMAL_64) -> 4:decimal(10,3)) 
-> 22:timestamp
                       Statistics: Num rows: 455 Data size: 100294 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
                         key expressions: _col0 (type: decimal(11,3)), _col1 
(type: decimal(11,3)), _col2 (type: decimal(21,11)), _col3 (type: 
decimal(23,9)), _col4 (type: decimal(5,3)), _col5 (type: int), _col6 (type: 
smallint), _col7 (type: tinyint), _col8 (type: bigint), _col9 (type: boolean), 
_col10 (type: double), _col11 (type: float), _col12 (type: string), _col13 
(type: timestamp)
                         sort order: ++++++++++++++
                         Reduce Sink Vectorization:
                             className: VectorReduceSinkObjectHashOperator
-                            keyColumnNums: [4, 6, 8, 10, 11, 12, 13, 14, 15, 
16, 17, 18, 19, 20]
+                            keyColumnNums: [6, 8, 10, 12, 13, 14, 15, 16, 17, 
18, 19, 20, 21, 22]
                             native: true
                             nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
                             valueColumnNums: []
@@ -267,8 +267,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -276,9 +276,9 @@ STAGE PLANS:
                 rowBatchContext:
                     dataColumnCount: 3
                     includeColumns: [0, 1, 2]
-                    dataColumns: cdouble:double, cdecimal1:decimal(10,3), 
cdecimal2:decimal(7,2)
+                    dataColumns: cdouble:double, 
cdecimal1:decimal(10,3)/DECIMAL_64, cdecimal2:decimal(7,2)/DECIMAL_64
                     partitionColumnCount: 0
-                    scratchColumnTypeNames: [decimal(11,3), decimal(9,2), 
decimal(11,3), decimal(11,3), decimal(21,11), decimal(12,6), decimal(23,9), 
decimal(5,3), bigint, bigint, bigint, bigint, bigint, double, double, string, 
timestamp]
+                    scratchColumnTypeNames: [decimal(10,3), decimal(7,2), 
decimal(11,3), decimal(9,2), decimal(11,3), decimal(11,3)/DECIMAL_64, 
decimal(21,11), decimal(12,6), decimal(23,9), decimal(5,3), bigint, bigint, 
bigint, bigint, bigint, double, double, string, timestamp, decimal(11,3)]
         Reducer 2 
             Execution mode: vectorized, llap
             Reduce Vectorization:
@@ -360,4 +360,4 @@ ORDER BY c1, c2, c3, c4, c5, c6, c7, c8, c9, c10, c11, c12, 
c13, c14) q
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@decimal_test_small_n0
 #### A masked pattern was here ####
-774841630076
+1273824888155

http://git-wip-us.apache.org/repos/asf/hive/blob/dd512593/ql/src/test/results/clientpositive/llap/vector_decimal_mapjoin.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_decimal_mapjoin.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_mapjoin.q.out
index 7b56d27..61f6609 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_mapjoin.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_mapjoin.q.out
@@ -152,8 +152,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false
@@ -203,8 +203,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -338,6 +338,17 @@ POSTHOOK: Input: default@t2_n29
 9.00   9
 9.00   9
 9.00   9
+PREHOOK: query: select count(*) from (select t1_n48.`dec`, t2_n29.`dec` from 
t1_n48 join t2_n29 on (t1_n48.`dec`=t2_n29.`dec`)) as t
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1_n48
+PREHOOK: Input: default@t2_n29
+#### A masked pattern was here ####
+POSTHOOK: query: select count(*) from (select t1_n48.`dec`, t2_n29.`dec` from 
t1_n48 join t2_n29 on (t1_n48.`dec`=t2_n29.`dec`)) as t
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1_n48
+POSTHOOK: Input: default@t2_n29
+#### A masked pattern was here ####
+106
 PREHOOK: query: explain vectorization detail
 select t1_n48.`dec`, t1_n48.value_dec, t2_n29.`dec`, t2_n29.value_dec from 
t1_n48 join t2_n29 on (t1_n48.`dec`=t2_n29.`dec`)
 PREHOOK: type: QUERY
@@ -416,8 +427,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false
@@ -468,8 +479,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -603,6 +614,17 @@ POSTHOOK: Input: default@t2_n29
 9.00   48.96   9       5
 9.00   48.96   9       7
 9.00   48.96   9       7
+PREHOOK: query: select count(*) from (select t1_n48.`dec`, t1_n48.value_dec, 
t2_n29.`dec`, t2_n29.value_dec from t1_n48 join t2_n29 on 
(t1_n48.`dec`=t2_n29.`dec`)) as t
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1_n48
+PREHOOK: Input: default@t2_n29
+#### A masked pattern was here ####
+POSTHOOK: query: select count(*) from (select t1_n48.`dec`, t1_n48.value_dec, 
t2_n29.`dec`, t2_n29.value_dec from t1_n48 join t2_n29 on 
(t1_n48.`dec`=t2_n29.`dec`)) as t
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1_n48
+POSTHOOK: Input: default@t2_n29
+#### A masked pattern was here ####
+106
 PREHOOK: query: CREATE TABLE over1k_small(t tinyint,
            si smallint,
            i int,
@@ -708,12 +730,12 @@ STAGE PLANS:
                   Statistics: Num rows: 1049 Data size: 111776 Basic stats: 
COMPLETE Column stats: NONE
                   TableScan Vectorization:
                       native: true
-                      vectorizationSchemaColumns: [0:dec:decimal(14,2), 
1:value_dec:decimal(14,2), 
2:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
+                      vectorizationSchemaColumns: 
[0:dec:decimal(14,2)/DECIMAL_64, 1:value_dec:decimal(14,2)/DECIMAL_64, 
2:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
                   Filter Operator
                     Filter Vectorization:
                         className: VectorFilterOperator
                         native: true
-                        predicateExpression: SelectColumnIsNotNull(col 
0:decimal(14,2))
+                        predicateExpression: SelectColumnIsNotNull(col 
3:decimal(14,2))(children: ConvertDecimal64ToDecimal(col 
0:decimal(14,2)/DECIMAL_64) -> 3:decimal(14,2))
                     predicate: dec is not null (type: boolean)
                     Statistics: Num rows: 997 Data size: 106235 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
@@ -731,8 +753,8 @@ STAGE PLANS:
                           0 _col0 (type: decimal(16,2))
                           1 _col0 (type: decimal(16,2))
                         Map Join Vectorization:
-                            bigTableKeyExpressions: col 0:decimal(16,2)
-                            bigTableValueExpressions: col 0:decimal(16,2)
+                            bigTableKeyExpressions: 
ConvertDecimal64ToDecimal(col 0:decimal(16,2)/DECIMAL_64) -> 4:decimal(16,2)
+                            bigTableValueExpressions: 
ConvertDecimal64ToDecimal(col 0:decimal(16,2)/DECIMAL_64) -> 5:decimal(16,2)
                             className: VectorMapJoinOperator
                             native: false
                             nativeConditionsMet: 
hive.mapjoin.optimized.hashtable IS true, 
hive.vectorized.execution.mapjoin.native.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, One MapJoin Condition IS true, No nullsafe IS 
true, Small table vectorizes IS true
@@ -758,8 +780,7 @@ STAGE PLANS:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vector.serde.deserialize IS true
                 inputFormatFeatureSupport: [DECIMAL_64]
-                vectorizationSupportRemovedReasons: [DECIMAL_64 disabled 
because LLAP is enabled]
-                featureSupportInUse: []
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: org.apache.hadoop.mapred.TextInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false
@@ -767,9 +788,9 @@ STAGE PLANS:
                 rowBatchContext:
                     dataColumnCount: 2
                     includeColumns: [0]
-                    dataColumns: dec:decimal(14,2), value_dec:decimal(14,2)
+                    dataColumns: dec:decimal(14,2)/DECIMAL_64, 
value_dec:decimal(14,2)/DECIMAL_64
                     partitionColumnCount: 0
-                    scratchColumnTypeNames: []
+                    scratchColumnTypeNames: [decimal(14,2), decimal(16,2), 
decimal(16,2)]
         Map 2 
             Map Operator Tree:
                 TableScan
@@ -777,12 +798,12 @@ STAGE PLANS:
                   Statistics: Num rows: 1049 Data size: 111776 Basic stats: 
COMPLETE Column stats: NONE
                   TableScan Vectorization:
                       native: true
-                      vectorizationSchemaColumns: [0:dec:decimal(14,0), 
1:value_dec:decimal(14,0), 
2:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
+                      vectorizationSchemaColumns: 
[0:dec:decimal(14,0)/DECIMAL_64, 1:value_dec:decimal(14,0)/DECIMAL_64, 
2:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
                   Filter Operator
                     Filter Vectorization:
                         className: VectorFilterOperator
                         native: true
-                        predicateExpression: SelectColumnIsNotNull(col 
0:decimal(14,0))
+                        predicateExpression: SelectColumnIsNotNull(col 
3:decimal(14,0))(children: ConvertDecimal64ToDecimal(col 
0:decimal(14,0)/DECIMAL_64) -> 3:decimal(14,0))
                     predicate: dec is not null (type: boolean)
                     Statistics: Num rows: 997 Data size: 106235 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
@@ -810,8 +831,7 @@ STAGE PLANS:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vector.serde.deserialize IS true
                 inputFormatFeatureSupport: [DECIMAL_64]
-                vectorizationSupportRemovedReasons: [DECIMAL_64 disabled 
because LLAP is enabled]
-                featureSupportInUse: []
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: org.apache.hadoop.mapred.TextInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -819,9 +839,9 @@ STAGE PLANS:
                 rowBatchContext:
                     dataColumnCount: 2
                     includeColumns: [0]
-                    dataColumns: dec:decimal(14,0), value_dec:decimal(14,0)
+                    dataColumns: dec:decimal(14,0)/DECIMAL_64, 
value_dec:decimal(14,0)/DECIMAL_64
                     partitionColumnCount: 0
-                    scratchColumnTypeNames: []
+                    scratchColumnTypeNames: [decimal(14,0)]
 
   Stage: Stage-0
     Fetch Operator
@@ -839,112 +859,18 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@t1_small
 POSTHOOK: Input: default@t2_small
 #### A masked pattern was here ####
-14.00  14
-14.00  14
-14.00  14
-14.00  14
-14.00  14
-14.00  14
-14.00  14
-14.00  14
-14.00  14
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-45.00  45
-45.00  45
-45.00  45
-45.00  45
-45.00  45
-6.00   6
-6.00   6
-6.00   6
-6.00   6
-6.00   6
-6.00   6
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-70.00  70
-70.00  70
-70.00  70
-70.00  70
-70.00  70
-70.00  70
-70.00  70
-79.00  79
-79.00  79
-79.00  79
-79.00  79
-79.00  79
-79.00  79
-89.00  89
 89.00  89
-89.00  89
-89.00  89
-89.00  89
-89.00  89
-89.00  89
-89.00  89
-89.00  89
-89.00  89
-89.00  89
-89.00  89
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
+PREHOOK: query: select count(*) from (select t1_small.`dec`, t2_small.`dec` 
from t1_small join t2_small on (t1_small.`dec`=t2_small.`dec`)) as t
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1_small
+PREHOOK: Input: default@t2_small
+#### A masked pattern was here ####
+POSTHOOK: query: select count(*) from (select t1_small.`dec`, t2_small.`dec` 
from t1_small join t2_small on (t1_small.`dec`=t2_small.`dec`)) as t
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1_small
+POSTHOOK: Input: default@t2_small
+#### A masked pattern was here ####
+1
 PREHOOK: query: explain vectorization detail
 select t1_small.`dec`, t1_small.value_dec, t2_small.`dec`, t2_small.value_dec 
from t1_small join t2_small on (t1_small.`dec`=t2_small.`dec`)
 PREHOOK: type: QUERY
@@ -974,12 +900,12 @@ STAGE PLANS:
                   Statistics: Num rows: 1049 Data size: 223552 Basic stats: 
COMPLETE Column stats: NONE
                   TableScan Vectorization:
                       native: true
-                      vectorizationSchemaColumns: [0:dec:decimal(14,2), 
1:value_dec:decimal(14,2), 
2:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
+                      vectorizationSchemaColumns: 
[0:dec:decimal(14,2)/DECIMAL_64, 1:value_dec:decimal(14,2)/DECIMAL_64, 
2:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
                   Filter Operator
                     Filter Vectorization:
                         className: VectorFilterOperator
                         native: true
-                        predicateExpression: SelectColumnIsNotNull(col 
0:decimal(14,2))
+                        predicateExpression: SelectColumnIsNotNull(col 
3:decimal(14,2))(children: ConvertDecimal64ToDecimal(col 
0:decimal(14,2)/DECIMAL_64) -> 3:decimal(14,2))
                     predicate: dec is not null (type: boolean)
                     Statistics: Num rows: 997 Data size: 212470 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
@@ -997,8 +923,8 @@ STAGE PLANS:
                           0 _col0 (type: decimal(16,2))
                           1 _col0 (type: decimal(16,2))
                         Map Join Vectorization:
-                            bigTableKeyExpressions: col 0:decimal(16,2)
-                            bigTableValueExpressions: col 0:decimal(16,2), col 
1:decimal(14,2)
+                            bigTableKeyExpressions: 
ConvertDecimal64ToDecimal(col 0:decimal(16,2)/DECIMAL_64) -> 4:decimal(16,2)
+                            bigTableValueExpressions: 
ConvertDecimal64ToDecimal(col 0:decimal(16,2)/DECIMAL_64) -> 5:decimal(16,2), 
ConvertDecimal64ToDecimal(col 1:decimal(14,2)/DECIMAL_64) -> 3:decimal(14,2)
                             className: VectorMapJoinOperator
                             native: false
                             nativeConditionsMet: 
hive.mapjoin.optimized.hashtable IS true, 
hive.vectorized.execution.mapjoin.native.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, One MapJoin Condition IS true, No nullsafe IS 
true, Small table vectorizes IS true
@@ -1024,8 +950,7 @@ STAGE PLANS:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vector.serde.deserialize IS true
                 inputFormatFeatureSupport: [DECIMAL_64]
-                vectorizationSupportRemovedReasons: [DECIMAL_64 disabled 
because LLAP is enabled]
-                featureSupportInUse: []
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: org.apache.hadoop.mapred.TextInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false
@@ -1033,9 +958,9 @@ STAGE PLANS:
                 rowBatchContext:
                     dataColumnCount: 2
                     includeColumns: [0, 1]
-                    dataColumns: dec:decimal(14,2), value_dec:decimal(14,2)
+                    dataColumns: dec:decimal(14,2)/DECIMAL_64, 
value_dec:decimal(14,2)/DECIMAL_64
                     partitionColumnCount: 0
-                    scratchColumnTypeNames: [decimal(14,0)]
+                    scratchColumnTypeNames: [decimal(14,2), decimal(16,2), 
decimal(16,2), decimal(14,0)]
         Map 2 
             Map Operator Tree:
                 TableScan
@@ -1043,12 +968,12 @@ STAGE PLANS:
                   Statistics: Num rows: 1049 Data size: 223552 Basic stats: 
COMPLETE Column stats: NONE
                   TableScan Vectorization:
                       native: true
-                      vectorizationSchemaColumns: [0:dec:decimal(14,0), 
1:value_dec:decimal(14,0), 
2:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
+                      vectorizationSchemaColumns: 
[0:dec:decimal(14,0)/DECIMAL_64, 1:value_dec:decimal(14,0)/DECIMAL_64, 
2:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
                   Filter Operator
                     Filter Vectorization:
                         className: VectorFilterOperator
                         native: true
-                        predicateExpression: SelectColumnIsNotNull(col 
0:decimal(14,0))
+                        predicateExpression: SelectColumnIsNotNull(col 
3:decimal(14,0))(children: ConvertDecimal64ToDecimal(col 
0:decimal(14,0)/DECIMAL_64) -> 3:decimal(14,0))
                     predicate: dec is not null (type: boolean)
                     Statistics: Num rows: 997 Data size: 212470 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
@@ -1077,8 +1002,7 @@ STAGE PLANS:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vector.serde.deserialize IS true
                 inputFormatFeatureSupport: [DECIMAL_64]
-                vectorizationSupportRemovedReasons: [DECIMAL_64 disabled 
because LLAP is enabled]
-                featureSupportInUse: []
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: org.apache.hadoop.mapred.TextInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -1086,9 +1010,9 @@ STAGE PLANS:
                 rowBatchContext:
                     dataColumnCount: 2
                     includeColumns: [0, 1]
-                    dataColumns: dec:decimal(14,0), value_dec:decimal(14,0)
+                    dataColumns: dec:decimal(14,0)/DECIMAL_64, 
value_dec:decimal(14,0)/DECIMAL_64
                     partitionColumnCount: 0
-                    scratchColumnTypeNames: []
+                    scratchColumnTypeNames: [decimal(14,0)]
 
   Stage: Stage-0
     Fetch Operator
@@ -1106,112 +1030,18 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@t1_small
 POSTHOOK: Input: default@t2_small
 #### A masked pattern was here ####
-14.00  33.66   14      10
-14.00  33.66   14      22
-14.00  33.66   14      34
-14.00  33.66   14      39
-14.00  33.66   14      42
-14.00  33.66   14      45
-14.00  33.66   14      46
-14.00  33.66   14      49
-14.00  33.66   14      5
-17.00  14.26   17      1
-17.00  14.26   17      14
-17.00  14.26   17      16
-17.00  14.26   17      19
-17.00  14.26   17      2
-17.00  14.26   17      22
-17.00  14.26   17      29
-17.00  14.26   17      3
-17.00  14.26   17      4
-17.00  14.26   17      44
-45.00  23.55   45      1
-45.00  23.55   45      2
-45.00  23.55   45      22
-45.00  23.55   45      24
-45.00  23.55   45      42
-6.00   29.78   6       16
-6.00   29.78   6       28
-6.00   29.78   6       30
-6.00   29.78   6       34
-6.00   29.78   6       36
-6.00   29.78   6       44
-62.00  21.02   62      15
-62.00  21.02   62      15
-62.00  21.02   62      21
-62.00  21.02   62      21
-62.00  21.02   62      22
-62.00  21.02   62      25
-62.00  21.02   62      29
-62.00  21.02   62      3
-62.00  21.02   62      34
-62.00  21.02   62      47
-62.00  21.02   62      47
-62.00  21.02   62      49
-64.00  37.76   64      0
-64.00  37.76   64      10
-64.00  37.76   64      10
-64.00  37.76   64      13
-64.00  37.76   64      23
-64.00  37.76   64      25
-64.00  37.76   64      26
-64.00  37.76   64      27
-64.00  37.76   64      27
-64.00  37.76   64      30
-64.00  37.76   64      32
-64.00  37.76   64      34
-64.00  37.76   64      35
-64.00  37.76   64      38
-64.00  37.76   64      40
-64.00  37.76   64      43
-64.00  37.76   64      5
-64.00  37.76   64      50
-70.00  24.59   70      2
-70.00  24.59   70      25
-70.00  24.59   70      27
-70.00  24.59   70      28
-70.00  24.59   70      3
-70.00  24.59   70      32
-70.00  24.59   70      44
-79.00  15.12   79      1
-79.00  15.12   79      15
-79.00  15.12   79      25
-79.00  15.12   79      30
-79.00  15.12   79      35
-79.00  15.12   79      35
-89.00  15.09   89      1
 89.00  15.09   89      15
-89.00  15.09   89      23
-89.00  15.09   89      27
-89.00  15.09   89      28
-89.00  15.09   89      29
-89.00  15.09   89      30
-89.00  15.09   89      32
-89.00  15.09   89      39
-89.00  15.09   89      40
-89.00  15.09   89      45
-89.00  15.09   89      7
-9.00   48.96   9       12
-9.00   48.96   9       15
-9.00   48.96   9       2
-9.00   48.96   9       2
-9.00   48.96   9       2
-9.00   48.96   9       20
-9.00   48.96   9       20
-9.00   48.96   9       21
-9.00   48.96   9       21
-9.00   48.96   9       26
-9.00   48.96   9       27
-9.00   48.96   9       34
-9.00   48.96   9       38
-9.00   48.96   9       41
-9.00   48.96   9       42
-9.00   48.96   9       45
-9.00   48.96   9       48
-9.00   48.96   9       49
-9.00   48.96   9       5
-9.00   48.96   9       7
-9.00   48.96   9       7
+PREHOOK: query: select count(*) from (select t1_small.`dec`, 
t1_small.value_dec, t2_small.`dec`, t2_small.value_dec from t1_small join 
t2_small on (t1_small.`dec`=t2_small.`dec`)) as t
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1_small
+PREHOOK: Input: default@t2_small
+#### A masked pattern was here ####
+POSTHOOK: query: select count(*) from (select t1_small.`dec`, 
t1_small.value_dec, t2_small.`dec`, t2_small.value_dec from t1_small join 
t2_small on (t1_small.`dec`=t2_small.`dec`)) as t
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1_small
+POSTHOOK: Input: default@t2_small
+#### A masked pattern was here ####
+1
 PREHOOK: query: explain vectorization detail
 select t1_small.`dec`, t2_small.`dec` from t1_small join t2_small on 
(t1_small.`dec`=t2_small.`dec`)
 PREHOOK: type: QUERY
@@ -1372,112 +1202,18 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@t1_small
 POSTHOOK: Input: default@t2_small
 #### A masked pattern was here ####
-14.00  14
-14.00  14
-14.00  14
-14.00  14
-14.00  14
-14.00  14
-14.00  14
-14.00  14
-14.00  14
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-17.00  17
-45.00  45
-45.00  45
-45.00  45
-45.00  45
-45.00  45
-6.00   6
-6.00   6
-6.00   6
-6.00   6
-6.00   6
-6.00   6
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-62.00  62
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-64.00  64
-70.00  70
-70.00  70
-70.00  70
-70.00  70
-70.00  70
-70.00  70
-70.00  70
-79.00  79
-79.00  79
-79.00  79
-79.00  79
-79.00  79
-79.00  79
-89.00  89
-89.00  89
-89.00  89
-89.00  89
-89.00  89
 89.00  89
-89.00  89
-89.00  89
-89.00  89
-89.00  89
-89.00  89
-89.00  89
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
-9.00   9
+PREHOOK: query: select count(*) from (select t1_small.`dec`, t2_small.`dec` 
from t1_small join t2_small on (t1_small.`dec`=t2_small.`dec`)) as t
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1_small
+PREHOOK: Input: default@t2_small
+#### A masked pattern was here ####
+POSTHOOK: query: select count(*) from (select t1_small.`dec`, t2_small.`dec` 
from t1_small join t2_small on (t1_small.`dec`=t2_small.`dec`)) as t
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1_small
+POSTHOOK: Input: default@t2_small
+#### A masked pattern was here ####
+1
 PREHOOK: query: explain vectorization detail
 select t1_small.`dec`, t1_small.value_dec, t2_small.`dec`, t2_small.value_dec 
from t1_small join t2_small on (t1_small.`dec`=t2_small.`dec`)
 PREHOOK: type: QUERY
@@ -1639,109 +1375,15 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@t1_small
 POSTHOOK: Input: default@t2_small
 #### A masked pattern was here ####
-14.00  33.66   14      10
-14.00  33.66   14      22
-14.00  33.66   14      34
-14.00  33.66   14      39
-14.00  33.66   14      42
-14.00  33.66   14      45
-14.00  33.66   14      46
-14.00  33.66   14      49
-14.00  33.66   14      5
-17.00  14.26   17      1
-17.00  14.26   17      14
-17.00  14.26   17      16
-17.00  14.26   17      19
-17.00  14.26   17      2
-17.00  14.26   17      22
-17.00  14.26   17      29
-17.00  14.26   17      3
-17.00  14.26   17      4
-17.00  14.26   17      44
-45.00  23.55   45      1
-45.00  23.55   45      2
-45.00  23.55   45      22
-45.00  23.55   45      24
-45.00  23.55   45      42
-6.00   29.78   6       16
-6.00   29.78   6       28
-6.00   29.78   6       30
-6.00   29.78   6       34
-6.00   29.78   6       36
-6.00   29.78   6       44
-62.00  21.02   62      15
-62.00  21.02   62      15
-62.00  21.02   62      21
-62.00  21.02   62      21
-62.00  21.02   62      22
-62.00  21.02   62      25
-62.00  21.02   62      29
-62.00  21.02   62      3
-62.00  21.02   62      34
-62.00  21.02   62      47
-62.00  21.02   62      47
-62.00  21.02   62      49
-64.00  37.76   64      0
-64.00  37.76   64      10
-64.00  37.76   64      10
-64.00  37.76   64      13
-64.00  37.76   64      23
-64.00  37.76   64      25
-64.00  37.76   64      26
-64.00  37.76   64      27
-64.00  37.76   64      27
-64.00  37.76   64      30
-64.00  37.76   64      32
-64.00  37.76   64      34
-64.00  37.76   64      35
-64.00  37.76   64      38
-64.00  37.76   64      40
-64.00  37.76   64      43
-64.00  37.76   64      5
-64.00  37.76   64      50
-70.00  24.59   70      2
-70.00  24.59   70      25
-70.00  24.59   70      27
-70.00  24.59   70      28
-70.00  24.59   70      3
-70.00  24.59   70      32
-70.00  24.59   70      44
-79.00  15.12   79      1
-79.00  15.12   79      15
-79.00  15.12   79      25
-79.00  15.12   79      30
-79.00  15.12   79      35
-79.00  15.12   79      35
-89.00  15.09   89      1
 89.00  15.09   89      15
-89.00  15.09   89      23
-89.00  15.09   89      27
-89.00  15.09   89      28
-89.00  15.09   89      29
-89.00  15.09   89      30
-89.00  15.09   89      32
-89.00  15.09   89      39
-89.00  15.09   89      40
-89.00  15.09   89      45
-89.00  15.09   89      7
-9.00   48.96   9       12
-9.00   48.96   9       15
-9.00   48.96   9       2
-9.00   48.96   9       2
-9.00   48.96   9       2
-9.00   48.96   9       20
-9.00   48.96   9       20
-9.00   48.96   9       21
-9.00   48.96   9       21
-9.00   48.96   9       26
-9.00   48.96   9       27
-9.00   48.96   9       34
-9.00   48.96   9       38
-9.00   48.96   9       41
-9.00   48.96   9       42
-9.00   48.96   9       45
-9.00   48.96   9       48
-9.00   48.96   9       49
-9.00   48.96   9       5
-9.00   48.96   9       7
-9.00   48.96   9       7
+PREHOOK: query: select count(*) from (select t1_small.`dec`, 
t1_small.value_dec, t2_small.`dec`, t2_small.value_dec from t1_small join 
t2_small on (t1_small.`dec`=t2_small.`dec`)) as t
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1_small
+PREHOOK: Input: default@t2_small
+#### A masked pattern was here ####
+POSTHOOK: query: select count(*) from (select t1_small.`dec`, 
t1_small.value_dec, t2_small.`dec`, t2_small.value_dec from t1_small join 
t2_small on (t1_small.`dec`=t2_small.`dec`)) as t
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1_small
+POSTHOOK: Input: default@t2_small
+#### A masked pattern was here ####
+1

http://git-wip-us.apache.org/repos/asf/hive/blob/dd512593/ql/src/test/results/clientpositive/llap/vector_decimal_math_funcs.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_decimal_math_funcs.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_math_funcs.q.out
index ec43f62..e3d5044 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_math_funcs.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_math_funcs.q.out
@@ -150,8 +150,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false
@@ -364,12 +364,12 @@ STAGE PLANS:
                   Statistics: Num rows: 12288 Data size: 1401000 Basic stats: 
COMPLETE Column stats: NONE
                   TableScan Vectorization:
                       native: true
-                      vectorizationSchemaColumns: [0:cbigint:bigint, 
1:cdouble:double, 2:cdecimal1:decimal(12,4), 3:cdecimal2:decimal(14,8), 
4:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
+                      vectorizationSchemaColumns: [0:cbigint:bigint, 
1:cdouble:double, 2:cdecimal1:decimal(12,4)/DECIMAL_64, 
3:cdecimal2:decimal(14,8)/DECIMAL_64, 
4:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
                   Filter Operator
                     Filter Vectorization:
                         className: VectorFilterOperator
                         native: true
-                        predicateExpression: FilterExprAndExpr(children: 
FilterLongColEqualLongScalar(col 5:bigint, val 0)(children: 
LongColModuloLongScalar(col 0:bigint, val 500) -> 5:bigint), 
FilterDoubleColGreaterEqualDoubleScalar(col 7:double, val -1.0)(children: 
FuncSinDoubleToDouble(col 6:double)(children: CastDecimalToDouble(col 
2:decimal(12,4)) -> 6:double) -> 7:double))
+                        predicateExpression: FilterExprAndExpr(children: 
FilterLongColEqualLongScalar(col 5:bigint, val 0)(children: 
LongColModuloLongScalar(col 0:bigint, val 500) -> 5:bigint), 
FilterDoubleColGreaterEqualDoubleScalar(col 8:double, val -1.0)(children: 
FuncSinDoubleToDouble(col 7:double)(children: CastDecimalToDouble(col 
6:decimal(12,4))(children: ConvertDecimal64ToDecimal(col 
2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) -> 7:double) -> 8:double))
                     predicate: (((cbigint % 500) = 0) and (sin(cdecimal1) >= 
-1.0D)) (type: boolean)
                     Statistics: Num rows: 2048 Data size: 233500 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
@@ -378,8 +378,8 @@ STAGE PLANS:
                       Select Vectorization:
                           className: VectorSelectOperator
                           native: true
-                          projectedOutputColumnNums: [2, 8, 9, 10, 11, 6, 12, 
13, 14, 16, 17, 7, 18, 20, 21, 22, 23, 24, 25, 26, 27, 28, 2, 29, 5, 30]
-                          selectExpressions: 
FuncRoundWithNumDigitsDecimalToDecimal(col 2:decimal(12,4), decimalPlaces 2) -> 
8:decimal(11,2), FuncRoundDecimalToDecimal(col 2:decimal(12,4)) -> 
9:decimal(9,0), FuncFloorDecimalToDecimal(col 2:decimal(12,4)) -> 
10:decimal(9,0), FuncCeilDecimalToDecimal(col 2:decimal(12,4)) -> 
11:decimal(9,0), RoundWithNumDigitsDoubleToDouble(col 7, decimalPlaces 
58)(children: FuncExpDoubleToDouble(col 6:double)(children: 
CastDecimalToDouble(col 2:decimal(12,4)) -> 6:double) -> 7:double) -> 6:double, 
FuncLnDoubleToDouble(col 7:double)(children: CastDecimalToDouble(col 
2:decimal(12,4)) -> 7:double) -> 12:double, FuncLog10DoubleToDouble(col 
7:double)(children: CastDecimalToDouble(col 2:decimal(12,4)) -> 7:double) -> 
13:double, FuncLog2DoubleToDouble(col 7:double)(children: 
CastDecimalToDouble(col 2:decimal(12,4)) -> 7:double) -> 14:double, 
FuncLog2DoubleToDouble(col 7:double)(children: CastDecimalToDouble(col 
15:decimal(13,4))(children: DecimalColSubtractD
 ecimalScalar(col 2:decimal(12,4), val 15601) -> 15:decimal(13,4)) -> 7:double) 
-> 16:double, FuncLogWithBaseDoubleToDouble(col 7:double)(children: 
CastDecimalToDouble(col 2:decimal(12,4)) -> 7:double) -> 17:double, 
FuncPowerDoubleToDouble(col 18:double)(children: FuncLog2DoubleToDouble(col 
7:double)(children: CastDecimalToDouble(col 2:decimal(12,4)) -> 7:double) -> 
18:double) -> 7:double, FuncPowerDoubleToDouble(col 19:double)(children: 
FuncLog2DoubleToDouble(col 18:double)(children: CastDecimalToDouble(col 
2:decimal(12,4)) -> 18:double) -> 19:double) -> 18:double, 
FuncSqrtDoubleToDouble(col 19:double)(children: CastDecimalToDouble(col 
2:decimal(12,4)) -> 19:double) -> 20:double, FuncAbsDecimalToDecimal(col 
2:decimal(12,4)) -> 21:decimal(12,4), FuncSinDoubleToDouble(col 
19:double)(children: CastDecimalToDouble(col 2:decimal(12,4)) -> 19:double) -> 
22:double, FuncASinDoubleToDouble(col 19:double)(children: 
CastDecimalToDouble(col 2:decimal(12,4)) -> 19:double) -> 23:double, FuncCosDo
 ubleToDouble(col 19:double)(children: CastDecimalToDouble(col 2:decimal(12,4)) 
-> 19:double) -> 24:double, FuncACosDoubleToDouble(col 19:double)(children: 
CastDecimalToDouble(col 2:decimal(12,4)) -> 19:double) -> 25:double, 
FuncATanDoubleToDouble(col 19:double)(children: CastDecimalToDouble(col 
2:decimal(12,4)) -> 19:double) -> 26:double, FuncDegreesDoubleToDouble(col 
19:double)(children: CastDecimalToDouble(col 2:decimal(12,4)) -> 19:double) -> 
27:double, FuncRadiansDoubleToDouble(col 19:double)(children: 
CastDecimalToDouble(col 2:decimal(12,4)) -> 19:double) -> 28:double, 
FuncNegateDecimalToDecimal(col 2:decimal(12,4)) -> 29:decimal(12,4), 
FuncSignDecimalToLong(col 2:decimal(12,4)) -> 5:int, FuncCosDoubleToDouble(col 
19:double)(children: DoubleColAddDoubleScalar(col 30:double, val 
3.14159)(children: DoubleColUnaryMinus(col 19:double)(children: 
FuncSinDoubleToDouble(col 30:double)(children: FuncLnDoubleToDouble(col 
19:double)(children: CastDecimalToDouble(col 2:decimal(12,4)) -> 19
 :double) -> 30:double) -> 19:double) -> 30:double) -> 19:double) -> 30:double
+                          projectedOutputColumnNums: [2, 9, 10, 11, 12, 7, 13, 
14, 15, 17, 18, 8, 19, 21, 22, 23, 24, 25, 26, 27, 28, 29, 2, 30, 5, 31]
+                          selectExpressions: 
FuncRoundWithNumDigitsDecimalToDecimal(col 6:decimal(12,4), decimalPlaces 
2)(children: ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 
6:decimal(12,4)) -> 9:decimal(11,2), FuncRoundDecimalToDecimal(col 
6:decimal(12,4))(children: ConvertDecimal64ToDecimal(col 
2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) -> 10:decimal(9,0), 
FuncFloorDecimalToDecimal(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 11:decimal(9,0), FuncCeilDecimalToDecimal(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 12:decimal(9,0), RoundWithNumDigitsDoubleToDouble(col 8, decimalPlaces 
58)(children: FuncExpDoubleToDouble(col 7:double)(children: 
CastDecimalToDouble(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 7:double) -> 8:double) -> 7:double, FuncLnDoubleToDouble(
 col 8:double)(children: CastDecimalToDouble(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 8:double) -> 13:double, FuncLog10DoubleToDouble(col 8:double)(children: 
CastDecimalToDouble(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 8:double) -> 14:double, FuncLog2DoubleToDouble(col 8:double)(children: 
CastDecimalToDouble(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 8:double) -> 15:double, FuncLog2DoubleToDouble(col 8:double)(children: 
CastDecimalToDouble(col 32:decimal(13,4))(children: 
ConvertDecimal64ToDecimal(col 16:decimal(13,4)/DECIMAL_64)(children: 
Decimal64ColSubtractDecimal64Scalar(col 2:decimal(12,4)/DECIMAL_64, 
decimal64Val 156010000, decimalVal 15601) -> 16:decimal(13,4)/DECIMAL_64) -> 
32:decimal(13,4)) -> 8:double) -> 17:double, FuncLogWithBaseDoubleToDouble(col 
8:double)(children
 : CastDecimalToDouble(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 8:double) -> 18:double, FuncPowerDoubleToDouble(col 19:double)(children: 
FuncLog2DoubleToDouble(col 8:double)(children: CastDecimalToDouble(col 
6:decimal(12,4))(children: ConvertDecimal64ToDecimal(col 
2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) -> 8:double) -> 19:double) -> 
8:double, FuncPowerDoubleToDouble(col 20:double)(children: 
FuncLog2DoubleToDouble(col 19:double)(children: CastDecimalToDouble(col 
6:decimal(12,4))(children: ConvertDecimal64ToDecimal(col 
2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) -> 19:double) -> 20:double) -> 
19:double, FuncSqrtDoubleToDouble(col 20:double)(children: 
CastDecimalToDouble(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 20:double) -> 21:double, FuncAbsDecimalToDecimal(col 
6:decimal(12,4))(children: ConvertDecimal64ToDecimal(col 2:decimal(12,4)/
 DECIMAL_64) -> 6:decimal(12,4)) -> 22:decimal(12,4), FuncSinDoubleToDouble(col 
20:double)(children: CastDecimalToDouble(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 20:double) -> 23:double, FuncASinDoubleToDouble(col 20:double)(children: 
CastDecimalToDouble(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 20:double) -> 24:double, FuncCosDoubleToDouble(col 20:double)(children: 
CastDecimalToDouble(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 20:double) -> 25:double, FuncACosDoubleToDouble(col 20:double)(children: 
CastDecimalToDouble(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 20:double) -> 26:double, FuncATanDoubleToDouble(col 20:double)(children: 
CastDecimalToDouble(col 6:decimal(12,4))(children: ConvertDecimal64ToDecimal(c
 ol 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) -> 20:double) -> 27:double, 
FuncDegreesDoubleToDouble(col 20:double)(children: CastDecimalToDouble(col 
6:decimal(12,4))(children: ConvertDecimal64ToDecimal(col 
2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) -> 20:double) -> 28:double, 
FuncRadiansDoubleToDouble(col 20:double)(children: CastDecimalToDouble(col 
6:decimal(12,4))(children: ConvertDecimal64ToDecimal(col 
2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) -> 20:double) -> 29:double, 
FuncNegateDecimalToDecimal(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 30:decimal(12,4), FuncSignDecimalToLong(col 6:decimal(12,4))(children: 
ConvertDecimal64ToDecimal(col 2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) 
-> 5:int, FuncCosDoubleToDouble(col 20:double)(children: 
DoubleColAddDoubleScalar(col 31:double, val 3.14159)(children: 
DoubleColUnaryMinus(col 20:double)(children: FuncSinDoubleToDouble(col 
31:double)(children: Fu
 ncLnDoubleToDouble(col 20:double)(children: CastDecimalToDouble(col 
6:decimal(12,4))(children: ConvertDecimal64ToDecimal(col 
2:decimal(12,4)/DECIMAL_64) -> 6:decimal(12,4)) -> 20:double) -> 31:double) -> 
20:double) -> 31:double) -> 20:double) -> 31:double
                       Statistics: Num rows: 2048 Data size: 233500 Basic 
stats: COMPLETE Column stats: NONE
                       File Output Operator
                         compressed: false
@@ -396,8 +396,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false
@@ -405,9 +405,9 @@ STAGE PLANS:
                 rowBatchContext:
                     dataColumnCount: 4
                     includeColumns: [0, 2]
-                    dataColumns: cbigint:bigint, cdouble:double, 
cdecimal1:decimal(12,4), cdecimal2:decimal(14,8)
+                    dataColumns: cbigint:bigint, cdouble:double, 
cdecimal1:decimal(12,4)/DECIMAL_64, cdecimal2:decimal(14,8)/DECIMAL_64
                     partitionColumnCount: 0
-                    scratchColumnTypeNames: [bigint, double, double, 
decimal(11,2), decimal(9,0), decimal(9,0), decimal(9,0), double, double, 
double, decimal(13,4), double, double, double, double, double, decimal(12,4), 
double, double, double, double, double, double, double, decimal(12,4), double]
+                    scratchColumnTypeNames: [bigint, decimal(12,4), double, 
double, decimal(11,2), decimal(9,0), decimal(9,0), decimal(9,0), double, 
double, double, decimal(13,4)/DECIMAL_64, double, double, double, double, 
double, decimal(12,4), double, double, double, double, double, double, double, 
decimal(12,4), double, decimal(13,4)]
 
   Stage: Stage-0
     Fetch Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/dd512593/ql/src/test/results/clientpositive/llap/vector_decimal_precision.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_decimal_precision.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_precision.q.out
index 50e4305..5e7e8ca 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_precision.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_precision.q.out
@@ -612,8 +612,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false
@@ -1232,8 +1232,7 @@ STAGE PLANS:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vector.serde.deserialize IS true
                 inputFormatFeatureSupport: [DECIMAL_64]
-                vectorizationSupportRemovedReasons: [DECIMAL_64 disabled 
because LLAP is enabled]
-                featureSupportInUse: []
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: org.apache.hadoop.mapred.TextInputFormat
                 allNative: false
                 usesVectorUDFAdaptor: false

http://git-wip-us.apache.org/repos/asf/hive/blob/dd512593/ql/src/test/results/clientpositive/llap/vector_decimal_round.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_decimal_round.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_round.q.out
index 6737052..eb4a588 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_round.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_round.q.out
@@ -59,15 +59,15 @@ STAGE PLANS:
                   Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE 
Column stats: COMPLETE
                   TableScan Vectorization:
                       native: true
-                      vectorizationSchemaColumns: [0:dec:decimal(10,0), 
1:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
+                      vectorizationSchemaColumns: 
[0:dec:decimal(10,0)/DECIMAL_64, 
1:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
                   Select Operator
                     expressions: dec (type: decimal(10,0)), round(dec, -1) 
(type: decimal(11,0))
                     outputColumnNames: _col0, _col1
                     Select Vectorization:
                         className: VectorSelectOperator
                         native: true
-                        projectedOutputColumnNums: [0, 2]
-                        selectExpressions: 
FuncRoundWithNumDigitsDecimalToDecimal(col 0:decimal(10,0), decimalPlaces -1) 
-> 2:decimal(11,0)
+                        projectedOutputColumnNums: [0, 3]
+                        selectExpressions: 
FuncRoundWithNumDigitsDecimalToDecimal(col 2:decimal(10,0), decimalPlaces 
-1)(children: ConvertDecimal64ToDecimal(col 0:decimal(10,0)/DECIMAL_64) -> 
2:decimal(10,0)) -> 3:decimal(11,0)
                     Statistics: Num rows: 1 Data size: 224 Basic stats: 
COMPLETE Column stats: COMPLETE
                     Reduce Output Operator
                       key expressions: _col0 (type: decimal(10,0))
@@ -77,7 +77,7 @@ STAGE PLANS:
                           keyColumnNums: [0]
                           native: true
                           nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
-                          valueColumnNums: [2]
+                          valueColumnNums: [3]
                       Statistics: Num rows: 1 Data size: 224 Basic stats: 
COMPLETE Column stats: COMPLETE
                       value expressions: _col1 (type: decimal(11,0))
             Execution mode: vectorized, llap
@@ -86,8 +86,7 @@ STAGE PLANS:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vector.serde.deserialize IS true
                 inputFormatFeatureSupport: [DECIMAL_64]
-                vectorizationSupportRemovedReasons: [DECIMAL_64 disabled 
because LLAP is enabled]
-                featureSupportInUse: []
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: org.apache.hadoop.mapred.TextInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -95,9 +94,9 @@ STAGE PLANS:
                 rowBatchContext:
                     dataColumnCount: 1
                     includeColumns: [0]
-                    dataColumns: dec:decimal(10,0)
+                    dataColumns: dec:decimal(10,0)/DECIMAL_64
                     partitionColumnCount: 0
-                    scratchColumnTypeNames: [decimal(11,0)]
+                    scratchColumnTypeNames: [decimal(10,0), decimal(11,0)]
         Reducer 2 
             Execution mode: vectorized, llap
             Reduce Vectorization:
@@ -177,22 +176,22 @@ STAGE PLANS:
                   Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE 
Column stats: COMPLETE
                   TableScan Vectorization:
                       native: true
-                      vectorizationSchemaColumns: [0:dec:decimal(10,0), 
1:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
+                      vectorizationSchemaColumns: 
[0:dec:decimal(10,0)/DECIMAL_64, 
1:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
                   Select Operator
                     expressions: dec (type: decimal(10,0)), round(dec, -1) 
(type: decimal(11,0))
                     outputColumnNames: _col0, _col2
                     Select Vectorization:
                         className: VectorSelectOperator
                         native: true
-                        projectedOutputColumnNums: [0, 2]
-                        selectExpressions: 
FuncRoundWithNumDigitsDecimalToDecimal(col 0:decimal(10,0), decimalPlaces -1) 
-> 2:decimal(11,0)
+                        projectedOutputColumnNums: [0, 3]
+                        selectExpressions: 
FuncRoundWithNumDigitsDecimalToDecimal(col 2:decimal(10,0), decimalPlaces 
-1)(children: ConvertDecimal64ToDecimal(col 0:decimal(10,0)/DECIMAL_64) -> 
2:decimal(10,0)) -> 3:decimal(11,0)
                     Statistics: Num rows: 1 Data size: 336 Basic stats: 
COMPLETE Column stats: COMPLETE
                     Reduce Output Operator
                       key expressions: _col2 (type: decimal(11,0))
                       sort order: +
                       Reduce Sink Vectorization:
                           className: VectorReduceSinkObjectHashOperator
-                          keyColumnNums: [2]
+                          keyColumnNums: [3]
                           native: true
                           nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
                           valueColumnNums: [0]
@@ -204,8 +203,7 @@ STAGE PLANS:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vector.serde.deserialize IS true
                 inputFormatFeatureSupport: [DECIMAL_64]
-                vectorizationSupportRemovedReasons: [DECIMAL_64 disabled 
because LLAP is enabled]
-                featureSupportInUse: []
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: org.apache.hadoop.mapred.TextInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -213,9 +211,9 @@ STAGE PLANS:
                 rowBatchContext:
                     dataColumnCount: 1
                     includeColumns: [0]
-                    dataColumns: dec:decimal(10,0)
+                    dataColumns: dec:decimal(10,0)/DECIMAL_64
                     partitionColumnCount: 0
-                    scratchColumnTypeNames: [decimal(11,0)]
+                    scratchColumnTypeNames: [decimal(10,0), decimal(11,0)]
         Reducer 2 
             Execution mode: vectorized, llap
             Reduce Vectorization:
@@ -585,15 +583,15 @@ STAGE PLANS:
                   Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE 
Column stats: COMPLETE
                   TableScan Vectorization:
                       native: true
-                      vectorizationSchemaColumns: [0:dec:decimal(10,0), 
1:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
+                      vectorizationSchemaColumns: 
[0:dec:decimal(10,0)/DECIMAL_64, 
1:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
                   Select Operator
                     expressions: dec (type: decimal(10,0)), round(dec, -1) 
(type: decimal(11,0))
                     outputColumnNames: _col0, _col1
                     Select Vectorization:
                         className: VectorSelectOperator
                         native: true
-                        projectedOutputColumnNums: [0, 2]
-                        selectExpressions: 
FuncRoundWithNumDigitsDecimalToDecimal(col 0:decimal(10,0), decimalPlaces -1) 
-> 2:decimal(11,0)
+                        projectedOutputColumnNums: [0, 3]
+                        selectExpressions: 
FuncRoundWithNumDigitsDecimalToDecimal(col 2:decimal(10,0), decimalPlaces 
-1)(children: ConvertDecimal64ToDecimal(col 0:decimal(10,0)/DECIMAL_64) -> 
2:decimal(10,0)) -> 3:decimal(11,0)
                     Statistics: Num rows: 1 Data size: 224 Basic stats: 
COMPLETE Column stats: COMPLETE
                     Reduce Output Operator
                       key expressions: _col0 (type: decimal(10,0))
@@ -603,7 +601,7 @@ STAGE PLANS:
                           keyColumnNums: [0]
                           native: true
                           nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
-                          valueColumnNums: [2]
+                          valueColumnNums: [3]
                       Statistics: Num rows: 1 Data size: 224 Basic stats: 
COMPLETE Column stats: COMPLETE
                       value expressions: _col1 (type: decimal(11,0))
             Execution mode: vectorized, llap
@@ -611,8 +609,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -620,9 +618,9 @@ STAGE PLANS:
                 rowBatchContext:
                     dataColumnCount: 1
                     includeColumns: [0]
-                    dataColumns: dec:decimal(10,0)
+                    dataColumns: dec:decimal(10,0)/DECIMAL_64
                     partitionColumnCount: 0
-                    scratchColumnTypeNames: [decimal(11,0)]
+                    scratchColumnTypeNames: [decimal(10,0), decimal(11,0)]
         Reducer 2 
             Execution mode: vectorized, llap
             Reduce Vectorization:
@@ -702,22 +700,22 @@ STAGE PLANS:
                   Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE 
Column stats: COMPLETE
                   TableScan Vectorization:
                       native: true
-                      vectorizationSchemaColumns: [0:dec:decimal(10,0), 
1:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
+                      vectorizationSchemaColumns: 
[0:dec:decimal(10,0)/DECIMAL_64, 
1:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
                   Select Operator
                     expressions: dec (type: decimal(10,0)), round(dec, -1) 
(type: decimal(11,0))
                     outputColumnNames: _col0, _col2
                     Select Vectorization:
                         className: VectorSelectOperator
                         native: true
-                        projectedOutputColumnNums: [0, 2]
-                        selectExpressions: 
FuncRoundWithNumDigitsDecimalToDecimal(col 0:decimal(10,0), decimalPlaces -1) 
-> 2:decimal(11,0)
+                        projectedOutputColumnNums: [0, 3]
+                        selectExpressions: 
FuncRoundWithNumDigitsDecimalToDecimal(col 2:decimal(10,0), decimalPlaces 
-1)(children: ConvertDecimal64ToDecimal(col 0:decimal(10,0)/DECIMAL_64) -> 
2:decimal(10,0)) -> 3:decimal(11,0)
                     Statistics: Num rows: 1 Data size: 336 Basic stats: 
COMPLETE Column stats: COMPLETE
                     Reduce Output Operator
                       key expressions: _col2 (type: decimal(11,0))
                       sort order: +
                       Reduce Sink Vectorization:
                           className: VectorReduceSinkObjectHashOperator
-                          keyColumnNums: [2]
+                          keyColumnNums: [3]
                           native: true
                           nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
                           valueColumnNums: [0]
@@ -728,8 +726,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -737,9 +735,9 @@ STAGE PLANS:
                 rowBatchContext:
                     dataColumnCount: 1
                     includeColumns: [0]
-                    dataColumns: dec:decimal(10,0)
+                    dataColumns: dec:decimal(10,0)/DECIMAL_64
                     partitionColumnCount: 0
-                    scratchColumnTypeNames: [decimal(11,0)]
+                    scratchColumnTypeNames: [decimal(10,0), decimal(11,0)]
         Reducer 2 
             Execution mode: vectorized, llap
             Reduce Vectorization:

http://git-wip-us.apache.org/repos/asf/hive/blob/dd512593/ql/src/test/results/clientpositive/llap/vector_decimal_round_2.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_decimal_round_2.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_round_2.q.out
index e3d4f40..bb0cbfc 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_round_2.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_round_2.q.out
@@ -89,8 +89,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -257,8 +257,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -452,8 +452,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -636,8 +636,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false

http://git-wip-us.apache.org/repos/asf/hive/blob/dd512593/ql/src/test/results/clientpositive/llap/vector_decimal_trailing.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_decimal_trailing.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_trailing.q.out
index b33f090..a35f6fe 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_trailing.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_trailing.q.out
@@ -94,7 +94,7 @@ STAGE PLANS:
                   Statistics: Num rows: 30 Data size: 4936 Basic stats: 
COMPLETE Column stats: COMPLETE
                   TableScan Vectorization:
                       native: true
-                      vectorizationSchemaColumns: [0:id:int, 
1:a:decimal(10,4), 2:b:decimal(15,8), 
3:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
+                      vectorizationSchemaColumns: [0:id:int, 
1:a:decimal(10,4)/DECIMAL_64, 2:b:decimal(15,8)/DECIMAL_64, 
3:ROW__ID:struct<writeid:bigint,bucketid:int,rowid:bigint>]
                   Select Operator
                     expressions: id (type: int), a (type: decimal(10,4)), b 
(type: decimal(15,8))
                     outputColumnNames: _col0, _col1, _col2
@@ -119,8 +119,8 @@ STAGE PLANS:
             Map Vectorization:
                 enabled: true
                 enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
+                inputFormatFeatureSupport: [DECIMAL_64]
+                featureSupportInUse: [DECIMAL_64]
                 inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                 allNative: true
                 usesVectorUDFAdaptor: false
@@ -128,7 +128,7 @@ STAGE PLANS:
                 rowBatchContext:
                     dataColumnCount: 3
                     includeColumns: [0, 1, 2]
-                    dataColumns: id:int, a:decimal(10,4), b:decimal(15,8)
+                    dataColumns: id:int, a:decimal(10,4)/DECIMAL_64, 
b:decimal(15,8)/DECIMAL_64
                     partitionColumnCount: 0
                     scratchColumnTypeNames: []
         Reducer 2 

Reply via email to