http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinAddNotNullRule.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinAddNotNullRule.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinAddNotNullRule.java
index c4a40bf..2a415d5 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinAddNotNullRule.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinAddNotNullRule.java
@@ -47,7 +47,6 @@ import 
org.apache.hadoop.hive.ql.optimizer.calcite.reloperators.HiveFilter;
 import 
org.apache.hadoop.hive.ql.optimizer.calcite.translator.SqlFunctionConverter;
 import org.apache.hadoop.hive.ql.parse.SemanticException;
 
-import com.esotericsoftware.minlog.Log;
 import com.google.common.collect.ImmutableList;
 
 public final class HiveJoinAddNotNullRule extends RelOptRule {
@@ -92,7 +91,6 @@ public final class HiveJoinAddNotNullRule extends RelOptRule {
     try {
       joinPredInfo = 
HiveCalciteUtil.JoinPredicateInfo.constructJoinPredicateInfo(join);
     } catch (CalciteSemanticException e) {
-      Log.trace("Failed to add is not null filter on join ", e);
       return;
     }
 
@@ -183,7 +181,7 @@ public final class HiveJoinAddNotNullRule extends 
RelOptRule {
     }
     return newConditions;
   }
-  
+
   private static Map<String,RexNode> splitCondition(RexNode condition) {
     Map<String,RexNode> newConditions = new HashMap<String,RexNode>();
     if (condition.getKind() == SqlKind.AND) {
@@ -196,10 +194,10 @@ public final class HiveJoinAddNotNullRule extends 
RelOptRule {
     }
     return newConditions;
   }
-  
+
   private static RelNode createHiveFilterConjunctiveCondition(FilterFactory 
filterFactory,
           RexBuilder rexBuilder, RelNode input, Collection<RexNode> 
conditions) {
     final RexNode newCondition = RexUtil.composeConjunction(rexBuilder, 
conditions, false);
     return filterFactory.createFilter(input, newCondition);
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinToMultiJoinRule.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinToMultiJoinRule.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinToMultiJoinRule.java
index 35dbda9..a8b16cc 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinToMultiJoinRule.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinToMultiJoinRule.java
@@ -35,8 +35,8 @@ import org.apache.calcite.rex.RexNode;
 import org.apache.calcite.rex.RexUtil;
 import org.apache.calcite.util.ImmutableBitSet;
 import org.apache.calcite.util.Pair;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.optimizer.calcite.CalciteSemanticException;
 import org.apache.hadoop.hive.ql.optimizer.calcite.HiveCalciteUtil;
 import 
org.apache.hadoop.hive.ql.optimizer.calcite.HiveCalciteUtil.JoinPredicateInfo;
@@ -59,7 +59,7 @@ public class HiveJoinToMultiJoinRule extends RelOptRule {
 
   private final ProjectFactory projectFactory;
 
-  private static transient final Log LOG = 
LogFactory.getLog(HiveJoinToMultiJoinRule.class);
+  private static transient final Logger LOG = 
LoggerFactory.getLogger(HiveJoinToMultiJoinRule.class);
 
   //~ Constructors -----------------------------------------------------------
 

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePreFilteringRule.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePreFilteringRule.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePreFilteringRule.java
index 5824127..82d9600 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePreFilteringRule.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePreFilteringRule.java
@@ -18,10 +18,8 @@
 package org.apache.hadoop.hive.ql.optimizer.calcite.rules;
 
 import java.util.ArrayList;
-import java.util.Collection;
 import java.util.EnumSet;
 import java.util.List;
-import java.util.Map.Entry;
 import java.util.Set;
 
 import org.apache.calcite.plan.RelOptPredicateList;
@@ -40,8 +38,8 @@ import org.apache.calcite.rex.RexLiteral;
 import org.apache.calcite.rex.RexNode;
 import org.apache.calcite.rex.RexUtil;
 import org.apache.calcite.sql.SqlKind;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.optimizer.calcite.reloperators.HiveFilter;
 
 import com.google.common.collect.ImmutableList;
@@ -52,8 +50,8 @@ import com.google.common.collect.Sets;
 
 public class HivePreFilteringRule extends RelOptRule {
 
-  protected static final Log LOG = LogFactory
-          .getLog(HivePreFilteringRule.class.getName());
+  protected static final Logger LOG = LoggerFactory
+          .getLogger(HivePreFilteringRule.class.getName());
 
 
   public static final HivePreFilteringRule INSTANCE =

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/PartitionPrune.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/PartitionPrune.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/PartitionPrune.java
index 9098c64..0e5e2b9 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/PartitionPrune.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/PartitionPrune.java
@@ -24,8 +24,8 @@ import java.util.List;
 import java.util.Set;
 import java.io.IOException;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.calcite.plan.RelOptCluster;
 import org.apache.calcite.rel.type.RelDataType;
 import org.apache.calcite.rel.type.RelDataTypeField;
@@ -67,7 +67,7 @@ public class PartitionPrune {
 
   public static class ExtractPartPruningPredicate extends
       RexVisitorImpl<RexNode> {
-    private static final Log LOG = 
LogFactory.getLog(ExtractPartPruningPredicate.class);
+    private static final Logger LOG = 
LoggerFactory.getLogger(ExtractPartPruningPredicate.class);
     final RelOptHiveTable hiveTable;
     final RelDataType rType;
     final Set<String> partCols;

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/stats/HiveRelMdRowCount.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/stats/HiveRelMdRowCount.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/stats/HiveRelMdRowCount.java
index 6948fb1..728c5aa 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/stats/HiveRelMdRowCount.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/stats/HiveRelMdRowCount.java
@@ -44,13 +44,13 @@ import org.apache.calcite.sql.fun.SqlStdOperatorTable;
 import org.apache.calcite.util.BuiltInMethod;
 import org.apache.calcite.util.ImmutableBitSet;
 import org.apache.calcite.util.Pair;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.optimizer.calcite.reloperators.HiveTableScan;
 
 public class HiveRelMdRowCount extends RelMdRowCount {
 
-  protected static final Log LOG  = 
LogFactory.getLog(HiveRelMdRowCount.class.getName());
+  protected static final Logger LOG  = 
LoggerFactory.getLogger(HiveRelMdRowCount.class.getName());
 
 
   public static final RelMetadataProvider SOURCE = 
ReflectiveRelMetadataProvider
@@ -66,24 +66,21 @@ public class HiveRelMdRowCount extends RelMdRowCount {
       double selectivity = (pkfk.pkInfo.selectivity * pkfk.ndvScalingFactor);
       selectivity = Math.min(1.0, selectivity);
       if (LOG.isDebugEnabled()) {
-        LOG.debug("Identified Primary - Foreign Key relation:");
-        LOG.debug(RelOptUtil.toString(join));
-        LOG.debug(pkfk);
+        LOG.debug("Identified Primary - Foreign Key relation: {} 
{}",RelOptUtil.toString(join), pkfk);
       }
       return pkfk.fkInfo.rowCount * selectivity;
     }
     return join.getRows();
   }
 
+  @Override
   public Double getRowCount(SemiJoin rel) {
     PKFKRelationInfo pkfk = analyzeJoinForPKFK(rel);
     if (pkfk != null) {
       double selectivity = (pkfk.pkInfo.selectivity * pkfk.ndvScalingFactor);
       selectivity = Math.min(1.0, selectivity);
       if (LOG.isDebugEnabled()) {
-        LOG.debug("Identified Primary - Foreign Key relation:");
-        LOG.debug(RelOptUtil.toString(rel));
-        LOG.debug(pkfk);
+        LOG.debug("Identified Primary - Foreign Key relation: {} {}", 
RelOptUtil.toString(rel), pkfk);
       }
       return pkfk.fkInfo.rowCount * selectivity;
     }
@@ -109,6 +106,7 @@ public class HiveRelMdRowCount extends RelMdRowCount {
       this.isPKSideSimple = isPKSideSimple;
     }
 
+    @Override
     public String toString() {
       return String.format(
           "Primary - Foreign Key join:\n\tfkSide = %d\n\tFKInfo:%s\n" +
@@ -129,6 +127,7 @@ public class HiveRelMdRowCount extends RelMdRowCount {
       this.distinctCount = distinctCount;
     }
 
+    @Override
     public String toString() {
       return String.format("FKInfo(rowCount=%.2f,ndv=%.2f)", rowCount, 
distinctCount);
     }
@@ -141,6 +140,7 @@ public class HiveRelMdRowCount extends RelMdRowCount {
       this.selectivity = selectivity;
     }
 
+    @Override
     public String toString() {
       return String.format("PKInfo(rowCount=%.2f,ndv=%.2f,selectivity=%.2f)", 
rowCount, distinctCount,selectivity);
     }
@@ -230,7 +230,7 @@ public class HiveRelMdRowCount extends RelMdRowCount {
 
     int pkSide = leftIsKey ? 0 : rightIsKey ? 1 : -1;
 
-    boolean isPKSideSimpleTree = pkSide != -1 ? 
+    boolean isPKSideSimpleTree = pkSide != -1 ?
         IsSimpleTreeOnJoinKey.check(
             pkSide == 0 ? left : right,
             pkSide == 0 ? leftColIdx : rightColIdx) : false;

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTConverter.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTConverter.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTConverter.java
index 14946b3..e4ac154 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTConverter.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTConverter.java
@@ -52,8 +52,8 @@ import org.apache.calcite.sql.SqlKind;
 import org.apache.calcite.sql.SqlOperator;
 import org.apache.calcite.sql.type.SqlTypeName;
 import org.apache.calcite.util.ImmutableBitSet;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.metastore.api.FieldSchema;
 import org.apache.hadoop.hive.ql.metadata.VirtualColumn;
 import org.apache.hadoop.hive.ql.optimizer.calcite.CalciteSemanticException;
@@ -68,7 +68,7 @@ import org.apache.hadoop.hive.ql.parse.ParseDriver;
 import com.google.common.collect.Iterables;
 
 public class ASTConverter {
-  private static final Log LOG = LogFactory.getLog(ASTConverter.class);
+  private static final Logger LOG = 
LoggerFactory.getLogger(ASTConverter.class);
 
   private final RelNode          root;
   private final HiveAST          hiveAST;

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ExprNodeConverter.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ExprNodeConverter.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ExprNodeConverter.java
index 42f1ab6..b42e78f 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ExprNodeConverter.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ExprNodeConverter.java
@@ -41,8 +41,8 @@ import org.apache.calcite.rex.RexWindow;
 import org.apache.calcite.rex.RexWindowBound;
 import org.apache.calcite.sql.SqlKind;
 import org.apache.calcite.sql.type.SqlTypeUtil;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.common.type.HiveDecimal;
 import org.apache.hadoop.hive.common.type.HiveIntervalDayTime;
 import org.apache.hadoop.hive.common.type.HiveIntervalYearMonth;
@@ -79,12 +79,11 @@ import com.google.common.collect.ImmutableSet;
 public class ExprNodeConverter extends RexVisitorImpl<ExprNodeDesc> {
 
   private final String             tabAlias;
-  private final String             columnAlias;
   private final RelDataType        inputRowType;
   private final ImmutableSet<Integer>       inputVCols;
-  private List<WindowFunctionSpec> windowFunctionSpecs = new ArrayList<>();
+  private final List<WindowFunctionSpec> windowFunctionSpecs = new 
ArrayList<>();
   private final RelDataTypeFactory dTFactory;
-  protected final Log LOG = LogFactory.getLog(this.getClass().getName());
+  protected final Logger LOG = 
LoggerFactory.getLogger(this.getClass().getName());
   private static long uniqueCounter = 0;
 
   public ExprNodeConverter(String tabAlias, RelDataType inputRowType,
@@ -96,7 +95,6 @@ public class ExprNodeConverter extends 
RexVisitorImpl<ExprNodeDesc> {
           RelDataType outputRowType, Set<Integer> inputVCols, 
RelDataTypeFactory dTFactory) {
     super(true);
     this.tabAlias = tabAlias;
-    this.columnAlias = columnAlias;
     this.inputRowType = inputRowType;
     this.inputVCols = ImmutableSet.copyOf(inputVCols);
     this.dTFactory = dTFactory;
@@ -146,7 +144,7 @@ public class ExprNodeConverter extends 
RexVisitorImpl<ExprNodeDesc> {
       try {
         gfDesc = ExprNodeGenericFuncDesc.newInstance(hiveUdf, args);
       } catch (UDFArgumentException e) {
-        LOG.error(e);
+        LOG.error("Failed to instantiate udf: ", e);
         throw new RuntimeException(e);
       }
     }

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverter.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverter.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverter.java
index f6c0114..f0f8aa8 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverter.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverter.java
@@ -40,8 +40,8 @@ import org.apache.calcite.rex.RexLiteral;
 import org.apache.calcite.rex.RexNode;
 import org.apache.calcite.util.ImmutableBitSet;
 import org.apache.calcite.util.Pair;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.ql.ErrorMsg;
 import org.apache.hadoop.hive.ql.exec.ColumnInfo;
@@ -105,7 +105,7 @@ import com.google.common.collect.Lists;
 
 public class HiveOpConverter {
 
-  private static final Log LOG = LogFactory.getLog(HiveOpConverter.class);
+  private static final Logger LOG = 
LoggerFactory.getLogger(HiveOpConverter.class);
 
   public static enum HIVEAGGOPMODE {
     NO_SKEW_NO_MAP_SIDE_AGG, // Corresponds to SemAnalyzer genGroupByPlan1MR

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverterPostProc.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverterPostProc.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverterPostProc.java
index 5080992..1d0a254 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverterPostProc.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverterPostProc.java
@@ -25,8 +25,8 @@ import java.util.Map;
 import java.util.Set;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.ql.exec.JoinOperator;
 import org.apache.hadoop.hive.ql.exec.Operator;
@@ -48,7 +48,7 @@ import org.apache.hadoop.hive.ql.plan.OperatorDesc;
 
 public class HiveOpConverterPostProc implements Transform {
 
-  private static final Log LOG = 
LogFactory.getLog(HiveOpConverterPostProc.class);
+  private static final Logger LOG = 
LoggerFactory.getLogger(HiveOpConverterPostProc.class);
 
   private ParseContext                                  pctx;
   private Map<String, Operator<? extends OperatorDesc>> aliasToOpInfo;

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierForASTConv.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierForASTConv.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierForASTConv.java
index b11cdfd..e820496 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierForASTConv.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierForASTConv.java
@@ -38,8 +38,8 @@ import org.apache.calcite.rel.type.RelDataTypeFactory;
 import org.apache.calcite.rex.RexNode;
 import org.apache.calcite.sql.SqlAggFunction;
 import org.apache.calcite.util.Pair;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.metastore.api.FieldSchema;
 import org.apache.hadoop.hive.ql.optimizer.calcite.CalciteSemanticException;
 import org.apache.hadoop.hive.ql.optimizer.calcite.HiveCalciteUtil;
@@ -53,7 +53,7 @@ import com.google.common.collect.ImmutableList;
 
 public class PlanModifierForASTConv {
 
-  private static final Log LOG = 
LogFactory.getLog(PlanModifierForASTConv.class);
+  private static final Logger LOG = 
LoggerFactory.getLogger(PlanModifierForASTConv.class);
 
 
   public static RelNode convertOpTree(RelNode rel, List<FieldSchema> 
resultSchema)

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierUtil.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierUtil.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierUtil.java
index 988d6d3..570eea2 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierUtil.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierUtil.java
@@ -31,8 +31,8 @@ import org.apache.calcite.rex.RexCall;
 import org.apache.calcite.rex.RexInputRef;
 import org.apache.calcite.rex.RexNode;
 import org.apache.calcite.util.Pair;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.metastore.api.FieldSchema;
 import org.apache.hadoop.hive.ql.optimizer.calcite.CalciteSemanticException;
 import org.apache.hadoop.hive.ql.optimizer.calcite.HiveCalciteUtil;
@@ -43,7 +43,7 @@ import com.google.common.collect.ImmutableMap;
 
 public class PlanModifierUtil {
 
-  private static final Log LOG = LogFactory.getLog(PlanModifierUtil.class);
+  private static final Logger LOG = 
LoggerFactory.getLogger(PlanModifierUtil.class);
 
 
   protected static void fixTopOBSchema(final RelNode rootRel,

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java
index 3d05161..d315497 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java
@@ -43,8 +43,8 @@ import org.apache.calcite.sql.SqlOperator;
 import org.apache.calcite.sql.fun.SqlCastFunction;
 import org.apache.calcite.sql.parser.SqlParserPos;
 import org.apache.calcite.sql.type.SqlTypeName;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.common.type.Decimal128;
 import org.apache.hadoop.hive.common.type.HiveChar;
 import org.apache.hadoop.hive.common.type.HiveDecimal;
@@ -88,7 +88,7 @@ import com.google.common.collect.ImmutableList.Builder;
 import com.google.common.collect.ImmutableMap;
 
 public class RexNodeConverter {
-  private static final Log LOG = LogFactory.getLog(RexNodeConverter.class);
+  private static final Logger LOG = 
LoggerFactory.getLogger(RexNodeConverter.class);
 
   private static class InputCtx {
     private final RelDataType                   calciteInpDataType;

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/SqlFunctionConverter.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/SqlFunctionConverter.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/SqlFunctionConverter.java
index d59c6bb..a17fb94 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/SqlFunctionConverter.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/SqlFunctionConverter.java
@@ -37,8 +37,8 @@ import org.apache.calcite.sql.type.SqlReturnTypeInference;
 import org.apache.calcite.sql.type.SqlTypeFamily;
 import org.apache.calcite.util.Util;
 import org.apache.commons.lang3.StringUtils;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.Description;
 import org.apache.hadoop.hive.ql.exec.FunctionInfo;
 import org.apache.hadoop.hive.ql.exec.FunctionRegistry;
@@ -70,7 +70,7 @@ import com.google.common.collect.ImmutableMap;
 import com.google.common.collect.Maps;
 
 public class SqlFunctionConverter {
-  private static final Log LOG = LogFactory.getLog(SqlFunctionConverter.class);
+  private static final Logger LOG = 
LoggerFactory.getLogger(SqlFunctionConverter.class);
 
   static final Map<String, SqlOperator>    hiveToCalcite;
   static final Map<SqlOperator, HiveToken> calciteToHiveToken;

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationOptimizer.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationOptimizer.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationOptimizer.java
index c1f1519..1b5d921 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationOptimizer.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationOptimizer.java
@@ -30,8 +30,8 @@ import java.util.Map;
 import java.util.Set;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.fs.ContentSummary;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
@@ -84,7 +84,7 @@ import org.apache.hadoop.hive.ql.plan.OperatorDesc;
  */
 public class CorrelationOptimizer implements Transform {
 
-  private static final Log LOG = 
LogFactory.getLog(CorrelationOptimizer.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(CorrelationOptimizer.class.getName());
 
   private boolean abort; // if correlation optimizer will not try to optimize 
this query
 

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/QueryPlanTreeTransformation.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/QueryPlanTreeTransformation.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/QueryPlanTreeTransformation.java
index e8ae2f7..315a650 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/QueryPlanTreeTransformation.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/QueryPlanTreeTransformation.java
@@ -25,8 +25,8 @@ import java.util.List;
 import java.util.Map;
 import java.util.Set;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.DemuxOperator;
 import org.apache.hadoop.hive.ql.exec.GroupByOperator;
 import org.apache.hadoop.hive.ql.exec.Operator;
@@ -48,7 +48,7 @@ import org.apache.hadoop.hive.ql.plan.TableDesc;
  * detected by Correlation Optimizer.
  */
 public class QueryPlanTreeTransformation {
-  private static final Log LOG = 
LogFactory.getLog(QueryPlanTreeTransformation.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(QueryPlanTreeTransformation.class.getName());
 
   private static void setNewTag(IntraQueryCorrelation correlation,
       List<Operator<? extends OperatorDesc>> childrenOfDemux,

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteCanApplyCtx.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteCanApplyCtx.java 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteCanApplyCtx.java
index 988bb29..55b85d9 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteCanApplyCtx.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteCanApplyCtx.java
@@ -25,8 +25,8 @@ import java.util.List;
 import java.util.Map;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.FilterOperator;
 import org.apache.hadoop.hive.ql.exec.GroupByOperator;
 import org.apache.hadoop.hive.ql.exec.ReduceSinkOperator;
@@ -51,7 +51,7 @@ import org.apache.hadoop.hive.ql.parse.SemanticException;
  */
 public final class RewriteCanApplyCtx implements NodeProcessorCtx {
 
-  private static final Log LOG = 
LogFactory.getLog(RewriteCanApplyCtx.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(RewriteCanApplyCtx.class.getName());
 
   private RewriteCanApplyCtx(ParseContext parseContext) {
     this.parseContext = parseContext;

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteGBUsingIndex.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteGBUsingIndex.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteGBUsingIndex.java
index 5afe21e..ea1ece6 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteGBUsingIndex.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteGBUsingIndex.java
@@ -26,8 +26,8 @@ import java.util.List;
 import java.util.Map;
 import java.util.Set;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.metastore.api.FieldSchema;
 import org.apache.hadoop.hive.metastore.api.Index;
@@ -93,7 +93,7 @@ public class RewriteGBUsingIndex implements Transform {
   private ParseContext parseContext;
   private Hive hiveDb;
   private HiveConf hiveConf;
-  private static final Log LOG = 
LogFactory.getLog(RewriteGBUsingIndex.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(RewriteGBUsingIndex.class.getName());
 
   /*
    * Stores the list of top TableScanOperator names for which the rewrite

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteParseContextGenerator.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteParseContextGenerator.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteParseContextGenerator.java
index 0c111bc..48105de 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteParseContextGenerator.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteParseContextGenerator.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.optimizer.index;
 
 import java.io.IOException;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.ql.Context;
 import org.apache.hadoop.hive.ql.exec.Operator;
@@ -48,7 +48,7 @@ import org.apache.hadoop.hive.ql.plan.OperatorDesc;
  */
 public final class RewriteParseContextGenerator {
 
-  private static final Log LOG = 
LogFactory.getLog(RewriteParseContextGenerator.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(RewriteParseContextGenerator.class.getName());
 
   /**
    * Parse the input {@link String} command and generate an operator tree.

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteQueryUsingAggregateIndexCtx.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteQueryUsingAggregateIndexCtx.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteQueryUsingAggregateIndexCtx.java
index 624ee7f..9acc7b7 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteQueryUsingAggregateIndexCtx.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteQueryUsingAggregateIndexCtx.java
@@ -24,8 +24,8 @@ import java.util.HashMap;
 import java.util.List;
 import java.util.Map;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hive.ql.exec.ColumnInfo;
 import org.apache.hadoop.hive.ql.exec.FunctionRegistry;
@@ -64,7 +64,7 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils;
  */
 
 public final class RewriteQueryUsingAggregateIndexCtx  implements 
NodeProcessorCtx {
-  private static final Log LOG = 
LogFactory.getLog(RewriteQueryUsingAggregateIndexCtx.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(RewriteQueryUsingAggregateIndexCtx.class.getName());
   private RewriteQueryUsingAggregateIndexCtx(ParseContext parseContext, Hive 
hiveDb,
       RewriteCanApplyCtx canApplyCtx) {
     this.parseContext = parseContext;

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java
index 0304196..28eade4 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java
@@ -17,8 +17,8 @@
  */
 package org.apache.hadoop.hive.ql.optimizer.listbucketingpruner;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.FilterOperator;
 import org.apache.hadoop.hive.ql.exec.TableScanOperator;
 import org.apache.hadoop.hive.ql.exec.UDFArgumentException;
@@ -40,7 +40,7 @@ import org.apache.hadoop.hive.ql.parse.SemanticException;
  *
  */
 public class LBPartitionProcFactory extends PrunerOperatorFactory {
-  static final Log LOG = 
LogFactory.getLog(ListBucketingPruner.class.getName());
+  static final Logger LOG = 
LoggerFactory.getLogger(ListBucketingPruner.class.getName());
 
   /**
    * Retrieve partitions for the filter. This is called only when

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPruner.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPruner.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPruner.java
index 9f12602..f399a20 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPruner.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPruner.java
@@ -24,8 +24,8 @@ import java.util.List;
 import java.util.Map;
 import java.util.Set;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hive.common.FileUtils;
 import org.apache.hadoop.hive.ql.lib.NodeProcessorCtx;
@@ -42,7 +42,7 @@ import org.apache.hadoop.hive.ql.plan.ExprNodeDesc;
  *
  */
 public class ListBucketingPruner implements Transform {
-  static final Log LOG = 
LogFactory.getLog(ListBucketingPruner.class.getName());
+  static final Logger LOG = 
LoggerFactory.getLogger(ListBucketingPruner.class.getName());
 
   /*
    * (non-Javadoc)

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PartitionConditionRemover.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PartitionConditionRemover.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PartitionConditionRemover.java
index cbed375..812ca51 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PartitionConditionRemover.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PartitionConditionRemover.java
@@ -23,8 +23,8 @@ import java.util.LinkedHashMap;
 import java.util.List;
 import java.util.Map;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.Operator;
 import org.apache.hadoop.hive.ql.exec.FilterOperator;
 import org.apache.hadoop.hive.ql.exec.TableScanOperator;
@@ -47,8 +47,8 @@ import org.apache.hadoop.hive.ql.parse.SemanticException;
 public class PartitionConditionRemover implements Transform {
 
   // The log
-  private static final Log LOG = LogFactory
-      .getLog("hive.ql.optimizer.pcr.PartitionConditionRemover");
+  private static final Logger LOG = LoggerFactory
+      .getLogger("hive.ql.optimizer.pcr.PartitionConditionRemover");
 
   /*
    * (non-Javadoc)

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrExprProcFactory.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrExprProcFactory.java 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrExprProcFactory.java
index 825938a..7cdc730 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrExprProcFactory.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrExprProcFactory.java
@@ -25,8 +25,8 @@ import java.util.List;
 import java.util.Map;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.FunctionRegistry;
 import org.apache.hadoop.hive.ql.lib.DefaultGraphWalker;
 import org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher;
@@ -61,7 +61,7 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory;
  */
 public final class PcrExprProcFactory {
 
-  public static final Log LOG = 
LogFactory.getLog(PcrExprProcFactory.class.getName());
+  public static final Logger LOG = 
LoggerFactory.getLogger(PcrExprProcFactory.class.getName());
 
   static Object evalExprWithPart(ExprNodeDesc expr, Partition p, 
List<VirtualColumn> vcs)
       throws SemanticException {

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java
index 8955cbd..65505b3 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java
@@ -21,8 +21,8 @@ package org.apache.hadoop.hive.ql.optimizer.pcr;
 import java.util.ArrayList;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.FilterOperator;
 import org.apache.hadoop.hive.ql.exec.Operator;
 import org.apache.hadoop.hive.ql.exec.TableScanOperator;
@@ -49,8 +49,8 @@ import org.apache.hadoop.hive.ql.plan.OperatorDesc;
 public final class PcrOpProcFactory {
 
   // The log
-  private static final Log LOG = LogFactory
-      .getLog("hive.ql.optimizer.pcr.OpProcFactory");
+  private static final Logger LOG = LoggerFactory
+      .getLogger("hive.ql.optimizer.pcr.OpProcFactory");
 
   /**
    * Remove partition condition in a filter operator when possible. This is

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/CrossProductCheck.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/CrossProductCheck.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/CrossProductCheck.java
index 6bdb0a7..9ad33fd 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/CrossProductCheck.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/CrossProductCheck.java
@@ -28,8 +28,8 @@ import java.util.List;
 import java.util.Map;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.AbstractMapJoinOperator;
 import org.apache.hadoop.hive.ql.exec.ConditionalTask;
 import org.apache.hadoop.hive.ql.exec.CommonMergeJoinOperator;
@@ -87,8 +87,8 @@ import org.apache.hadoop.hive.ql.session.SessionState;
  */
 public class CrossProductCheck implements PhysicalPlanResolver, Dispatcher {
 
-  protected static transient final Log LOG = LogFactory
-      .getLog(CrossProductCheck.class);
+  protected static transient final Logger LOG = LoggerFactory
+      .getLogger(CrossProductCheck.class);
 
   @Override
   public PhysicalContext resolve(PhysicalContext pctx) throws 
SemanticException {

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenSparkSkewJoinProcessor.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenSparkSkewJoinProcessor.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenSparkSkewJoinProcessor.java
index f88fd0a..895e64e 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenSparkSkewJoinProcessor.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenSparkSkewJoinProcessor.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.optimizer.physical;
 
 import com.google.common.base.Preconditions;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.ql.exec.ColumnInfo;
@@ -74,7 +74,7 @@ import java.util.Map;
  *
  */
 public class GenSparkSkewJoinProcessor {
-  private static final Log LOG = 
LogFactory.getLog(GenSparkSkewJoinProcessor.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(GenSparkSkewJoinProcessor.class.getName());
 
   private GenSparkSkewJoinProcessor() {
     // prevent instantiation

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LlapDecider.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LlapDecider.java 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LlapDecider.java
index 9a8cb0f..af6129a 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LlapDecider.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LlapDecider.java
@@ -35,8 +35,8 @@ import java.util.List;
 import java.util.Map;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.ql.exec.FilterOperator;
 import org.apache.hadoop.hive.ql.exec.FunctionInfo;
@@ -84,8 +84,8 @@ import org.apache.hadoop.hive.ql.plan.TezWork;
  */
 public class LlapDecider implements PhysicalPlanResolver {
 
-  protected static transient final Log LOG
-    = LogFactory.getLog(LlapDecider.class);
+  protected static transient final Logger LOG
+    = LoggerFactory.getLogger(LlapDecider.class);
 
   private PhysicalContext physicalContext;
 

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LocalMapJoinProcFactory.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LocalMapJoinProcFactory.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LocalMapJoinProcFactory.java
index 9076d48..df598e7 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LocalMapJoinProcFactory.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LocalMapJoinProcFactory.java
@@ -24,8 +24,8 @@ import java.util.List;
 import java.util.Map;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.ql.exec.GroupByOperator;
 import org.apache.hadoop.hive.ql.exec.HashTableDummyOperator;
@@ -62,7 +62,7 @@ import org.apache.hadoop.hive.ql.plan.TableDesc;
  * OOM in group by operator.
  */
 public final class LocalMapJoinProcFactory {
-  private static final Log LOG = 
LogFactory.getLog(LocalMapJoinProcFactory.class);
+  private static final Logger LOG = 
LoggerFactory.getLogger(LocalMapJoinProcFactory.class);
 
   public static NodeProcessor getJoinProc() {
     return new LocalMapJoinProcessor();

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MemoryDecider.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MemoryDecider.java 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MemoryDecider.java
index eb8597d..9d67722 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MemoryDecider.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MemoryDecider.java
@@ -15,8 +15,8 @@ import java.util.SortedSet;
 import java.util.Stack;
 import java.util.TreeSet;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.ql.exec.MapJoinOperator;
 import org.apache.hadoop.hive.ql.exec.ReduceSinkOperator;
@@ -53,7 +53,7 @@ import org.apache.hadoop.hive.ql.plan.TezWork;
  */
 public class MemoryDecider implements PhysicalPlanResolver {
 
-  protected static transient final Log LOG = 
LogFactory.getLog(MemoryDecider.class);
+  protected static transient final Logger LOG = 
LoggerFactory.getLogger(MemoryDecider.class);
 
   public class MemoryCalculator implements Dispatcher {
 

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MetadataOnlyOptimizer.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MetadataOnlyOptimizer.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MetadataOnlyOptimizer.java
index 24610d9..d47d3c2 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MetadataOnlyOptimizer.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MetadataOnlyOptimizer.java
@@ -24,8 +24,8 @@ import java.util.List;
 import java.util.Map;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.FileSinkOperator;
 import org.apache.hadoop.hive.ql.exec.GroupByOperator;
 import org.apache.hadoop.hive.ql.exec.TableScanOperator;
@@ -55,7 +55,7 @@ import org.apache.hadoop.hive.ql.plan.TableScanDesc;
  *
  */
 public class MetadataOnlyOptimizer implements PhysicalPlanResolver {
-  static final Log LOG = 
LogFactory.getLog(MetadataOnlyOptimizer.class.getName());
+  static final Logger LOG = 
LoggerFactory.getLogger(MetadataOnlyOptimizer.class.getName());
 
   static class WalkerCtx implements NodeProcessorCtx {
     /* operators for which there is chance the optimization can be applied */

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanOptimizer.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanOptimizer.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanOptimizer.java
index 080a0e6..200e098 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanOptimizer.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanOptimizer.java
@@ -23,8 +23,8 @@ import java.util.LinkedHashMap;
 import java.util.Map;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.FilterOperator;
 import org.apache.hadoop.hive.ql.exec.LimitOperator;
 import org.apache.hadoop.hive.ql.exec.TableScanOperator;
@@ -49,7 +49,7 @@ import org.apache.hadoop.hive.ql.plan.ExprNodeDesc;
  */
 public class NullScanOptimizer implements PhysicalPlanResolver {
 
-  private static final Log LOG = 
LogFactory.getLog(NullScanOptimizer.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(NullScanOptimizer.class.getName());
   @Override
   public PhysicalContext resolve(PhysicalContext pctx) throws 
SemanticException {
 

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java
index 8dcd283..2e924fb 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java
@@ -30,8 +30,8 @@ import java.util.Map;
 import java.util.Map.Entry;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hive.ql.exec.Operator;
 import org.apache.hadoop.hive.ql.exec.TableScanOperator;
@@ -60,7 +60,7 @@ import org.apache.hadoop.hive.serde2.NullStructSerDe;
  */
 public class NullScanTaskDispatcher implements Dispatcher {
 
-  static final Log LOG = 
LogFactory.getLog(NullScanTaskDispatcher.class.getName());
+  static final Logger LOG = 
LoggerFactory.getLogger(NullScanTaskDispatcher.class.getName());
 
   private final PhysicalContext physicalContext;
   private final Map<Rule, NodeProcessor> rules;
@@ -203,4 +203,4 @@ public class NullScanTaskDispatcher implements Dispatcher {
     }
     return null;
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SerializeFilter.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SerializeFilter.java 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SerializeFilter.java
index f3c1d42..64e7ba2 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SerializeFilter.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SerializeFilter.java
@@ -15,8 +15,8 @@ import java.util.SortedSet;
 import java.util.Stack;
 import java.util.TreeSet;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.ql.exec.Operator;
 import org.apache.hadoop.hive.ql.exec.StatsTask;
@@ -49,7 +49,7 @@ import org.apache.hadoop.hive.ql.plan.TezWork;
  */
 public class SerializeFilter implements PhysicalPlanResolver {
 
-  protected static transient final Log LOG = 
LogFactory.getLog(SerializeFilter.class);
+  protected static transient final Logger LOG = 
LoggerFactory.getLogger(SerializeFilter.class);
 
   public class Serializer implements Dispatcher {
 

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
index dc283e8..4dead18 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
@@ -31,8 +31,8 @@ import java.util.Set;
 import java.util.Stack;
 import java.util.regex.Pattern;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.metastore.api.hive_metastoreConstants;
 import org.apache.hadoop.hive.ql.exec.*;
@@ -151,7 +151,7 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils;
 
 public class Vectorizer implements PhysicalPlanResolver {
 
-  protected static transient final Log LOG = 
LogFactory.getLog(Vectorizer.class);
+  protected static transient final Logger LOG = 
LoggerFactory.getLogger(Vectorizer.class);
 
   Pattern supportedDataTypesPattern;
   List<Task<? extends Serializable>> vectorizableTasks =

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcCtx.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcCtx.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcCtx.java
index a115c67..adfc96e 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcCtx.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcCtx.java
@@ -20,15 +20,15 @@ package org.apache.hadoop.hive.ql.optimizer.physical.index;
 
 import java.io.Serializable;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.Task;
 import org.apache.hadoop.hive.ql.lib.NodeProcessorCtx;
 import org.apache.hadoop.hive.ql.parse.ParseContext;
 
 public class IndexWhereProcCtx implements NodeProcessorCtx {
 
-  private static final Log LOG = 
LogFactory.getLog(IndexWhereProcCtx.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(IndexWhereProcCtx.class.getName());
 
   private final Task<? extends Serializable> currentTask;
   private final ParseContext parseCtx;

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcessor.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcessor.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcessor.java
index 61ff61e..81e99fc 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcessor.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcessor.java
@@ -27,8 +27,8 @@ import java.util.Map;
 import java.util.Set;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.fs.ContentSummary;
 import org.apache.hadoop.hive.metastore.api.Index;
 import org.apache.hadoop.hive.ql.exec.TableScanOperator;
@@ -62,7 +62,7 @@ import org.apache.hadoop.hive.ql.plan.TableScanDesc;
 */
 public class IndexWhereProcessor implements NodeProcessor {
 
-  private static final Log LOG = 
LogFactory.getLog(IndexWhereProcessor.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(IndexWhereProcessor.class.getName());
   private final Map<TableScanOperator, List<Index>> tsToIndices;
 
   public IndexWhereProcessor(Map<TableScanOperator, List<Index>> tsToIndices) {

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionExpressionForMetastore.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionExpressionForMetastore.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionExpressionForMetastore.java
index b76229c..f9978b4 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionExpressionForMetastore.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionExpressionForMetastore.java
@@ -22,8 +22,8 @@ import java.io.IOException;
 import java.nio.ByteBuffer;
 import java.util.List;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.metastore.Metastore.SplitInfo;
 import org.apache.hadoop.hive.metastore.Metastore.SplitInfos;
 import org.apache.hadoop.hive.metastore.PartitionExpressionProxy;
@@ -43,7 +43,7 @@ import 
org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo;
  * The basic implementation of PartitionExpressionProxy that uses ql package 
classes.
  */
 public class PartitionExpressionForMetastore implements 
PartitionExpressionProxy {
-  private static final Log LOG = 
LogFactory.getLog(PartitionExpressionForMetastore.class);
+  private static final Logger LOG = 
LoggerFactory.getLogger(PartitionExpressionForMetastore.class);
 
   @Override
   public String convertExprToFilter(byte[] exprBytes) throws MetaException {

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java
index 5644662..d99362a 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java
@@ -29,8 +29,8 @@ import java.util.Set;
 
 import com.google.common.annotations.VisibleForTesting;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.common.ObjectPair;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.metastore.IMetaStoreClient;
@@ -74,7 +74,7 @@ public class PartitionPruner implements Transform {
 
   // The log
   public static final String CLASS_NAME = PartitionPruner.class.getName();
-  public static final Log LOG = LogFactory.getLog(CLASS_NAME);
+  public static final Logger LOG = LoggerFactory.getLogger(CLASS_NAME);
 
   /*
    * (non-Javadoc)

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java
index 9c4c25e..41e9ba6 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java
@@ -29,8 +29,8 @@ import java.util.Stack;
 
 import com.google.common.collect.Maps;
 import com.google.common.collect.Sets;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.MapJoinOperator;
 import org.apache.hadoop.hive.ql.exec.Operator;
 import org.apache.hadoop.hive.ql.exec.spark.SparkTask;
@@ -53,7 +53,7 @@ import org.apache.hadoop.hive.ql.plan.SparkWork;
  * works.
  */
 public class CombineEquivalentWorkResolver implements PhysicalPlanResolver {
-  protected static transient Log LOG = 
LogFactory.getLog(CombineEquivalentWorkResolver.class);
+  protected static transient Logger LOG = 
LoggerFactory.getLogger(CombineEquivalentWorkResolver.class);
 
   @Override
   public PhysicalContext resolve(PhysicalContext pctx) throws 
SemanticException {

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java
index 5f9225c..ff4924d 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java
@@ -21,8 +21,8 @@ package org.apache.hadoop.hive.ql.optimizer.spark;
 import java.util.List;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.common.ObjectPair;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.metastore.api.hive_metastoreConstants;
@@ -51,7 +51,7 @@ import org.apache.hadoop.hive.ql.plan.ReduceSinkDesc;
  */
 public class SetSparkReducerParallelism implements NodeProcessor {
 
-  private static final Log LOG = 
LogFactory.getLog(SetSparkReducerParallelism.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(SetSparkReducerParallelism.class.getName());
 
   // Spark memory per task, and total number of cores
   private ObjectPair<Long, Integer> sparkMemoryAndCores;

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java
index 39d1f18..4d2b8d6 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java
@@ -24,8 +24,8 @@ import java.util.Map;
 import java.util.Set;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.ql.exec.GroupByOperator;
 import org.apache.hadoop.hive.ql.exec.JoinOperator;
@@ -56,7 +56,7 @@ import org.apache.hadoop.hive.ql.plan.Statistics;
  */
 public class SparkMapJoinOptimizer implements NodeProcessor {
 
-  private static final Log LOG = 
LogFactory.getLog(SparkMapJoinOptimizer.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(SparkMapJoinOptimizer.class.getName());
 
   @Override
   /**

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkReduceSinkMapJoinProc.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkReduceSinkMapJoinProc.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkReduceSinkMapJoinProc.java
index 76517e4..5b3125b 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkReduceSinkMapJoinProc.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkReduceSinkMapJoinProc.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.optimizer.spark;
 
 import com.google.common.base.Preconditions;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.ql.exec.GroupByOperator;
 import org.apache.hadoop.hive.ql.exec.HashTableDummyOperator;
@@ -62,7 +62,7 @@ import java.util.Stack;
 
 public class SparkReduceSinkMapJoinProc implements NodeProcessor {
 
-  public static final Log LOG = 
LogFactory.getLog(SparkReduceSinkMapJoinProc.class.getName());
+  public static final Logger LOG = 
LoggerFactory.getLogger(SparkReduceSinkMapJoinProc.class.getName());
 
   public static class SparkMapJoinFollowedByGroupByProcessor implements 
NodeProcessor {
     private boolean hasGroupBy = false;

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java
index 6b3e715..a8ff158 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java
@@ -26,8 +26,8 @@ import java.util.Map;
 import java.util.Map.Entry;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.ql.ErrorMsg;
 import org.apache.hadoop.hive.ql.exec.ColumnInfo;
@@ -82,7 +82,7 @@ import com.google.common.collect.Maps;
 
 public class StatsRulesProcFactory {
 
-  private static final Log LOG = 
LogFactory.getLog(StatsRulesProcFactory.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(StatsRulesProcFactory.class.getName());
   private static final boolean isDebugEnabled = LOG.isDebugEnabled();
 
   /**
@@ -117,7 +117,7 @@ public class StatsRulesProcFactory {
       } catch (CloneNotSupportedException e) {
         throw new 
SemanticException(ErrorMsg.STATISTICS_CLONING_FAILED.getMsg());
       } catch (HiveException e) {
-        LOG.debug(e);
+        LOG.debug("Failed to retrieve stats ",e);
         throw new SemanticException(e);
       }
       return null;
@@ -1192,7 +1192,7 @@ public class StatsRulesProcFactory {
           wcStats.setNumRows(newNumRows);
           wcStats.setDataSize(newDataSize);
           jop.setStatistics(wcStats);
- 
+
           if (isDebugEnabled) {
             LOG.debug("[1] STATS-" + jop.toString() + ": " + 
wcStats.extendedToString());
           }

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/AppMasterEventProcessor.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/parse/AppMasterEventProcessor.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/AppMasterEventProcessor.java
index 9ef0dd7..9442ba3 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/AppMasterEventProcessor.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/AppMasterEventProcessor.java
@@ -22,8 +22,8 @@ import java.util.ArrayList;
 import java.util.List;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.AppMasterEventOperator;
 import org.apache.hadoop.hive.ql.lib.Node;
 import org.apache.hadoop.hive.ql.lib.NodeProcessor;
@@ -37,7 +37,7 @@ import org.apache.hadoop.hive.ql.plan.DynamicPruningEventDesc;
  */
 public class AppMasterEventProcessor implements NodeProcessor {
 
-  static final private Log LOG = 
LogFactory.getLog(AppMasterEventProcessor.class.getName());
+  static final private Logger LOG = 
LoggerFactory.getLogger(AppMasterEventProcessor.class.getName());
 
   @Override
   public Object process(Node nd, Stack<Node> stack, NodeProcessorCtx procCtx, 
Object... nodeOutputs)

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java
index fbe93f9..cead5ae 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java
@@ -34,16 +34,12 @@ import java.util.Map;
 import java.util.Map.Entry;
 import java.util.Set;
 
-import org.antlr.runtime.tree.CommonTree;
 import org.antlr.runtime.tree.Tree;
 import org.apache.commons.lang3.StringUtils;
 import org.apache.commons.lang3.tuple.Pair;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hive.common.FileUtils;
 import org.apache.hadoop.hive.conf.HiveConf;
-import org.apache.hadoop.hive.conf.HiveConf.ConfVars;
 import org.apache.hadoop.hive.metastore.HiveMetaStore;
 import org.apache.hadoop.hive.metastore.api.Database;
 import org.apache.hadoop.hive.metastore.api.FieldSchema;
@@ -75,8 +71,9 @@ import org.apache.hadoop.hive.serde2.io.DateWritable;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters;
 import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo;
-import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory;
 import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 
 import com.google.common.annotations.VisibleForTesting;
 
@@ -85,12 +82,12 @@ import com.google.common.annotations.VisibleForTesting;
  *
  */
 public abstract class BaseSemanticAnalyzer {
-  protected static final Log STATIC_LOG = 
LogFactory.getLog(BaseSemanticAnalyzer.class.getName());
+  protected static final Logger STATIC_LOG = 
LoggerFactory.getLogger(BaseSemanticAnalyzer.class.getName());
   protected final Hive db;
   protected final HiveConf conf;
   protected List<Task<? extends Serializable>> rootTasks;
   protected FetchTask fetchTask;
-  protected final Log LOG;
+  protected final Logger LOG;
   protected final LogHelper console;
 
   protected Context ctx;
@@ -203,7 +200,7 @@ public abstract class BaseSemanticAnalyzer {
       this.conf = conf;
       this.db = db;
       rootTasks = new ArrayList<Task<? extends Serializable>>();
-      LOG = LogFactory.getLog(this.getClass().getName());
+      LOG = LoggerFactory.getLogger(this.getClass().getName());
       console = new LogHelper(LOG);
       idToTableNameMap = new HashMap<String, String>();
       inputs = new LinkedHashSet<ReadEntity>();

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnAccessAnalyzer.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnAccessAnalyzer.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnAccessAnalyzer.java
index c2286a0..ea58917 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnAccessAnalyzer.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnAccessAnalyzer.java
@@ -22,15 +22,15 @@ import java.util.HashMap;
 import java.util.List;
 import java.util.Map;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.Operator;
 import org.apache.hadoop.hive.ql.exec.TableScanOperator;
 import org.apache.hadoop.hive.ql.metadata.Table;
 import org.apache.hadoop.hive.ql.plan.OperatorDesc;
 
 public class ColumnAccessAnalyzer {
-  private static final Log   LOG = 
LogFactory.getLog(ColumnAccessAnalyzer.class.getName());
+  private static final Logger   LOG = 
LoggerFactory.getLogger(ColumnAccessAnalyzer.class.getName());
   private final ParseContext pGraphContext;
 
   public ColumnAccessAnalyzer() {

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java
index 533bcdf..543bc0f 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java
@@ -24,8 +24,8 @@ import java.util.LinkedList;
 import java.util.List;
 import java.util.Map;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.conf.HiveConf.ConfVars;
 import org.apache.hadoop.hive.conf.HiveVariableSource;
@@ -46,8 +46,8 @@ import org.apache.hadoop.hive.serde.serdeConstants;
  *
  */
 public class ColumnStatsSemanticAnalyzer extends SemanticAnalyzer {
-  private static final Log LOG = LogFactory
-      .getLog(ColumnStatsSemanticAnalyzer.class);
+  private static final Logger LOG = LoggerFactory
+      .getLogger(ColumnStatsSemanticAnalyzer.class);
 
   private ASTNode originalTree;
   private ASTNode rewrittenTree;

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java
index 2d7d9d4..b4546e1 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java
@@ -21,8 +21,8 @@ package org.apache.hadoop.hive.ql.parse;
 import com.google.common.collect.Lists;
 import org.antlr.runtime.tree.CommonTree;
 import org.antlr.runtime.tree.Tree;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hive.common.JavaUtils;
@@ -167,7 +167,7 @@ import static 
org.apache.hadoop.hive.ql.parse.HiveParser.TOK_DATABASEPROPERTIES;
  *
  */
 public class DDLSemanticAnalyzer extends BaseSemanticAnalyzer {
-  private static final Log LOG = LogFactory.getLog(DDLSemanticAnalyzer.class);
+  private static final Logger LOG = 
LoggerFactory.getLogger(DDLSemanticAnalyzer.class);
   private static final Map<Integer, String> TokenToTypeName = new 
HashMap<Integer, String>();
 
   private final Set<String> reservedPartitionValues;

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java
index bdf0ed7..179f9c2 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.parse;
 
 import com.google.common.base.Function;
 import org.apache.commons.lang.StringUtils;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.fs.FSDataInputStream;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
@@ -63,7 +63,7 @@ import java.util.TreeMap;
  */
 public class EximUtil {
 
-  private static Log LOG = LogFactory.getLog(EximUtil.class);
+  private static final Logger LOG = LoggerFactory.getLogger(EximUtil.class);
 
   private EximUtil() {
   }

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/FileSinkProcessor.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/FileSinkProcessor.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/FileSinkProcessor.java
index 2f769d4..191c82e 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/FileSinkProcessor.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/FileSinkProcessor.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.parse;
 
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.ql.exec.FileSinkOperator;
 import org.apache.hadoop.hive.ql.lib.Node;
 import org.apache.hadoop.hive.ql.lib.NodeProcessor;
@@ -34,7 +34,7 @@ import org.apache.hadoop.hive.ql.lib.NodeProcessorCtx;
  */
 public class FileSinkProcessor implements NodeProcessor {
 
-  static final private Log LOG = 
LogFactory.getLog(FileSinkProcessor.class.getName());
+  static final private Logger LOG = 
LoggerFactory.getLogger(FileSinkProcessor.class.getName());
 
   @Override
   public Object process(Node nd, Stack<Node> stack,
@@ -48,4 +48,4 @@ public class FileSinkProcessor implements NodeProcessor {
     context.fileSinkSet.add(fileSink);
     return true;
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/FunctionSemanticAnalyzer.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/parse/FunctionSemanticAnalyzer.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/FunctionSemanticAnalyzer.java
index 418b4ad..be908d3 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/FunctionSemanticAnalyzer.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/FunctionSemanticAnalyzer.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.parse;
 import java.util.ArrayList;
 import java.util.List;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.conf.HiveConf.ConfVars;
 import org.apache.hadoop.hive.metastore.api.Database;
@@ -46,8 +46,8 @@ import org.apache.hadoop.hive.ql.plan.PlanUtils;
  *
  */
 public class FunctionSemanticAnalyzer extends BaseSemanticAnalyzer {
-  private static final Log LOG = LogFactory
-      .getLog(FunctionSemanticAnalyzer.class);
+  private static final Logger LOG = LoggerFactory
+      .getLogger(FunctionSemanticAnalyzer.class);
 
   public FunctionSemanticAnalyzer(HiveConf conf) throws SemanticException {
     super(conf);
@@ -178,7 +178,7 @@ public class FunctionSemanticAnalyzer extends 
BaseSemanticAnalyzer {
         functionName = qualifiedNameParts[1];
         database = getDatabase(dbName);
       } catch (HiveException e) {
-        LOG.error(e);
+        LOG.error("Failed to get database ", e);
         throw new SemanticException(e);
       }
     }

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezUtils.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezUtils.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezUtils.java
index 3450a26..c5f7426 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezUtils.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezUtils.java
@@ -26,8 +26,8 @@ import java.util.LinkedList;
 import java.util.List;
 import java.util.Set;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hive.conf.HiveConf;
@@ -67,7 +67,7 @@ import static 
org.apache.hadoop.hive.ql.plan.ReduceSinkDesc.ReducerTraits.AUTOPA
  * Methods are made non-static on as needed basis.
  */
 public class GenTezUtils {
-  static final private Log LOG = LogFactory.getLog(GenTezUtils.class);
+  static final private Logger LOG = LoggerFactory.getLogger(GenTezUtils.class);
 
   public GenTezUtils() {
   }

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezWork.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezWork.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezWork.java
index c4e0413..1a49de1 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezWork.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezWork.java
@@ -25,8 +25,6 @@ import java.util.Map;
 import java.util.Map.Entry;
 import java.util.Stack;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.ql.exec.CommonMergeJoinOperator;
 import org.apache.hadoop.hive.ql.exec.DummyStoreOperator;
@@ -51,6 +49,8 @@ import 
org.apache.hadoop.hive.ql.plan.TezEdgeProperty.EdgeType;
 import org.apache.hadoop.hive.ql.plan.TezWork;
 import org.apache.hadoop.hive.ql.plan.TezWork.VertexType;
 import org.apache.hadoop.hive.ql.plan.UnionWork;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 
 /**
  * GenTezWork separates the operator tree into tez tasks.
@@ -60,7 +60,7 @@ import org.apache.hadoop.hive.ql.plan.UnionWork;
  */
 public class GenTezWork implements NodeProcessor {
 
-  static final private Log LOG = LogFactory.getLog(GenTezWork.class.getName());
+  private static final Logger LOG = 
LoggerFactory.getLogger(GenTezWork.class.getName());
 
   private final GenTezUtils utils;
 
@@ -337,7 +337,7 @@ public class GenTezWork implements NodeProcessor {
         unionWork = context.rootUnionWorkMap.get(root);
         if (unionWork == null) {
           // if unionWork is null, it means it is the first time. we need to
-          // create a union work object and add this work to it. Subsequent 
+          // create a union work object and add this work to it. Subsequent
           // work should reference the union and not the actual work.
           unionWork = GenTezUtils.createUnionWork(context, root, operator, 
tezWork);
           // finally connect the union work with work
@@ -495,7 +495,7 @@ public class GenTezWork implements NodeProcessor {
     int pos = stack.indexOf(currentMergeJoinOperator);
     return (Operator<? extends OperatorDesc>) stack.get(pos - 1);
   }
-  
+
   private void connectUnionWorkWithWork(UnionWork unionWork, BaseWork work, 
TezWork tezWork,
       GenTezProcContext context) {
     LOG.debug("Connecting union work (" + unionWork + ") with work (" + work + 
")");

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g
index d8fb83d..009e72e 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g
@@ -375,8 +375,6 @@ import java.util.Collection;
 import java.util.HashMap;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hive.conf.HiveConf;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
 }
 
 
@@ -619,14 +617,12 @@ import org.apache.commons.logging.LogFactory;
     return msg;
   }
   
-  public static final Log LOG = LogFactory.getLog("HiveParser");
   public void pushMsg(String msg, RecognizerSharedState state) {
     // ANTLR generated code does not wrap the @init code wit this backtracking 
check,
     //  even if the matching @after has it. If we have parser rules with that 
are doing
     // some lookahead with syntactic predicates this can cause the push() and 
pop() calls
     // to become unbalanced, so make sure both push/pop check the backtracking 
state.
     if (state.backtracking == 0) {
-      // LOG.debug("Push " + msg);
       msgs.push(msg);
     }
   }
@@ -634,7 +630,6 @@ import org.apache.commons.logging.LogFactory;
   public void popMsg(RecognizerSharedState state) {
     if (state.backtracking == 0) {
       Object o = msgs.pop();
-      // LOG.debug("Pop " + o);
     }
   }
 

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/InputSignature.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/InputSignature.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/InputSignature.java
index 213d446..7692223 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/InputSignature.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/InputSignature.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.parse;
 
 import java.util.ArrayList;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo;
 import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory;
 
@@ -36,7 +36,7 @@ public class InputSignature {
   private final ArrayList<TypeInfo> typeArray;
 
   @SuppressWarnings("unused")
-  private static final Log LOG = LogFactory.getLog(InputSignature.class
+  private static final Logger LOG = 
LoggerFactory.getLogger(InputSignature.class
       .getName());
 
   public InputSignature(String name) {

http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/MacroSemanticAnalyzer.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/parse/MacroSemanticAnalyzer.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/MacroSemanticAnalyzer.java
index e3ba201..e394914 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/MacroSemanticAnalyzer.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/MacroSemanticAnalyzer.java
@@ -28,8 +28,8 @@ import java.util.Set;
 import java.util.Stack;
 import java.util.LinkedHashSet;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.conf.HiveConf.ConfVars;
 import org.apache.hadoop.hive.metastore.MetaStoreUtils;
@@ -59,8 +59,8 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils;
  *
  */
 public class MacroSemanticAnalyzer extends BaseSemanticAnalyzer {
-  private static final Log LOG = LogFactory
-      .getLog(MacroSemanticAnalyzer.class);
+  private static final Logger LOG = LoggerFactory
+      .getLogger(MacroSemanticAnalyzer.class);
 
   public MacroSemanticAnalyzer(HiveConf conf) throws SemanticException {
     super(conf);

Reply via email to