http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinAddNotNullRule.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinAddNotNullRule.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinAddNotNullRule.java index c4a40bf..2a415d5 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinAddNotNullRule.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinAddNotNullRule.java @@ -47,7 +47,6 @@ import org.apache.hadoop.hive.ql.optimizer.calcite.reloperators.HiveFilter; import org.apache.hadoop.hive.ql.optimizer.calcite.translator.SqlFunctionConverter; import org.apache.hadoop.hive.ql.parse.SemanticException; -import com.esotericsoftware.minlog.Log; import com.google.common.collect.ImmutableList; public final class HiveJoinAddNotNullRule extends RelOptRule { @@ -92,7 +91,6 @@ public final class HiveJoinAddNotNullRule extends RelOptRule { try { joinPredInfo = HiveCalciteUtil.JoinPredicateInfo.constructJoinPredicateInfo(join); } catch (CalciteSemanticException e) { - Log.trace("Failed to add is not null filter on join ", e); return; } @@ -183,7 +181,7 @@ public final class HiveJoinAddNotNullRule extends RelOptRule { } return newConditions; } - + private static Map<String,RexNode> splitCondition(RexNode condition) { Map<String,RexNode> newConditions = new HashMap<String,RexNode>(); if (condition.getKind() == SqlKind.AND) { @@ -196,10 +194,10 @@ public final class HiveJoinAddNotNullRule extends RelOptRule { } return newConditions; } - + private static RelNode createHiveFilterConjunctiveCondition(FilterFactory filterFactory, RexBuilder rexBuilder, RelNode input, Collection<RexNode> conditions) { final RexNode newCondition = RexUtil.composeConjunction(rexBuilder, conditions, false); return filterFactory.createFilter(input, newCondition); } -} \ No newline at end of file +}
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinToMultiJoinRule.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinToMultiJoinRule.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinToMultiJoinRule.java index 35dbda9..a8b16cc 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinToMultiJoinRule.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinToMultiJoinRule.java @@ -35,8 +35,8 @@ import org.apache.calcite.rex.RexNode; import org.apache.calcite.rex.RexUtil; import org.apache.calcite.util.ImmutableBitSet; import org.apache.calcite.util.Pair; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.optimizer.calcite.CalciteSemanticException; import org.apache.hadoop.hive.ql.optimizer.calcite.HiveCalciteUtil; import org.apache.hadoop.hive.ql.optimizer.calcite.HiveCalciteUtil.JoinPredicateInfo; @@ -59,7 +59,7 @@ public class HiveJoinToMultiJoinRule extends RelOptRule { private final ProjectFactory projectFactory; - private static transient final Log LOG = LogFactory.getLog(HiveJoinToMultiJoinRule.class); + private static transient final Logger LOG = LoggerFactory.getLogger(HiveJoinToMultiJoinRule.class); //~ Constructors ----------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePreFilteringRule.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePreFilteringRule.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePreFilteringRule.java index 5824127..82d9600 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePreFilteringRule.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePreFilteringRule.java @@ -18,10 +18,8 @@ package org.apache.hadoop.hive.ql.optimizer.calcite.rules; import java.util.ArrayList; -import java.util.Collection; import java.util.EnumSet; import java.util.List; -import java.util.Map.Entry; import java.util.Set; import org.apache.calcite.plan.RelOptPredicateList; @@ -40,8 +38,8 @@ import org.apache.calcite.rex.RexLiteral; import org.apache.calcite.rex.RexNode; import org.apache.calcite.rex.RexUtil; import org.apache.calcite.sql.SqlKind; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.optimizer.calcite.reloperators.HiveFilter; import com.google.common.collect.ImmutableList; @@ -52,8 +50,8 @@ import com.google.common.collect.Sets; public class HivePreFilteringRule extends RelOptRule { - protected static final Log LOG = LogFactory - .getLog(HivePreFilteringRule.class.getName()); + protected static final Logger LOG = LoggerFactory + .getLogger(HivePreFilteringRule.class.getName()); public static final HivePreFilteringRule INSTANCE = http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/PartitionPrune.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/PartitionPrune.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/PartitionPrune.java index 9098c64..0e5e2b9 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/PartitionPrune.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/PartitionPrune.java @@ -24,8 +24,8 @@ import java.util.List; import java.util.Set; import java.io.IOException; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.calcite.plan.RelOptCluster; import org.apache.calcite.rel.type.RelDataType; import org.apache.calcite.rel.type.RelDataTypeField; @@ -67,7 +67,7 @@ public class PartitionPrune { public static class ExtractPartPruningPredicate extends RexVisitorImpl<RexNode> { - private static final Log LOG = LogFactory.getLog(ExtractPartPruningPredicate.class); + private static final Logger LOG = LoggerFactory.getLogger(ExtractPartPruningPredicate.class); final RelOptHiveTable hiveTable; final RelDataType rType; final Set<String> partCols; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/stats/HiveRelMdRowCount.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/stats/HiveRelMdRowCount.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/stats/HiveRelMdRowCount.java index 6948fb1..728c5aa 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/stats/HiveRelMdRowCount.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/stats/HiveRelMdRowCount.java @@ -44,13 +44,13 @@ import org.apache.calcite.sql.fun.SqlStdOperatorTable; import org.apache.calcite.util.BuiltInMethod; import org.apache.calcite.util.ImmutableBitSet; import org.apache.calcite.util.Pair; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.optimizer.calcite.reloperators.HiveTableScan; public class HiveRelMdRowCount extends RelMdRowCount { - protected static final Log LOG = LogFactory.getLog(HiveRelMdRowCount.class.getName()); + protected static final Logger LOG = LoggerFactory.getLogger(HiveRelMdRowCount.class.getName()); public static final RelMetadataProvider SOURCE = ReflectiveRelMetadataProvider @@ -66,24 +66,21 @@ public class HiveRelMdRowCount extends RelMdRowCount { double selectivity = (pkfk.pkInfo.selectivity * pkfk.ndvScalingFactor); selectivity = Math.min(1.0, selectivity); if (LOG.isDebugEnabled()) { - LOG.debug("Identified Primary - Foreign Key relation:"); - LOG.debug(RelOptUtil.toString(join)); - LOG.debug(pkfk); + LOG.debug("Identified Primary - Foreign Key relation: {} {}",RelOptUtil.toString(join), pkfk); } return pkfk.fkInfo.rowCount * selectivity; } return join.getRows(); } + @Override public Double getRowCount(SemiJoin rel) { PKFKRelationInfo pkfk = analyzeJoinForPKFK(rel); if (pkfk != null) { double selectivity = (pkfk.pkInfo.selectivity * pkfk.ndvScalingFactor); selectivity = Math.min(1.0, selectivity); if (LOG.isDebugEnabled()) { - LOG.debug("Identified Primary - Foreign Key relation:"); - LOG.debug(RelOptUtil.toString(rel)); - LOG.debug(pkfk); + LOG.debug("Identified Primary - Foreign Key relation: {} {}", RelOptUtil.toString(rel), pkfk); } return pkfk.fkInfo.rowCount * selectivity; } @@ -109,6 +106,7 @@ public class HiveRelMdRowCount extends RelMdRowCount { this.isPKSideSimple = isPKSideSimple; } + @Override public String toString() { return String.format( "Primary - Foreign Key join:\n\tfkSide = %d\n\tFKInfo:%s\n" + @@ -129,6 +127,7 @@ public class HiveRelMdRowCount extends RelMdRowCount { this.distinctCount = distinctCount; } + @Override public String toString() { return String.format("FKInfo(rowCount=%.2f,ndv=%.2f)", rowCount, distinctCount); } @@ -141,6 +140,7 @@ public class HiveRelMdRowCount extends RelMdRowCount { this.selectivity = selectivity; } + @Override public String toString() { return String.format("PKInfo(rowCount=%.2f,ndv=%.2f,selectivity=%.2f)", rowCount, distinctCount,selectivity); } @@ -230,7 +230,7 @@ public class HiveRelMdRowCount extends RelMdRowCount { int pkSide = leftIsKey ? 0 : rightIsKey ? 1 : -1; - boolean isPKSideSimpleTree = pkSide != -1 ? + boolean isPKSideSimpleTree = pkSide != -1 ? IsSimpleTreeOnJoinKey.check( pkSide == 0 ? left : right, pkSide == 0 ? leftColIdx : rightColIdx) : false; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTConverter.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTConverter.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTConverter.java index 14946b3..e4ac154 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTConverter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTConverter.java @@ -52,8 +52,8 @@ import org.apache.calcite.sql.SqlKind; import org.apache.calcite.sql.SqlOperator; import org.apache.calcite.sql.type.SqlTypeName; import org.apache.calcite.util.ImmutableBitSet; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.metastore.api.FieldSchema; import org.apache.hadoop.hive.ql.metadata.VirtualColumn; import org.apache.hadoop.hive.ql.optimizer.calcite.CalciteSemanticException; @@ -68,7 +68,7 @@ import org.apache.hadoop.hive.ql.parse.ParseDriver; import com.google.common.collect.Iterables; public class ASTConverter { - private static final Log LOG = LogFactory.getLog(ASTConverter.class); + private static final Logger LOG = LoggerFactory.getLogger(ASTConverter.class); private final RelNode root; private final HiveAST hiveAST; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ExprNodeConverter.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ExprNodeConverter.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ExprNodeConverter.java index 42f1ab6..b42e78f 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ExprNodeConverter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ExprNodeConverter.java @@ -41,8 +41,8 @@ import org.apache.calcite.rex.RexWindow; import org.apache.calcite.rex.RexWindowBound; import org.apache.calcite.sql.SqlKind; import org.apache.calcite.sql.type.SqlTypeUtil; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.common.type.HiveDecimal; import org.apache.hadoop.hive.common.type.HiveIntervalDayTime; import org.apache.hadoop.hive.common.type.HiveIntervalYearMonth; @@ -79,12 +79,11 @@ import com.google.common.collect.ImmutableSet; public class ExprNodeConverter extends RexVisitorImpl<ExprNodeDesc> { private final String tabAlias; - private final String columnAlias; private final RelDataType inputRowType; private final ImmutableSet<Integer> inputVCols; - private List<WindowFunctionSpec> windowFunctionSpecs = new ArrayList<>(); + private final List<WindowFunctionSpec> windowFunctionSpecs = new ArrayList<>(); private final RelDataTypeFactory dTFactory; - protected final Log LOG = LogFactory.getLog(this.getClass().getName()); + protected final Logger LOG = LoggerFactory.getLogger(this.getClass().getName()); private static long uniqueCounter = 0; public ExprNodeConverter(String tabAlias, RelDataType inputRowType, @@ -96,7 +95,6 @@ public class ExprNodeConverter extends RexVisitorImpl<ExprNodeDesc> { RelDataType outputRowType, Set<Integer> inputVCols, RelDataTypeFactory dTFactory) { super(true); this.tabAlias = tabAlias; - this.columnAlias = columnAlias; this.inputRowType = inputRowType; this.inputVCols = ImmutableSet.copyOf(inputVCols); this.dTFactory = dTFactory; @@ -146,7 +144,7 @@ public class ExprNodeConverter extends RexVisitorImpl<ExprNodeDesc> { try { gfDesc = ExprNodeGenericFuncDesc.newInstance(hiveUdf, args); } catch (UDFArgumentException e) { - LOG.error(e); + LOG.error("Failed to instantiate udf: ", e); throw new RuntimeException(e); } } http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverter.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverter.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverter.java index f6c0114..f0f8aa8 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverter.java @@ -40,8 +40,8 @@ import org.apache.calcite.rex.RexLiteral; import org.apache.calcite.rex.RexNode; import org.apache.calcite.util.ImmutableBitSet; import org.apache.calcite.util.Pair; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.ErrorMsg; import org.apache.hadoop.hive.ql.exec.ColumnInfo; @@ -105,7 +105,7 @@ import com.google.common.collect.Lists; public class HiveOpConverter { - private static final Log LOG = LogFactory.getLog(HiveOpConverter.class); + private static final Logger LOG = LoggerFactory.getLogger(HiveOpConverter.class); public static enum HIVEAGGOPMODE { NO_SKEW_NO_MAP_SIDE_AGG, // Corresponds to SemAnalyzer genGroupByPlan1MR http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverterPostProc.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverterPostProc.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverterPostProc.java index 5080992..1d0a254 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverterPostProc.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverterPostProc.java @@ -25,8 +25,8 @@ import java.util.Map; import java.util.Set; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.exec.JoinOperator; import org.apache.hadoop.hive.ql.exec.Operator; @@ -48,7 +48,7 @@ import org.apache.hadoop.hive.ql.plan.OperatorDesc; public class HiveOpConverterPostProc implements Transform { - private static final Log LOG = LogFactory.getLog(HiveOpConverterPostProc.class); + private static final Logger LOG = LoggerFactory.getLogger(HiveOpConverterPostProc.class); private ParseContext pctx; private Map<String, Operator<? extends OperatorDesc>> aliasToOpInfo; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierForASTConv.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierForASTConv.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierForASTConv.java index b11cdfd..e820496 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierForASTConv.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierForASTConv.java @@ -38,8 +38,8 @@ import org.apache.calcite.rel.type.RelDataTypeFactory; import org.apache.calcite.rex.RexNode; import org.apache.calcite.sql.SqlAggFunction; import org.apache.calcite.util.Pair; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.metastore.api.FieldSchema; import org.apache.hadoop.hive.ql.optimizer.calcite.CalciteSemanticException; import org.apache.hadoop.hive.ql.optimizer.calcite.HiveCalciteUtil; @@ -53,7 +53,7 @@ import com.google.common.collect.ImmutableList; public class PlanModifierForASTConv { - private static final Log LOG = LogFactory.getLog(PlanModifierForASTConv.class); + private static final Logger LOG = LoggerFactory.getLogger(PlanModifierForASTConv.class); public static RelNode convertOpTree(RelNode rel, List<FieldSchema> resultSchema) http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierUtil.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierUtil.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierUtil.java index 988d6d3..570eea2 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierUtil.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/PlanModifierUtil.java @@ -31,8 +31,8 @@ import org.apache.calcite.rex.RexCall; import org.apache.calcite.rex.RexInputRef; import org.apache.calcite.rex.RexNode; import org.apache.calcite.util.Pair; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.metastore.api.FieldSchema; import org.apache.hadoop.hive.ql.optimizer.calcite.CalciteSemanticException; import org.apache.hadoop.hive.ql.optimizer.calcite.HiveCalciteUtil; @@ -43,7 +43,7 @@ import com.google.common.collect.ImmutableMap; public class PlanModifierUtil { - private static final Log LOG = LogFactory.getLog(PlanModifierUtil.class); + private static final Logger LOG = LoggerFactory.getLogger(PlanModifierUtil.class); protected static void fixTopOBSchema(final RelNode rootRel, http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java index 3d05161..d315497 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java @@ -43,8 +43,8 @@ import org.apache.calcite.sql.SqlOperator; import org.apache.calcite.sql.fun.SqlCastFunction; import org.apache.calcite.sql.parser.SqlParserPos; import org.apache.calcite.sql.type.SqlTypeName; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.common.type.Decimal128; import org.apache.hadoop.hive.common.type.HiveChar; import org.apache.hadoop.hive.common.type.HiveDecimal; @@ -88,7 +88,7 @@ import com.google.common.collect.ImmutableList.Builder; import com.google.common.collect.ImmutableMap; public class RexNodeConverter { - private static final Log LOG = LogFactory.getLog(RexNodeConverter.class); + private static final Logger LOG = LoggerFactory.getLogger(RexNodeConverter.class); private static class InputCtx { private final RelDataType calciteInpDataType; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/SqlFunctionConverter.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/SqlFunctionConverter.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/SqlFunctionConverter.java index d59c6bb..a17fb94 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/SqlFunctionConverter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/SqlFunctionConverter.java @@ -37,8 +37,8 @@ import org.apache.calcite.sql.type.SqlReturnTypeInference; import org.apache.calcite.sql.type.SqlTypeFamily; import org.apache.calcite.util.Util; import org.apache.commons.lang3.StringUtils; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.Description; import org.apache.hadoop.hive.ql.exec.FunctionInfo; import org.apache.hadoop.hive.ql.exec.FunctionRegistry; @@ -70,7 +70,7 @@ import com.google.common.collect.ImmutableMap; import com.google.common.collect.Maps; public class SqlFunctionConverter { - private static final Log LOG = LogFactory.getLog(SqlFunctionConverter.class); + private static final Logger LOG = LoggerFactory.getLogger(SqlFunctionConverter.class); static final Map<String, SqlOperator> hiveToCalcite; static final Map<SqlOperator, HiveToken> calciteToHiveToken; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationOptimizer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationOptimizer.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationOptimizer.java index c1f1519..1b5d921 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationOptimizer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationOptimizer.java @@ -30,8 +30,8 @@ import java.util.Map; import java.util.Set; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.fs.ContentSummary; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -84,7 +84,7 @@ import org.apache.hadoop.hive.ql.plan.OperatorDesc; */ public class CorrelationOptimizer implements Transform { - private static final Log LOG = LogFactory.getLog(CorrelationOptimizer.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(CorrelationOptimizer.class.getName()); private boolean abort; // if correlation optimizer will not try to optimize this query http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/QueryPlanTreeTransformation.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/QueryPlanTreeTransformation.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/QueryPlanTreeTransformation.java index e8ae2f7..315a650 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/QueryPlanTreeTransformation.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/QueryPlanTreeTransformation.java @@ -25,8 +25,8 @@ import java.util.List; import java.util.Map; import java.util.Set; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.DemuxOperator; import org.apache.hadoop.hive.ql.exec.GroupByOperator; import org.apache.hadoop.hive.ql.exec.Operator; @@ -48,7 +48,7 @@ import org.apache.hadoop.hive.ql.plan.TableDesc; * detected by Correlation Optimizer. */ public class QueryPlanTreeTransformation { - private static final Log LOG = LogFactory.getLog(QueryPlanTreeTransformation.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(QueryPlanTreeTransformation.class.getName()); private static void setNewTag(IntraQueryCorrelation correlation, List<Operator<? extends OperatorDesc>> childrenOfDemux, http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteCanApplyCtx.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteCanApplyCtx.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteCanApplyCtx.java index 988bb29..55b85d9 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteCanApplyCtx.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteCanApplyCtx.java @@ -25,8 +25,8 @@ import java.util.List; import java.util.Map; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.FilterOperator; import org.apache.hadoop.hive.ql.exec.GroupByOperator; import org.apache.hadoop.hive.ql.exec.ReduceSinkOperator; @@ -51,7 +51,7 @@ import org.apache.hadoop.hive.ql.parse.SemanticException; */ public final class RewriteCanApplyCtx implements NodeProcessorCtx { - private static final Log LOG = LogFactory.getLog(RewriteCanApplyCtx.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(RewriteCanApplyCtx.class.getName()); private RewriteCanApplyCtx(ParseContext parseContext) { this.parseContext = parseContext; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteGBUsingIndex.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteGBUsingIndex.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteGBUsingIndex.java index 5afe21e..ea1ece6 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteGBUsingIndex.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteGBUsingIndex.java @@ -26,8 +26,8 @@ import java.util.List; import java.util.Map; import java.util.Set; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.api.FieldSchema; import org.apache.hadoop.hive.metastore.api.Index; @@ -93,7 +93,7 @@ public class RewriteGBUsingIndex implements Transform { private ParseContext parseContext; private Hive hiveDb; private HiveConf hiveConf; - private static final Log LOG = LogFactory.getLog(RewriteGBUsingIndex.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(RewriteGBUsingIndex.class.getName()); /* * Stores the list of top TableScanOperator names for which the rewrite http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteParseContextGenerator.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteParseContextGenerator.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteParseContextGenerator.java index 0c111bc..48105de 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteParseContextGenerator.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteParseContextGenerator.java @@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.optimizer.index; import java.io.IOException; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.Context; import org.apache.hadoop.hive.ql.exec.Operator; @@ -48,7 +48,7 @@ import org.apache.hadoop.hive.ql.plan.OperatorDesc; */ public final class RewriteParseContextGenerator { - private static final Log LOG = LogFactory.getLog(RewriteParseContextGenerator.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(RewriteParseContextGenerator.class.getName()); /** * Parse the input {@link String} command and generate an operator tree. http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteQueryUsingAggregateIndexCtx.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteQueryUsingAggregateIndexCtx.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteQueryUsingAggregateIndexCtx.java index 624ee7f..9acc7b7 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteQueryUsingAggregateIndexCtx.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/index/RewriteQueryUsingAggregateIndexCtx.java @@ -24,8 +24,8 @@ import java.util.HashMap; import java.util.List; import java.util.Map; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.ql.exec.ColumnInfo; import org.apache.hadoop.hive.ql.exec.FunctionRegistry; @@ -64,7 +64,7 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils; */ public final class RewriteQueryUsingAggregateIndexCtx implements NodeProcessorCtx { - private static final Log LOG = LogFactory.getLog(RewriteQueryUsingAggregateIndexCtx.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(RewriteQueryUsingAggregateIndexCtx.class.getName()); private RewriteQueryUsingAggregateIndexCtx(ParseContext parseContext, Hive hiveDb, RewriteCanApplyCtx canApplyCtx) { this.parseContext = parseContext; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java index 0304196..28eade4 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java @@ -17,8 +17,8 @@ */ package org.apache.hadoop.hive.ql.optimizer.listbucketingpruner; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.FilterOperator; import org.apache.hadoop.hive.ql.exec.TableScanOperator; import org.apache.hadoop.hive.ql.exec.UDFArgumentException; @@ -40,7 +40,7 @@ import org.apache.hadoop.hive.ql.parse.SemanticException; * */ public class LBPartitionProcFactory extends PrunerOperatorFactory { - static final Log LOG = LogFactory.getLog(ListBucketingPruner.class.getName()); + static final Logger LOG = LoggerFactory.getLogger(ListBucketingPruner.class.getName()); /** * Retrieve partitions for the filter. This is called only when http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPruner.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPruner.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPruner.java index 9f12602..f399a20 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPruner.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPruner.java @@ -24,8 +24,8 @@ import java.util.List; import java.util.Map; import java.util.Set; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.common.FileUtils; import org.apache.hadoop.hive.ql.lib.NodeProcessorCtx; @@ -42,7 +42,7 @@ import org.apache.hadoop.hive.ql.plan.ExprNodeDesc; * */ public class ListBucketingPruner implements Transform { - static final Log LOG = LogFactory.getLog(ListBucketingPruner.class.getName()); + static final Logger LOG = LoggerFactory.getLogger(ListBucketingPruner.class.getName()); /* * (non-Javadoc) http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PartitionConditionRemover.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PartitionConditionRemover.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PartitionConditionRemover.java index cbed375..812ca51 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PartitionConditionRemover.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PartitionConditionRemover.java @@ -23,8 +23,8 @@ import java.util.LinkedHashMap; import java.util.List; import java.util.Map; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.Operator; import org.apache.hadoop.hive.ql.exec.FilterOperator; import org.apache.hadoop.hive.ql.exec.TableScanOperator; @@ -47,8 +47,8 @@ import org.apache.hadoop.hive.ql.parse.SemanticException; public class PartitionConditionRemover implements Transform { // The log - private static final Log LOG = LogFactory - .getLog("hive.ql.optimizer.pcr.PartitionConditionRemover"); + private static final Logger LOG = LoggerFactory + .getLogger("hive.ql.optimizer.pcr.PartitionConditionRemover"); /* * (non-Javadoc) http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrExprProcFactory.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrExprProcFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrExprProcFactory.java index 825938a..7cdc730 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrExprProcFactory.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrExprProcFactory.java @@ -25,8 +25,8 @@ import java.util.List; import java.util.Map; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.FunctionRegistry; import org.apache.hadoop.hive.ql.lib.DefaultGraphWalker; import org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher; @@ -61,7 +61,7 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory; */ public final class PcrExprProcFactory { - public static final Log LOG = LogFactory.getLog(PcrExprProcFactory.class.getName()); + public static final Logger LOG = LoggerFactory.getLogger(PcrExprProcFactory.class.getName()); static Object evalExprWithPart(ExprNodeDesc expr, Partition p, List<VirtualColumn> vcs) throws SemanticException { http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java index 8955cbd..65505b3 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java @@ -21,8 +21,8 @@ package org.apache.hadoop.hive.ql.optimizer.pcr; import java.util.ArrayList; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.FilterOperator; import org.apache.hadoop.hive.ql.exec.Operator; import org.apache.hadoop.hive.ql.exec.TableScanOperator; @@ -49,8 +49,8 @@ import org.apache.hadoop.hive.ql.plan.OperatorDesc; public final class PcrOpProcFactory { // The log - private static final Log LOG = LogFactory - .getLog("hive.ql.optimizer.pcr.OpProcFactory"); + private static final Logger LOG = LoggerFactory + .getLogger("hive.ql.optimizer.pcr.OpProcFactory"); /** * Remove partition condition in a filter operator when possible. This is http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/CrossProductCheck.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/CrossProductCheck.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/CrossProductCheck.java index 6bdb0a7..9ad33fd 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/CrossProductCheck.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/CrossProductCheck.java @@ -28,8 +28,8 @@ import java.util.List; import java.util.Map; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.AbstractMapJoinOperator; import org.apache.hadoop.hive.ql.exec.ConditionalTask; import org.apache.hadoop.hive.ql.exec.CommonMergeJoinOperator; @@ -87,8 +87,8 @@ import org.apache.hadoop.hive.ql.session.SessionState; */ public class CrossProductCheck implements PhysicalPlanResolver, Dispatcher { - protected static transient final Log LOG = LogFactory - .getLog(CrossProductCheck.class); + protected static transient final Logger LOG = LoggerFactory + .getLogger(CrossProductCheck.class); @Override public PhysicalContext resolve(PhysicalContext pctx) throws SemanticException { http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenSparkSkewJoinProcessor.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenSparkSkewJoinProcessor.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenSparkSkewJoinProcessor.java index f88fd0a..895e64e 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenSparkSkewJoinProcessor.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenSparkSkewJoinProcessor.java @@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.optimizer.physical; import com.google.common.base.Preconditions; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.exec.ColumnInfo; @@ -74,7 +74,7 @@ import java.util.Map; * */ public class GenSparkSkewJoinProcessor { - private static final Log LOG = LogFactory.getLog(GenSparkSkewJoinProcessor.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(GenSparkSkewJoinProcessor.class.getName()); private GenSparkSkewJoinProcessor() { // prevent instantiation http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LlapDecider.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LlapDecider.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LlapDecider.java index 9a8cb0f..af6129a 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LlapDecider.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LlapDecider.java @@ -35,8 +35,8 @@ import java.util.List; import java.util.Map; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.exec.FilterOperator; import org.apache.hadoop.hive.ql.exec.FunctionInfo; @@ -84,8 +84,8 @@ import org.apache.hadoop.hive.ql.plan.TezWork; */ public class LlapDecider implements PhysicalPlanResolver { - protected static transient final Log LOG - = LogFactory.getLog(LlapDecider.class); + protected static transient final Logger LOG + = LoggerFactory.getLogger(LlapDecider.class); private PhysicalContext physicalContext; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LocalMapJoinProcFactory.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LocalMapJoinProcFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LocalMapJoinProcFactory.java index 9076d48..df598e7 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LocalMapJoinProcFactory.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/LocalMapJoinProcFactory.java @@ -24,8 +24,8 @@ import java.util.List; import java.util.Map; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.exec.GroupByOperator; import org.apache.hadoop.hive.ql.exec.HashTableDummyOperator; @@ -62,7 +62,7 @@ import org.apache.hadoop.hive.ql.plan.TableDesc; * OOM in group by operator. */ public final class LocalMapJoinProcFactory { - private static final Log LOG = LogFactory.getLog(LocalMapJoinProcFactory.class); + private static final Logger LOG = LoggerFactory.getLogger(LocalMapJoinProcFactory.class); public static NodeProcessor getJoinProc() { return new LocalMapJoinProcessor(); http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MemoryDecider.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MemoryDecider.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MemoryDecider.java index eb8597d..9d67722 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MemoryDecider.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MemoryDecider.java @@ -15,8 +15,8 @@ import java.util.SortedSet; import java.util.Stack; import java.util.TreeSet; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.exec.MapJoinOperator; import org.apache.hadoop.hive.ql.exec.ReduceSinkOperator; @@ -53,7 +53,7 @@ import org.apache.hadoop.hive.ql.plan.TezWork; */ public class MemoryDecider implements PhysicalPlanResolver { - protected static transient final Log LOG = LogFactory.getLog(MemoryDecider.class); + protected static transient final Logger LOG = LoggerFactory.getLogger(MemoryDecider.class); public class MemoryCalculator implements Dispatcher { http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MetadataOnlyOptimizer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MetadataOnlyOptimizer.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MetadataOnlyOptimizer.java index 24610d9..d47d3c2 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MetadataOnlyOptimizer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MetadataOnlyOptimizer.java @@ -24,8 +24,8 @@ import java.util.List; import java.util.Map; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.FileSinkOperator; import org.apache.hadoop.hive.ql.exec.GroupByOperator; import org.apache.hadoop.hive.ql.exec.TableScanOperator; @@ -55,7 +55,7 @@ import org.apache.hadoop.hive.ql.plan.TableScanDesc; * */ public class MetadataOnlyOptimizer implements PhysicalPlanResolver { - static final Log LOG = LogFactory.getLog(MetadataOnlyOptimizer.class.getName()); + static final Logger LOG = LoggerFactory.getLogger(MetadataOnlyOptimizer.class.getName()); static class WalkerCtx implements NodeProcessorCtx { /* operators for which there is chance the optimization can be applied */ http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanOptimizer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanOptimizer.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanOptimizer.java index 080a0e6..200e098 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanOptimizer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanOptimizer.java @@ -23,8 +23,8 @@ import java.util.LinkedHashMap; import java.util.Map; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.FilterOperator; import org.apache.hadoop.hive.ql.exec.LimitOperator; import org.apache.hadoop.hive.ql.exec.TableScanOperator; @@ -49,7 +49,7 @@ import org.apache.hadoop.hive.ql.plan.ExprNodeDesc; */ public class NullScanOptimizer implements PhysicalPlanResolver { - private static final Log LOG = LogFactory.getLog(NullScanOptimizer.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(NullScanOptimizer.class.getName()); @Override public PhysicalContext resolve(PhysicalContext pctx) throws SemanticException { http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java index 8dcd283..2e924fb 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java @@ -30,8 +30,8 @@ import java.util.Map; import java.util.Map.Entry; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.ql.exec.Operator; import org.apache.hadoop.hive.ql.exec.TableScanOperator; @@ -60,7 +60,7 @@ import org.apache.hadoop.hive.serde2.NullStructSerDe; */ public class NullScanTaskDispatcher implements Dispatcher { - static final Log LOG = LogFactory.getLog(NullScanTaskDispatcher.class.getName()); + static final Logger LOG = LoggerFactory.getLogger(NullScanTaskDispatcher.class.getName()); private final PhysicalContext physicalContext; private final Map<Rule, NodeProcessor> rules; @@ -203,4 +203,4 @@ public class NullScanTaskDispatcher implements Dispatcher { } return null; } -} \ No newline at end of file +} http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SerializeFilter.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SerializeFilter.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SerializeFilter.java index f3c1d42..64e7ba2 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SerializeFilter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SerializeFilter.java @@ -15,8 +15,8 @@ import java.util.SortedSet; import java.util.Stack; import java.util.TreeSet; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.exec.Operator; import org.apache.hadoop.hive.ql.exec.StatsTask; @@ -49,7 +49,7 @@ import org.apache.hadoop.hive.ql.plan.TezWork; */ public class SerializeFilter implements PhysicalPlanResolver { - protected static transient final Log LOG = LogFactory.getLog(SerializeFilter.class); + protected static transient final Logger LOG = LoggerFactory.getLogger(SerializeFilter.class); public class Serializer implements Dispatcher { http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java index dc283e8..4dead18 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java @@ -31,8 +31,8 @@ import java.util.Set; import java.util.Stack; import java.util.regex.Pattern; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.api.hive_metastoreConstants; import org.apache.hadoop.hive.ql.exec.*; @@ -151,7 +151,7 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils; public class Vectorizer implements PhysicalPlanResolver { - protected static transient final Log LOG = LogFactory.getLog(Vectorizer.class); + protected static transient final Logger LOG = LoggerFactory.getLogger(Vectorizer.class); Pattern supportedDataTypesPattern; List<Task<? extends Serializable>> vectorizableTasks = http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcCtx.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcCtx.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcCtx.java index a115c67..adfc96e 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcCtx.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcCtx.java @@ -20,15 +20,15 @@ package org.apache.hadoop.hive.ql.optimizer.physical.index; import java.io.Serializable; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.Task; import org.apache.hadoop.hive.ql.lib.NodeProcessorCtx; import org.apache.hadoop.hive.ql.parse.ParseContext; public class IndexWhereProcCtx implements NodeProcessorCtx { - private static final Log LOG = LogFactory.getLog(IndexWhereProcCtx.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(IndexWhereProcCtx.class.getName()); private final Task<? extends Serializable> currentTask; private final ParseContext parseCtx; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcessor.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcessor.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcessor.java index 61ff61e..81e99fc 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcessor.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/index/IndexWhereProcessor.java @@ -27,8 +27,8 @@ import java.util.Map; import java.util.Set; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.fs.ContentSummary; import org.apache.hadoop.hive.metastore.api.Index; import org.apache.hadoop.hive.ql.exec.TableScanOperator; @@ -62,7 +62,7 @@ import org.apache.hadoop.hive.ql.plan.TableScanDesc; */ public class IndexWhereProcessor implements NodeProcessor { - private static final Log LOG = LogFactory.getLog(IndexWhereProcessor.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(IndexWhereProcessor.class.getName()); private final Map<TableScanOperator, List<Index>> tsToIndices; public IndexWhereProcessor(Map<TableScanOperator, List<Index>> tsToIndices) { http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionExpressionForMetastore.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionExpressionForMetastore.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionExpressionForMetastore.java index b76229c..f9978b4 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionExpressionForMetastore.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionExpressionForMetastore.java @@ -22,8 +22,8 @@ import java.io.IOException; import java.nio.ByteBuffer; import java.util.List; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.metastore.Metastore.SplitInfo; import org.apache.hadoop.hive.metastore.Metastore.SplitInfos; import org.apache.hadoop.hive.metastore.PartitionExpressionProxy; @@ -43,7 +43,7 @@ import org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo; * The basic implementation of PartitionExpressionProxy that uses ql package classes. */ public class PartitionExpressionForMetastore implements PartitionExpressionProxy { - private static final Log LOG = LogFactory.getLog(PartitionExpressionForMetastore.class); + private static final Logger LOG = LoggerFactory.getLogger(PartitionExpressionForMetastore.class); @Override public String convertExprToFilter(byte[] exprBytes) throws MetaException { http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java index 5644662..d99362a 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java @@ -29,8 +29,8 @@ import java.util.Set; import com.google.common.annotations.VisibleForTesting; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.common.ObjectPair; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.IMetaStoreClient; @@ -74,7 +74,7 @@ public class PartitionPruner implements Transform { // The log public static final String CLASS_NAME = PartitionPruner.class.getName(); - public static final Log LOG = LogFactory.getLog(CLASS_NAME); + public static final Logger LOG = LoggerFactory.getLogger(CLASS_NAME); /* * (non-Javadoc) http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java index 9c4c25e..41e9ba6 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java @@ -29,8 +29,8 @@ import java.util.Stack; import com.google.common.collect.Maps; import com.google.common.collect.Sets; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.MapJoinOperator; import org.apache.hadoop.hive.ql.exec.Operator; import org.apache.hadoop.hive.ql.exec.spark.SparkTask; @@ -53,7 +53,7 @@ import org.apache.hadoop.hive.ql.plan.SparkWork; * works. */ public class CombineEquivalentWorkResolver implements PhysicalPlanResolver { - protected static transient Log LOG = LogFactory.getLog(CombineEquivalentWorkResolver.class); + protected static transient Logger LOG = LoggerFactory.getLogger(CombineEquivalentWorkResolver.class); @Override public PhysicalContext resolve(PhysicalContext pctx) throws SemanticException { http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java index 5f9225c..ff4924d 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java @@ -21,8 +21,8 @@ package org.apache.hadoop.hive.ql.optimizer.spark; import java.util.List; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.common.ObjectPair; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.api.hive_metastoreConstants; @@ -51,7 +51,7 @@ import org.apache.hadoop.hive.ql.plan.ReduceSinkDesc; */ public class SetSparkReducerParallelism implements NodeProcessor { - private static final Log LOG = LogFactory.getLog(SetSparkReducerParallelism.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(SetSparkReducerParallelism.class.getName()); // Spark memory per task, and total number of cores private ObjectPair<Long, Integer> sparkMemoryAndCores; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java index 39d1f18..4d2b8d6 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java @@ -24,8 +24,8 @@ import java.util.Map; import java.util.Set; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.exec.GroupByOperator; import org.apache.hadoop.hive.ql.exec.JoinOperator; @@ -56,7 +56,7 @@ import org.apache.hadoop.hive.ql.plan.Statistics; */ public class SparkMapJoinOptimizer implements NodeProcessor { - private static final Log LOG = LogFactory.getLog(SparkMapJoinOptimizer.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(SparkMapJoinOptimizer.class.getName()); @Override /** http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkReduceSinkMapJoinProc.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkReduceSinkMapJoinProc.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkReduceSinkMapJoinProc.java index 76517e4..5b3125b 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkReduceSinkMapJoinProc.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkReduceSinkMapJoinProc.java @@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.optimizer.spark; import com.google.common.base.Preconditions; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.exec.GroupByOperator; import org.apache.hadoop.hive.ql.exec.HashTableDummyOperator; @@ -62,7 +62,7 @@ import java.util.Stack; public class SparkReduceSinkMapJoinProc implements NodeProcessor { - public static final Log LOG = LogFactory.getLog(SparkReduceSinkMapJoinProc.class.getName()); + public static final Logger LOG = LoggerFactory.getLogger(SparkReduceSinkMapJoinProc.class.getName()); public static class SparkMapJoinFollowedByGroupByProcessor implements NodeProcessor { private boolean hasGroupBy = false; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java index 6b3e715..a8ff158 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java @@ -26,8 +26,8 @@ import java.util.Map; import java.util.Map.Entry; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.ErrorMsg; import org.apache.hadoop.hive.ql.exec.ColumnInfo; @@ -82,7 +82,7 @@ import com.google.common.collect.Maps; public class StatsRulesProcFactory { - private static final Log LOG = LogFactory.getLog(StatsRulesProcFactory.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(StatsRulesProcFactory.class.getName()); private static final boolean isDebugEnabled = LOG.isDebugEnabled(); /** @@ -117,7 +117,7 @@ public class StatsRulesProcFactory { } catch (CloneNotSupportedException e) { throw new SemanticException(ErrorMsg.STATISTICS_CLONING_FAILED.getMsg()); } catch (HiveException e) { - LOG.debug(e); + LOG.debug("Failed to retrieve stats ",e); throw new SemanticException(e); } return null; @@ -1192,7 +1192,7 @@ public class StatsRulesProcFactory { wcStats.setNumRows(newNumRows); wcStats.setDataSize(newDataSize); jop.setStatistics(wcStats); - + if (isDebugEnabled) { LOG.debug("[1] STATS-" + jop.toString() + ": " + wcStats.extendedToString()); } http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/AppMasterEventProcessor.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/AppMasterEventProcessor.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/AppMasterEventProcessor.java index 9ef0dd7..9442ba3 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/AppMasterEventProcessor.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/AppMasterEventProcessor.java @@ -22,8 +22,8 @@ import java.util.ArrayList; import java.util.List; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.AppMasterEventOperator; import org.apache.hadoop.hive.ql.lib.Node; import org.apache.hadoop.hive.ql.lib.NodeProcessor; @@ -37,7 +37,7 @@ import org.apache.hadoop.hive.ql.plan.DynamicPruningEventDesc; */ public class AppMasterEventProcessor implements NodeProcessor { - static final private Log LOG = LogFactory.getLog(AppMasterEventProcessor.class.getName()); + static final private Logger LOG = LoggerFactory.getLogger(AppMasterEventProcessor.class.getName()); @Override public Object process(Node nd, Stack<Node> stack, NodeProcessorCtx procCtx, Object... nodeOutputs) http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java index fbe93f9..cead5ae 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java @@ -34,16 +34,12 @@ import java.util.Map; import java.util.Map.Entry; import java.util.Set; -import org.antlr.runtime.tree.CommonTree; import org.antlr.runtime.tree.Tree; import org.apache.commons.lang3.StringUtils; import org.apache.commons.lang3.tuple.Pair; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.common.FileUtils; import org.apache.hadoop.hive.conf.HiveConf; -import org.apache.hadoop.hive.conf.HiveConf.ConfVars; import org.apache.hadoop.hive.metastore.HiveMetaStore; import org.apache.hadoop.hive.metastore.api.Database; import org.apache.hadoop.hive.metastore.api.FieldSchema; @@ -75,8 +71,9 @@ import org.apache.hadoop.hive.serde2.io.DateWritable; import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; -import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import com.google.common.annotations.VisibleForTesting; @@ -85,12 +82,12 @@ import com.google.common.annotations.VisibleForTesting; * */ public abstract class BaseSemanticAnalyzer { - protected static final Log STATIC_LOG = LogFactory.getLog(BaseSemanticAnalyzer.class.getName()); + protected static final Logger STATIC_LOG = LoggerFactory.getLogger(BaseSemanticAnalyzer.class.getName()); protected final Hive db; protected final HiveConf conf; protected List<Task<? extends Serializable>> rootTasks; protected FetchTask fetchTask; - protected final Log LOG; + protected final Logger LOG; protected final LogHelper console; protected Context ctx; @@ -203,7 +200,7 @@ public abstract class BaseSemanticAnalyzer { this.conf = conf; this.db = db; rootTasks = new ArrayList<Task<? extends Serializable>>(); - LOG = LogFactory.getLog(this.getClass().getName()); + LOG = LoggerFactory.getLogger(this.getClass().getName()); console = new LogHelper(LOG); idToTableNameMap = new HashMap<String, String>(); inputs = new LinkedHashSet<ReadEntity>(); http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnAccessAnalyzer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnAccessAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnAccessAnalyzer.java index c2286a0..ea58917 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnAccessAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnAccessAnalyzer.java @@ -22,15 +22,15 @@ import java.util.HashMap; import java.util.List; import java.util.Map; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.Operator; import org.apache.hadoop.hive.ql.exec.TableScanOperator; import org.apache.hadoop.hive.ql.metadata.Table; import org.apache.hadoop.hive.ql.plan.OperatorDesc; public class ColumnAccessAnalyzer { - private static final Log LOG = LogFactory.getLog(ColumnAccessAnalyzer.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(ColumnAccessAnalyzer.class.getName()); private final ParseContext pGraphContext; public ColumnAccessAnalyzer() { http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java index 533bcdf..543bc0f 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java @@ -24,8 +24,8 @@ import java.util.LinkedList; import java.util.List; import java.util.Map; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConf.ConfVars; import org.apache.hadoop.hive.conf.HiveVariableSource; @@ -46,8 +46,8 @@ import org.apache.hadoop.hive.serde.serdeConstants; * */ public class ColumnStatsSemanticAnalyzer extends SemanticAnalyzer { - private static final Log LOG = LogFactory - .getLog(ColumnStatsSemanticAnalyzer.class); + private static final Logger LOG = LoggerFactory + .getLogger(ColumnStatsSemanticAnalyzer.class); private ASTNode originalTree; private ASTNode rewrittenTree; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java index 2d7d9d4..b4546e1 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java @@ -21,8 +21,8 @@ package org.apache.hadoop.hive.ql.parse; import com.google.common.collect.Lists; import org.antlr.runtime.tree.CommonTree; import org.antlr.runtime.tree.Tree; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.common.JavaUtils; @@ -167,7 +167,7 @@ import static org.apache.hadoop.hive.ql.parse.HiveParser.TOK_DATABASEPROPERTIES; * */ public class DDLSemanticAnalyzer extends BaseSemanticAnalyzer { - private static final Log LOG = LogFactory.getLog(DDLSemanticAnalyzer.class); + private static final Logger LOG = LoggerFactory.getLogger(DDLSemanticAnalyzer.class); private static final Map<Integer, String> TokenToTypeName = new HashMap<Integer, String>(); private final Set<String> reservedPartitionValues; http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java index bdf0ed7..179f9c2 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java @@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.parse; import com.google.common.base.Function; import org.apache.commons.lang.StringUtils; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -63,7 +63,7 @@ import java.util.TreeMap; */ public class EximUtil { - private static Log LOG = LogFactory.getLog(EximUtil.class); + private static final Logger LOG = LoggerFactory.getLogger(EximUtil.class); private EximUtil() { } http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/FileSinkProcessor.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/FileSinkProcessor.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/FileSinkProcessor.java index 2f769d4..191c82e 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/FileSinkProcessor.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/FileSinkProcessor.java @@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.parse; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.ql.exec.FileSinkOperator; import org.apache.hadoop.hive.ql.lib.Node; import org.apache.hadoop.hive.ql.lib.NodeProcessor; @@ -34,7 +34,7 @@ import org.apache.hadoop.hive.ql.lib.NodeProcessorCtx; */ public class FileSinkProcessor implements NodeProcessor { - static final private Log LOG = LogFactory.getLog(FileSinkProcessor.class.getName()); + static final private Logger LOG = LoggerFactory.getLogger(FileSinkProcessor.class.getName()); @Override public Object process(Node nd, Stack<Node> stack, @@ -48,4 +48,4 @@ public class FileSinkProcessor implements NodeProcessor { context.fileSinkSet.add(fileSink); return true; } -} \ No newline at end of file +} http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/FunctionSemanticAnalyzer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/FunctionSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/FunctionSemanticAnalyzer.java index 418b4ad..be908d3 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/FunctionSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/FunctionSemanticAnalyzer.java @@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.parse; import java.util.ArrayList; import java.util.List; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConf.ConfVars; import org.apache.hadoop.hive.metastore.api.Database; @@ -46,8 +46,8 @@ import org.apache.hadoop.hive.ql.plan.PlanUtils; * */ public class FunctionSemanticAnalyzer extends BaseSemanticAnalyzer { - private static final Log LOG = LogFactory - .getLog(FunctionSemanticAnalyzer.class); + private static final Logger LOG = LoggerFactory + .getLogger(FunctionSemanticAnalyzer.class); public FunctionSemanticAnalyzer(HiveConf conf) throws SemanticException { super(conf); @@ -178,7 +178,7 @@ public class FunctionSemanticAnalyzer extends BaseSemanticAnalyzer { functionName = qualifiedNameParts[1]; database = getDatabase(dbName); } catch (HiveException e) { - LOG.error(e); + LOG.error("Failed to get database ", e); throw new SemanticException(e); } } http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezUtils.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezUtils.java index 3450a26..c5f7426 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezUtils.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezUtils.java @@ -26,8 +26,8 @@ import java.util.LinkedList; import java.util.List; import java.util.Set; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.conf.HiveConf; @@ -67,7 +67,7 @@ import static org.apache.hadoop.hive.ql.plan.ReduceSinkDesc.ReducerTraits.AUTOPA * Methods are made non-static on as needed basis. */ public class GenTezUtils { - static final private Log LOG = LogFactory.getLog(GenTezUtils.class); + static final private Logger LOG = LoggerFactory.getLogger(GenTezUtils.class); public GenTezUtils() { } http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezWork.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezWork.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezWork.java index c4e0413..1a49de1 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezWork.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/GenTezWork.java @@ -25,8 +25,6 @@ import java.util.Map; import java.util.Map.Entry; import java.util.Stack; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.exec.CommonMergeJoinOperator; import org.apache.hadoop.hive.ql.exec.DummyStoreOperator; @@ -51,6 +49,8 @@ import org.apache.hadoop.hive.ql.plan.TezEdgeProperty.EdgeType; import org.apache.hadoop.hive.ql.plan.TezWork; import org.apache.hadoop.hive.ql.plan.TezWork.VertexType; import org.apache.hadoop.hive.ql.plan.UnionWork; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * GenTezWork separates the operator tree into tez tasks. @@ -60,7 +60,7 @@ import org.apache.hadoop.hive.ql.plan.UnionWork; */ public class GenTezWork implements NodeProcessor { - static final private Log LOG = LogFactory.getLog(GenTezWork.class.getName()); + private static final Logger LOG = LoggerFactory.getLogger(GenTezWork.class.getName()); private final GenTezUtils utils; @@ -337,7 +337,7 @@ public class GenTezWork implements NodeProcessor { unionWork = context.rootUnionWorkMap.get(root); if (unionWork == null) { // if unionWork is null, it means it is the first time. we need to - // create a union work object and add this work to it. Subsequent + // create a union work object and add this work to it. Subsequent // work should reference the union and not the actual work. unionWork = GenTezUtils.createUnionWork(context, root, operator, tezWork); // finally connect the union work with work @@ -495,7 +495,7 @@ public class GenTezWork implements NodeProcessor { int pos = stack.indexOf(currentMergeJoinOperator); return (Operator<? extends OperatorDesc>) stack.get(pos - 1); } - + private void connectUnionWorkWithWork(UnionWork unionWork, BaseWork work, TezWork tezWork, GenTezProcContext context) { LOG.debug("Connecting union work (" + unionWork + ") with work (" + work + ")"); http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g b/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g index d8fb83d..009e72e 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g @@ -375,8 +375,6 @@ import java.util.Collection; import java.util.HashMap; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.conf.HiveConf; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; } @@ -619,14 +617,12 @@ import org.apache.commons.logging.LogFactory; return msg; } - public static final Log LOG = LogFactory.getLog("HiveParser"); public void pushMsg(String msg, RecognizerSharedState state) { // ANTLR generated code does not wrap the @init code wit this backtracking check, // even if the matching @after has it. If we have parser rules with that are doing // some lookahead with syntactic predicates this can cause the push() and pop() calls // to become unbalanced, so make sure both push/pop check the backtracking state. if (state.backtracking == 0) { - // LOG.debug("Push " + msg); msgs.push(msg); } } @@ -634,7 +630,6 @@ import org.apache.commons.logging.LogFactory; public void popMsg(RecognizerSharedState state) { if (state.backtracking == 0) { Object o = msgs.pop(); - // LOG.debug("Pop " + o); } } http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/InputSignature.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/InputSignature.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/InputSignature.java index 213d446..7692223 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/InputSignature.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/InputSignature.java @@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.parse; import java.util.ArrayList; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory; @@ -36,7 +36,7 @@ public class InputSignature { private final ArrayList<TypeInfo> typeArray; @SuppressWarnings("unused") - private static final Log LOG = LogFactory.getLog(InputSignature.class + private static final Logger LOG = LoggerFactory.getLogger(InputSignature.class .getName()); public InputSignature(String name) { http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/parse/MacroSemanticAnalyzer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/MacroSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/MacroSemanticAnalyzer.java index e3ba201..e394914 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/MacroSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/MacroSemanticAnalyzer.java @@ -28,8 +28,8 @@ import java.util.Set; import java.util.Stack; import java.util.LinkedHashSet; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConf.ConfVars; import org.apache.hadoop.hive.metastore.MetaStoreUtils; @@ -59,8 +59,8 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils; * */ public class MacroSemanticAnalyzer extends BaseSemanticAnalyzer { - private static final Log LOG = LogFactory - .getLog(MacroSemanticAnalyzer.class); + private static final Logger LOG = LoggerFactory + .getLogger(MacroSemanticAnalyzer.class); public MacroSemanticAnalyzer(HiveConf conf) throws SemanticException { super(conf);