This is an automated email from the ASF dual-hosted git repository. pvary pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/hive.git
The following commit(s) were added to refs/heads/master by this push: new 431e7d9e54 HIVE-26092: Fix javadoc errors for the 4.0.0 release (Peter Vary reviewed by Zoltan Haindrich) (#3185) 431e7d9e54 is described below commit 431e7d9e5431a808106d8db81e11aea74f040da5 Author: pvary <pv...@cloudera.com> AuthorDate: Tue Apr 12 13:52:52 2022 +0200 HIVE-26092: Fix javadoc errors for the 4.0.0 release (Peter Vary reviewed by Zoltan Haindrich) (#3185) --- Jenkinsfile | 12 ++++++++ .../format/datetime/HiveSqlDateTimeFormatter.java | 32 +++++++++++----------- .../org/apache/hadoop/hive/common/type/Date.java | 6 ++-- .../apache/hadoop/hive/common/type/Timestamp.java | 6 ++-- .../apache/hive/common/util/TimestampParser.java | 2 +- .../hive/benchmark/calcite/FieldTrimmerBench.java | 6 ++-- .../apache/hive/benchmark/hash/Murmur3Bench.java | 10 +++---- .../hive/benchmark/serde/LazySimpleSerDeBench.java | 6 ++-- .../vectorization/VectorizedArithmeticBench.java | 10 +++---- .../vectorization/VectorizedComparisonBench.java | 10 +++---- .../vectorization/VectorizedLikeBench.java | 10 +++---- .../vectorization/VectorizedLogicBench.java | 10 +++---- .../hive/ql/qoption/QTestOptionDispatcher.java | 2 +- .../org/apache/hive/jdbc/HiveBaseResultSet.java | 4 +-- .../apache/hive/jdbc/saml/IJdbcBrowserClient.java | 2 +- .../org/apache/hadoop/hive/llap/io/api/LlapIo.java | 2 +- .../security/DefaultJwtSharedSecretProvider.java | 2 +- .../tezplugins/metrics/LlapMetricsListener.java | 2 +- .../org/apache/hadoop/hive/llap/LlapHiveUtils.java | 3 +- .../ql/ddl/table/info/desc/DescTableAnalyzer.java | 6 ++-- .../hadoop/hive/ql/exec/AddToClassPathAction.java | 4 +-- .../java/org/apache/hadoop/hive/ql/exec/Task.java | 2 +- .../org/apache/hadoop/hive/ql/exec/Utilities.java | 3 +- .../hive/ql/exec/WindowFunctionDescription.java | 26 +++++++++--------- .../hive/ql/exec/repl/OptimisedBootstrapUtils.java | 4 +-- .../hadoop/hive/ql/exec/repl/ReplStatsTracker.java | 2 +- .../apache/hadoop/hive/ql/exec/tez/DagUtils.java | 1 - .../expressions/CastDateToCharWithFormat.java | 2 +- .../expressions/CastDateToStringWithFormat.java | 2 +- .../expressions/CastDateToVarCharWithFormat.java | 2 +- .../expressions/CastStringToDateWithFormat.java | 2 +- .../CastStringToTimestampWithFormat.java | 2 +- .../expressions/CastTimestampToCharWithFormat.java | 2 +- .../CastTimestampToStringWithFormat.java | 2 +- .../CastTimestampToVarCharWithFormat.java | 2 +- .../apache/hadoop/hive/ql/io/AcidInputFormat.java | 2 +- .../hadoop/hive/ql/io/orc/encoded/StreamUtils.java | 1 - .../hive/ql/log/syslog/SyslogInputFormat.java | 2 +- .../hadoop/hive/ql/log/syslog/SyslogParser.java | 11 ++++---- .../org/apache/hadoop/hive/ql/metadata/Hive.java | 2 +- .../hive/ql/metadata/HiveStorageHandler.java | 2 +- .../hive/ql/optimizer/ParallelEdgeFixer.java | 2 +- .../hive/ql/optimizer/SemiJoinReductionMerge.java | 6 ++-- .../calcite/functions/HiveMergeableAggregate.java | 2 +- .../calcite/rules/HiveAggregateSortLimitRule.java | 2 +- .../ql/optimizer/calcite/rules/HiveDruidRules.java | 2 +- .../calcite/rules/HiveHepExtractRelNodeRule.java | 3 +- .../HiveProjectSortExchangeTransposeRule.java | 2 +- .../rules/HiveRewriteToDataSketchesRules.java | 6 ++-- ...regateInsertDeleteIncrementalRewritingRule.java | 14 +++++----- ...iveAggregateInsertIncrementalRewritingRule.java | 6 ++-- ...AggregatePartitionIncrementalRewritingRule.java | 4 +-- ...veJoinInsertDeleteIncrementalRewritingRule.java | 4 +-- .../calcite/translator/RexNodeConverter.java | 4 +-- .../hive/ql/optimizer/topnkey/CommonKeyPrefix.java | 6 ++-- .../ql/optimizer/topnkey/TopNKeyProcessor.java | 2 +- .../hadoop/hive/ql/parse/CalcitePlanner.java | 1 - .../hadoop/hive/ql/parse/UnparseTranslator.java | 4 +-- .../hadoop/hive/ql/parse/type/FunctionHelper.java | 2 +- .../hive/ql/txn/compactor/CompactorThread.java | 2 +- .../hive/ql/udf/generic/GenericUDAFRank.java | 2 +- .../hive/ql/udf/generic/GenericUDFCastFormat.java | 2 +- .../org/apache/hadoop/hive/serde2/JsonSerDe.java | 4 +-- .../hadoop/hive/serde2/json/BinaryEncoding.java | 2 +- .../hadoop/hive/serde2/json/HiveJsonReader.java | 6 ++-- .../hive/service/cli/operation/QueryInfoCache.java | 2 +- .../org/apache/hadoop/hive/shims/HadoopShims.java | 2 -- .../hadoop/hive/metastore/HiveMetaStoreClient.java | 2 +- .../hadoop/hive/metastore/IMetaStoreClient.java | 11 ++------ .../hadoop/hive/metastore/utils/FileUtils.java | 1 - .../hadoop/hive/metastore/ExceptionHandler.java | 4 +-- .../metastore/IMetaStoreMetadataTransformer.java | 12 ++++---- .../hive/metastore/MetaStoreEventListener.java | 2 +- .../org/apache/hadoop/hive/metastore/RawStore.java | 15 +++++----- .../jdbc/DerbySQLConnectorProvider.java | 2 +- .../apache/hadoop/hive/metastore/txn/TxnStore.java | 2 +- 76 files changed, 182 insertions(+), 184 deletions(-) diff --git a/Jenkinsfile b/Jenkinsfile index 8d16e60e9a..ade004a6c3 100644 --- a/Jenkinsfile +++ b/Jenkinsfile @@ -350,6 +350,18 @@ tar -xzf packaging/target/apache-hive-*-nightly-*-src.tar.gz } } } + branches['javadoc-check'] = { + executorNode { + stage('Prepare') { + loadWS(); + } + stage('Generate javadoc') { + sh """#!/bin/bash -e +mvn install javadoc:javadoc javadoc:aggregate -DskipTests -pl '!itests/hive-jmh,!itests/util' +""" + } + } + } try { stage('Testing') { parallel branches diff --git a/common/src/java/org/apache/hadoop/hive/common/format/datetime/HiveSqlDateTimeFormatter.java b/common/src/java/org/apache/hadoop/hive/common/format/datetime/HiveSqlDateTimeFormatter.java index 2b79e03c6f..00d4796f7b 100644 --- a/common/src/java/org/apache/hadoop/hive/common/format/datetime/HiveSqlDateTimeFormatter.java +++ b/common/src/java/org/apache/hadoop/hive/common/format/datetime/HiveSqlDateTimeFormatter.java @@ -176,7 +176,7 @@ import java.util.stream.Collectors; * - Retains the exact format (capitalization and length) provided in the pattern string. If p.m. * is in the pattern, we expect a.m. or p.m. in the output; if AM is in the pattern, we expect * AM or PM in the output. If the case is mixed (Am or aM) then the output case will match the - * case of the pattern's first character (Am => AM, aM => am). + * case of the pattern's first character (Am => AM, aM => am). * - String to datetime conversion: * - Conflicts with HH24 and SSSSS. * - It doesn't matter which meridian indicator is in the pattern. @@ -253,21 +253,21 @@ import java.util.stream.Collectors; * - Output is right padded with trailing spaces unless the pattern is marked with the fill mode * modifier (FM). Capitalization happens as follows: * - If the first letter of the pattern is lowercase then the output is lowercase: - * 'mONTH' -> 'may' + * 'mONTH' -> 'may' * - If the first two letters of the pattern are uppercase then the output is uppercase: - * 'MOnth' -> 'MAY' + * 'MOnth' -> 'MAY' * - If the first letter of the pattern is uppercase and the second is lowercase then the output - * is capitalized: 'Month' -> 'May'. + * is capitalized: 'Month' -> 'May'. * - For string to datetime conversion, the case of the pattern does not matter. * * MONTH|Month|month * Name of month of year * - For datetime to string conversion, will include trailing spaces up to length 9 (length of * longest month of year name: "September"). Case is taken into account according to the - * following example (pattern => output): - * - MONTH => JANUARY - * - Month => January - * - month => january + * following example (pattern => output): + * - MONTH => JANUARY + * - Month => January + * - month => january * - For string to datetime conversion, neither the case of the pattern nor the case of the input * are taken into account. * - For string to datetime conversion, conflicts with MM and MON. @@ -276,10 +276,10 @@ import java.util.stream.Collectors; * MON|Mon|mon * Abbreviated name of month of year * - For datetime to string conversion, case is taken into account according to the following - * example (pattern => output): - * - MON => JAN - * - Mon => Jan - * - mon => jan + * example (pattern => output): + * - MON => JAN + * - Mon => Jan + * - mon => jan * - For string to datetime conversion, neither the case of the pattern nor the case of the input * are taken into account. * - For string to datetime conversion, conflicts with MM and MONTH. @@ -289,7 +289,7 @@ import java.util.stream.Collectors; * Name of day of week * - For datetime to string conversion, will include trailing spaces until length is 9 (length of * longest day of week name: "Wednesday"). Case is taken into account according to the following - * example (pattern => output): + * example (pattern => output): * - DAY = SUNDAY * - Day = Sunday * - day = sunday @@ -300,7 +300,7 @@ import java.util.stream.Collectors; * DY|Dy|dy * Abbreviated name of day of week * - For datetime to string conversion, case is taken into account according to the following - * example (pattern => output): + * example (pattern => output): * - DY = SUN * - Dy = Sun * - dy = sun @@ -329,11 +329,11 @@ import java.util.stream.Collectors; * zone agnostic. * * C. Separators - * -|.|/|,|'|;|:|<space> + * -|.|/|,|'|;|:|<space> * Separator * - Uses loose matching. Existence of a sequence of separators in the format should match the * existence of a sequence of separators in the input regardless of the types of the separator or - * the length of the sequence where length > 1. E.g. input=“2019-. ;10/10”, pattern=“YYYY-MM-DD” + * the length of the sequence where length > 1. E.g. input=“2019-. ;10/10”, pattern=“YYYY-MM-DD” * is valid; input=“20191010”, pattern=“YYYY-MM-DD” is not valid. * - If the last separator character in the separator substring is "-" and is immediately followed * by a time zone hour (tzh) token, it's a negative sign and not counted as a separator, UNLESS diff --git a/common/src/java/org/apache/hadoop/hive/common/type/Date.java b/common/src/java/org/apache/hadoop/hive/common/type/Date.java index 0a6de90bdf..13e710c5d8 100644 --- a/common/src/java/org/apache/hadoop/hive/common/type/Date.java +++ b/common/src/java/org/apache/hadoop/hive/common/type/Date.java @@ -39,7 +39,7 @@ import static java.time.temporal.ChronoField.YEAR; /** * This is the internal type for Date. The full qualified input format of Date * is "uuuu-MM-dd". For example: "2021-02-11". - * <table border="2"> + * <table border="2" summary=""> * <tr> * <th>Field</th> * <th>Format</th> @@ -73,8 +73,8 @@ import static java.time.temporal.ChronoField.YEAR; * positive. * </p> * - * @see {@link ChronoField#YEAR} - * @see {@link ChronoField#YEAR_OF_ERA} + * @see java.time.temporal.ChronoField#YEAR + * @see java.time.temporal.ChronoField#YEAR_OF_ERA */ public class Date implements Comparable<Date> { diff --git a/common/src/java/org/apache/hadoop/hive/common/type/Timestamp.java b/common/src/java/org/apache/hadoop/hive/common/type/Timestamp.java index 3efdc1cdef..30b1074c93 100644 --- a/common/src/java/org/apache/hadoop/hive/common/type/Timestamp.java +++ b/common/src/java/org/apache/hadoop/hive/common/type/Timestamp.java @@ -44,7 +44,7 @@ import static java.time.temporal.ChronoField.YEAR; * Timestamp is "uuuu-MM-dd HH:mm:ss[.SSS...]", where the time part is optional. * If time part is absent, a default '00:00:00.0' will be used. * - * <table border="2"> + * <table border="2" summary=""> * <tr> * <th>Field</th> * <th>Format</th> @@ -78,8 +78,8 @@ import static java.time.temporal.ChronoField.YEAR; * positive. * </p> * - * @see {@link ChronoField#YEAR} - * @see {@link ChronoField#YEAR_OF_ERA} + * @see java.time.temporal.ChronoField#YEAR + * @see java.time.temporal.ChronoField#YEAR_OF_ERA */ public class Timestamp implements Comparable<Timestamp> { diff --git a/common/src/java/org/apache/hive/common/util/TimestampParser.java b/common/src/java/org/apache/hive/common/util/TimestampParser.java index bb591fb63f..c8ffc6930c 100644 --- a/common/src/java/org/apache/hive/common/util/TimestampParser.java +++ b/common/src/java/org/apache/hive/common/util/TimestampParser.java @@ -48,7 +48,7 @@ import org.slf4j.LoggerFactory; * In addition to accepting format patterns, this parser provides support for * three pre-defined formats: * - * <table border="1"> + * <table border="1" summary=""> * <thead> * <tr> * <th>Formatter</th> diff --git a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/calcite/FieldTrimmerBench.java b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/calcite/FieldTrimmerBench.java index d98e2511a1..435732c1ab 100644 --- a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/calcite/FieldTrimmerBench.java +++ b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/calcite/FieldTrimmerBench.java @@ -55,13 +55,13 @@ import org.openjdk.jmh.runner.options.OptionsBuilder; /** * This test measures the performance for field trimmer. - * <p/> + * <p> * This test uses JMH framework for benchmarking. * You may execute this benchmark tool using JMH command line in different ways: - * <p/> + * <p> * To use the settings shown in the main() function, use: * $ java -cp target/benchmarks.jar org.apache.hive.benchmark.calcite.FieldTrimmerBench - * <p/> + * <p> * To use the default settings used by JMH, use: * $ java -jar target/benchmarks.jar org.apache.hive.benchmark.calcite.FieldTrimmerBench */ diff --git a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/hash/Murmur3Bench.java b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/hash/Murmur3Bench.java index cd85148ebb..f07420ae95 100644 --- a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/hash/Murmur3Bench.java +++ b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/hash/Murmur3Bench.java @@ -41,23 +41,23 @@ import org.openjdk.jmh.runner.options.OptionsBuilder; /** * This test measures the performance for vectorization. - * <p/> + * <p> * This test uses JMH framework for benchmarking. * You may execute this benchmark tool using JMH command line in different ways: - * <p/> + * <p> * To use the settings shown in the main() function, use: * $ java -cp target/benchmarks.jar org.apache.hive.benchmark.hash.Murmur3Bench - * <p/> + * <p> * To use the default settings used by JMH, use: * $ java -jar target/benchmarks.jar org.apache.hive.benchmark.hash.Murmur3Bench - * <p/> + * <p> * To specify different parameters, use: * - This command will use 10 warm-up iterations, 5 test iterations, and 2 forks. And it will * display the Average Time (avgt) in Microseconds (us) * - Benchmark mode. Available modes are: * [Throughput/thrpt, AverageTime/avgt, SampleTime/sample, SingleShotTime/ss, All/all] * - Output time unit. Available time units are: [m, s, ms, us, ns]. - * <p/> + * <p> * $ java -jar target/benchmarks.jar org.apache.hive.benchmark.hash.Murmur3Bench * -wi 10 -i 5 -f 2 -bm avgt -tu us */ diff --git a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/serde/LazySimpleSerDeBench.java b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/serde/LazySimpleSerDeBench.java index 5dbc571df5..43c148fe0d 100644 --- a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/serde/LazySimpleSerDeBench.java +++ b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/serde/LazySimpleSerDeBench.java @@ -50,13 +50,13 @@ import org.openjdk.jmh.runner.options.OptionsBuilder; public class LazySimpleSerDeBench { /** * This test measures the performance for LazySimpleSerDe. - * <p/> + * <p> * This test uses JMH framework for benchmarking. You may execute this * benchmark tool using JMH command line in different ways: - * <p/> + * <p> * To run using default settings, use: * $ java -cp target/benchmarks.jar org.apache.hive.benchmark.serde.LazySimpleSerDeBench - * <p/> + * <p> */ public static final int DEFAULT_ITER_TIME = 1000000; public static final int DEFAULT_DATA_SIZE = 4096; diff --git a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedArithmeticBench.java b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedArithmeticBench.java index 70ee9b7ddb..6e07d0f360 100644 --- a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedArithmeticBench.java +++ b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedArithmeticBench.java @@ -30,23 +30,23 @@ import org.openjdk.jmh.runner.options.OptionsBuilder; /** * This test measures the performance for vectorization. - * <p/> + * <p> * This test uses JMH framework for benchmarking. * You may execute this benchmark tool using JMH command line in different ways: - * <p/> + * <p> * To use the settings shown in the main() function, use: * $ java -cp target/benchmarks.jar org.apache.hive.benchmark.vectorization.VectorizedArithmeticBench - * <p/> + * <p> * To use the default settings used by JMH, use: * $ java -jar target/benchmarks.jar org.apache.hive.benchmark.vectorization.VectorizedArithmeticBench - * <p/> + * <p> * To specify different parameters, use: * - This command will use 10 warm-up iterations, 5 test iterations, and 2 forks. And it will * display the Average Time (avgt) in Microseconds (us) * - Benchmark mode. Available modes are: * [Throughput/thrpt, AverageTime/avgt, SampleTime/sample, SingleShotTime/ss, All/all] * - Output time unit. Available time units are: [m, s, ms, us, ns]. - * <p/> + * <p> * $ java -jar target/benchmarks.jar org.apache.hive.benchmark.vectorization.VectorizedArithmeticBench * -wi 10 -i 5 -f 2 -bm avgt -tu us */ diff --git a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedComparisonBench.java b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedComparisonBench.java index d54d1fa115..cfec3c8d4f 100644 --- a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedComparisonBench.java +++ b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedComparisonBench.java @@ -24,23 +24,23 @@ import org.openjdk.jmh.runner.options.OptionsBuilder; /** * This test measures the performance for vectorization. - * <p/> + * <p> * This test uses JMH framework for benchmarking. * You may execute this benchmark tool using JMH command line in different ways: - * <p/> + * <p> * To use the settings shown in the main() function, use: * $ java -cp target/benchmarks.jar org.apache.hive.benchmark.vectorization.VectorizedComparisonBench - * <p/> + * <p> * To use the default settings used by JMH, use: * $ java -jar target/benchmarks.jar org.apache.hive.benchmark.vectorization.VectorizedComparisonBench - * <p/> + * <p> * To specify different parameters, use: * - This command will use 10 warm-up iterations, 5 test iterations, and 2 forks. And it will * display the Average Time (avgt) in Microseconds (us) * - Benchmark mode. Available modes are: * [Throughput/thrpt, AverageTime/avgt, SampleTime/sample, SingleShotTime/ss, All/all] * - Output time unit. Available time units are: [m, s, ms, us, ns]. - * <p/> + * <p> * $ java -jar target/benchmarks.jar org.apache.hive.benchmark.vectorization.VectorizedComparisonBench * -wi 10 -i 5 -f 2 -bm avgt -tu us */ diff --git a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedLikeBench.java b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedLikeBench.java index 8b1b045917..30f3992165 100644 --- a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedLikeBench.java +++ b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedLikeBench.java @@ -29,23 +29,23 @@ import org.openjdk.jmh.runner.options.OptionsBuilder; /** * This test measures the performance for vectorization. - * <p/> + * <p> * This test uses JMH framework for benchmarking. * You may execute this benchmark tool using JMH command line in different ways: - * <p/> + * <p> * To use the settings shown in the main() function, use: * $ java -cp target/benchmarks.jar org.apache.hive.benchmark.vectorization.VectorizedLikeBench - * <p/> + * <p> * To use the default settings used by JMH, use: * $ java -jar target/benchmarks.jar org.apache.hive.benchmark.vectorization.VectorizedLikeBench - * <p/> + * <p> * To specify different parameters, use: * - This command will use 10 warm-up iterations, 5 test iterations, and 2 forks. And it will * display the Average Time (avgt) in Microseconds (us) * - Benchmark mode. Available modes are: * [Throughput/thrpt, AverageTime/avgt, SampleTime/sample, SingleShotTime/ss, All/all] * - Output time unit. Available time units are: [m, s, ms, us, ns]. - * <p/> + * <p> * $ java -jar target/benchmarks.jar org.apache.hive.benchmark.vectorization.VectorizedLikeBench * -wi 10 -i 5 -f 2 -bm avgt -tu us */ diff --git a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedLogicBench.java b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedLogicBench.java index bf2f4b4e53..0dede7e19a 100644 --- a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedLogicBench.java +++ b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/VectorizedLogicBench.java @@ -27,23 +27,23 @@ import org.openjdk.jmh.runner.options.OptionsBuilder; /** * This test measures the performance for vectorization. - * <p/> + * <p> * This test uses JMH framework for benchmarking. * You may execute this benchmark tool using JMH command line in different ways: - * <p/> + * <p> * To use the settings shown in the main() function, use: * $ java -cp target/benchmarks.jar org.apache.hive.benchmark.vectorization.VectorizedLogicBench - * <p/> + * <p> * To use the default settings used by JMH, use: * $ java -jar target/benchmarks.jar org.apache.hive.benchmark.vectorization.VectorizedLogicBench - * <p/> + * <p> * To specify different parameters, use: * - This command will use 10 warm-up iterations, 5 test iterations, and 2 forks. And it will * display the Average Time (avgt) in Microseconds (us) * - Benchmark mode. Available modes are: * [Throughput/thrpt, AverageTime/avgt, SampleTime/sample, SingleShotTime/ss, All/all] * - Output time unit. Available time units are: [m, s, ms, us, ns]. - * <p/> + * <p> * $ java -jar target/benchmarks.jar org.apache.hive.benchmark.vectorization.VectorizedLogicBench * -wi 10 -i 5 -f 2 -bm avgt -tu us */ diff --git a/itests/util/src/main/java/org/apache/hadoop/hive/ql/qoption/QTestOptionDispatcher.java b/itests/util/src/main/java/org/apache/hadoop/hive/ql/qoption/QTestOptionDispatcher.java index 64e451690e..75939a4638 100644 --- a/itests/util/src/main/java/org/apache/hadoop/hive/ql/qoption/QTestOptionDispatcher.java +++ b/itests/util/src/main/java/org/apache/hadoop/hive/ql/qoption/QTestOptionDispatcher.java @@ -32,7 +32,7 @@ import org.apache.hadoop.hive.ql.QTestUtil; * Provides facilities to invoke {@link QTestOptionHandler}-s. * * Enables to dispatch option arguments to a specific option handler. - * The option invocation format is '--! qt:<optionName>:<optionArgs> + * The option invocation format is '--! qt:<optionName>:<optionArgs> * * Please refer to specific implementations of {@link QTestOptionHandler} for more detailed information about them. */ diff --git a/jdbc/src/java/org/apache/hive/jdbc/HiveBaseResultSet.java b/jdbc/src/java/org/apache/hive/jdbc/HiveBaseResultSet.java index 45de932740..0ee0027d8c 100644 --- a/jdbc/src/java/org/apache/hive/jdbc/HiveBaseResultSet.java +++ b/jdbc/src/java/org/apache/hive/jdbc/HiveBaseResultSet.java @@ -222,7 +222,7 @@ public abstract class HiveBaseResultSet implements ResultSet { * @param columnIndex the first column is 1, the second is 2, ... * @return the column value; if the value is SQL NULL, the value returned is * false - * @throws if the columnIndex is not valid; if a database access error occurs + * @throws SQLException if the columnIndex is not valid; if a database access error occurs * or this method is called on a closed result set * @see ResultSet#getBoolean(int) */ @@ -261,7 +261,7 @@ public abstract class HiveBaseResultSet implements ResultSet { * the name of the column * @return the column value; if the value is SQL NULL, the value returned is * false - * @throws if the columnIndex is not valid; if a database access error occurs + * @throws SQLException if the columnIndex is not valid; if a database access error occurs * or this method is called on a closed result set * @see ResultSet#getBoolean(String) */ diff --git a/jdbc/src/java/org/apache/hive/jdbc/saml/IJdbcBrowserClient.java b/jdbc/src/java/org/apache/hive/jdbc/saml/IJdbcBrowserClient.java index 6bed4204da..a612583830 100644 --- a/jdbc/src/java/org/apache/hive/jdbc/saml/IJdbcBrowserClient.java +++ b/jdbc/src/java/org/apache/hive/jdbc/saml/IJdbcBrowserClient.java @@ -53,7 +53,7 @@ public interface IJdbcBrowserClient extends Closeable { /** * Initializes the browser client context. The client context contains a client * identifier which must be used to set the http header with key - * {@link HiveSamlUtils.SSO_CLIENT_IDENTIFIER}. + * {@link HiveSamlUtils#SSO_CLIENT_IDENTIFIER}. */ void init(JdbcBrowserClientContext context); diff --git a/llap-client/src/java/org/apache/hadoop/hive/llap/io/api/LlapIo.java b/llap-client/src/java/org/apache/hadoop/hive/llap/io/api/LlapIo.java index 3d6258f29e..a4fc13a0ee 100644 --- a/llap-client/src/java/org/apache/hadoop/hive/llap/io/api/LlapIo.java +++ b/llap-client/src/java/org/apache/hadoop/hive/llap/io/api/LlapIo.java @@ -56,7 +56,7 @@ public interface LlapIo<T> { * @param tag a CacheTag instance must be provided as that's needed for cache insertion * @param fileKey fileId of the ORC file (either the Long fileId of HDFS or the SyntheticFileId). * Optional, if it is not provided, it will be generated, see: - * {@link org.apache.hadoop.hive.ql.io.HdfsUtils.getFileId()} + * org.apache.hadoop.hive.ql.io.HdfsUtils#getFileId() * @return The tail of the ORC file * @throws IOException ex */ diff --git a/llap-common/src/java/org/apache/hadoop/hive/llap/security/DefaultJwtSharedSecretProvider.java b/llap-common/src/java/org/apache/hadoop/hive/llap/security/DefaultJwtSharedSecretProvider.java index dca8afc2c6..68f936ff5b 100644 --- a/llap-common/src/java/org/apache/hadoop/hive/llap/security/DefaultJwtSharedSecretProvider.java +++ b/llap-common/src/java/org/apache/hadoop/hive/llap/security/DefaultJwtSharedSecretProvider.java @@ -43,7 +43,7 @@ import static org.apache.hadoop.hive.conf.HiveConf.ConfVars.LLAP_EXTERNAL_CLIENT * * If secret is not found even after 1) and 2), {@link #init(Configuration)} methods throws {@link IllegalStateException}. * - * Length of shared secret provided in 1) or 2) should be > 32 bytes. + * Length of shared secret provided in 1) or 2) should be > 32 bytes. * * It uses the same encryption and decryption secret which can be used to sign and verify JWT. */ diff --git a/llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/metrics/LlapMetricsListener.java b/llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/metrics/LlapMetricsListener.java index 446100b415..b42434eb52 100644 --- a/llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/metrics/LlapMetricsListener.java +++ b/llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/metrics/LlapMetricsListener.java @@ -41,7 +41,7 @@ public interface LlapMetricsListener { /** * Handler will be called when new data is arrived for every active Llap Daemon in the cluster. - * @param newMetrics The map of the worker indentity -> metrics + * @param newMetrics The map of the worker indentity -> metrics */ void newClusterMetrics(Map<String, LlapMetrics> newMetrics); } diff --git a/ql/src/java/org/apache/hadoop/hive/llap/LlapHiveUtils.java b/ql/src/java/org/apache/hadoop/hive/llap/LlapHiveUtils.java index c988fd81a0..52126d971c 100644 --- a/ql/src/java/org/apache/hadoop/hive/llap/LlapHiveUtils.java +++ b/ql/src/java/org/apache/hadoop/hive/llap/LlapHiveUtils.java @@ -52,7 +52,7 @@ public final class LlapHiveUtils { } /** - * Takes a Path and looks up the PartitionDesc instance associated with it in a map of Path->PartitionDesc entries. + * Takes a Path and looks up the PartitionDesc instance associated with it in a map of Path->PartitionDesc entries. * If it is not found (e.g. Path denotes a partition path, but map contains table level instances only) we will try * to do the same with the parent of this path, traversing up until there's a match, if any. * @param path the absolute path used for the look up @@ -104,7 +104,6 @@ public final class LlapHiveUtils { * Returns MapWork based what is serialized in the JobConf instance provided. * @param job * @return the MapWork instance. Might be null if missing. - * @throws HiveException */ public static MapWork findMapWork(JobConf job) { String inputName = job.get(Utilities.INPUT_NAME, null); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/info/desc/DescTableAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/info/desc/DescTableAnalyzer.java index 0a1a99ab16..fb2e80b863 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/info/desc/DescTableAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/info/desc/DescTableAnalyzer.java @@ -46,14 +46,14 @@ import org.apache.hadoop.hive.ql.parse.SemanticException; * A query like this will generate a tree as follows * "describe formatted default.maptable partition (b=100) id;" * TOK_TABTYPE - * TOK_TABNAME --> root for tablename, 2 child nodes mean DB specified + * TOK_TABNAME --> root for tablename, 2 child nodes mean DB specified * default * maptable - * TOK_PARTSPEC --> root node for partition spec. else columnName + * TOK_PARTSPEC --> root node for partition spec. else columnName * TOK_PARTVAL * b * 100 - * id --> root node for columnName + * id --> root node for columnName * formatted */ @DDLType(types = HiveParser.TOK_DESCTABLE) diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/AddToClassPathAction.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/AddToClassPathAction.java index a2ffbb1f24..b5bcbe90cc 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/AddToClassPathAction.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/AddToClassPathAction.java @@ -26,8 +26,8 @@ import java.util.Objects; /** * Helper class to create UDFClassLoader when running under a security manager. To create a class loader: - * > AddToClassPathAction addAction = new AddToClassPathAction(parentLoader, newPaths, true); - * > UDFClassLoader childClassLoader = AccessController.doPrivileged(addAction); + * > AddToClassPathAction addAction = new AddToClassPathAction(parentLoader, newPaths, true); + * > UDFClassLoader childClassLoader = AccessController.doPrivileged(addAction); * To try to add to the class path of the existing class loader; call the above without forceNewClassLoader=true. * Note that a class loader might be still created as fallback method. * <p> diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/Task.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/Task.java index ba1decdc61..19bb543c21 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/Task.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/Task.java @@ -142,7 +142,7 @@ public abstract class Task<T extends Serializable> implements Serializable, Node protected List<Task<?>> parentTasks; /** * this can be set by the Task, to provide more info about the failure in TaskResult - * where the Driver can find it. This is checked if {@link Task#execute(org.apache.hadoop.hive.ql.TaskQueue)} + * where the Driver can find it. This is checked if {@link Task#execute()} * returns non-0 code. */ private Throwable exception; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java index 1f12d0cb76..09ff212923 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java @@ -1127,7 +1127,8 @@ public final class Utilities { * @param destFileName * the target filename * @return The final path the file was moved to. - * @throws IOException, HiveException + * @throws IOException + * @throws HiveException */ public static Path moveFile(FileSystem fs, Path srcFile, Path destDir, String destFileName) throws IOException, HiveException { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/WindowFunctionDescription.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/WindowFunctionDescription.java index 60932be297..bef82f25fd 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/WindowFunctionDescription.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/WindowFunctionDescription.java @@ -63,23 +63,23 @@ public @interface WindowFunctionDescription { /** * This property specifies whether the UDAF is an Ordered-set aggregate function. - * <ordered-set aggregate functions> ::= - * <hypothetical set function> | - * <inverse distribution function> + * <ordered-set aggregate functions> ::= + * <hypothetical set function> | + * <inverse distribution function> * - * <hypothetical set function> ::= - * <rank function type> <left paren> - * <hypothetical set function value expression list> <right paren> - * <within group specification> + * <hypothetical set function> ::= + * <rank function type> <left paren> + * <hypothetical set function value expression list> <right paren> + * <within group specification> * - * <rank function type> ::= RANK | DENSE_RANK | PERCENT_RANK | CUME_DIST + * <rank function type> ::= RANK | DENSE_RANK | PERCENT_RANK | CUME_DIST * - * <inverse distribution function> ::= - * <inverse distribution function type> <left paren> - * <inverse distribution function argument> <right paren> - * <within group specification> + * <inverse distribution function> ::= + * <inverse distribution function type> <left paren> + * <inverse distribution function argument> <right paren> + * <within group specification> * - * <inverse distribution function type> ::= PERCENTILE_CONT | PERCENTILE_DISC + * <inverse distribution function type> ::= PERCENTILE_CONT | PERCENTILE_DISC * * @return true if the function can be used as an ordered-set aggregate */ diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/OptimisedBootstrapUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/OptimisedBootstrapUtils.java index 002f635cc0..f3aa530283 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/OptimisedBootstrapUtils.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/OptimisedBootstrapUtils.java @@ -74,7 +74,7 @@ public class OptimisedBootstrapUtils { public static final String BOOTSTRAP_TABLES_LIST = "_failover_bootstrap_table_list"; /** - * Gets & checks whether the database is target of replication. + * Gets & checks whether the database is target of replication. * @param dbName name of database * @param hive hive object * @return true, if the database has repl.target.for property set. @@ -91,7 +91,7 @@ public class OptimisedBootstrapUtils { } /** - * Gets the source & target event id from the event ack file + * Gets the source & target event id from the event ack file * @param dumpPath the dump path * @param conf the hive configuration * @return the event id from file. diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplStatsTracker.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplStatsTracker.java index 06c147ce71..023c788aff 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplStatsTracker.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplStatsTracker.java @@ -107,7 +107,7 @@ public class ReplStatsTracker { /** * Get the DescriptiveStatistics for each event type. - * @return A HashMap, with key as event type & value as the DescriptiveAnalytics of the entire run. + * @return A HashMap, with key as event type & value as the DescriptiveAnalytics of the entire run. */ public ConcurrentHashMap<String, DescriptiveStatistics> getDescMap() { return descMap; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DagUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DagUtils.java index 7622bccae5..d8ac63acea 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DagUtils.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DagUtils.java @@ -1260,7 +1260,6 @@ public class DagUtils { * @param inputOutputJars The file names to localize. * @return Map<String, LocalResource> (srcPath, local resources) to add to execution * @throws IOException when hdfs operation fails. - * @throws LoginException when getDefaultDestDir fails with the same exception */ public Map<String, LocalResource> localizeTempFiles(String hdfsDirPathStr, Configuration conf, String[] inputOutputJars, String[] skipJars) throws IOException { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastDateToCharWithFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastDateToCharWithFormat.java index 17afbc3b2b..7ed9902a98 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastDateToCharWithFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastDateToCharWithFormat.java @@ -24,7 +24,7 @@ import org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector; import java.nio.charset.StandardCharsets; /** - * Vectorized UDF for CAST (<DATE> TO CHAR(<LENGTH>) WITH FORMAT <STRING>). + * Vectorized UDF for CAST (<DATE> TO CHAR(<LENGTH>) WITH FORMAT <STRING>). */ public class CastDateToCharWithFormat extends CastDateToChar { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastDateToStringWithFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastDateToStringWithFormat.java index acfa2ba27c..aec5c12721 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastDateToStringWithFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastDateToStringWithFormat.java @@ -24,7 +24,7 @@ import org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector; import java.nio.charset.StandardCharsets; /** - * Vectorized UDF for CAST (<DATE> TO STRING WITH FORMAT <STRING>). + * Vectorized UDF for CAST (<DATE> TO STRING WITH FORMAT <STRING>). */ public class CastDateToStringWithFormat extends CastDateToString { private static final long serialVersionUID = 1L; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastDateToVarCharWithFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastDateToVarCharWithFormat.java index 6ed44a6d57..9cf28fd67d 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastDateToVarCharWithFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastDateToVarCharWithFormat.java @@ -24,7 +24,7 @@ import org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector; import java.nio.charset.StandardCharsets; /** - * Vectorized UDF for CAST (<DATE> TO VARCHAR(<LENGTH>) WITH FORMAT <STRING>). + * Vectorized UDF for CAST (<DATE> TO VARCHAR(<LENGTH>) WITH FORMAT <STRING>). */ public class CastDateToVarCharWithFormat extends CastDateToVarChar { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastStringToDateWithFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastStringToDateWithFormat.java index b50ed0e70c..a2bcd56c6f 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastStringToDateWithFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastStringToDateWithFormat.java @@ -27,7 +27,7 @@ import org.apache.hadoop.hive.serde2.io.DateWritableV2; import java.nio.charset.StandardCharsets; /** - * Vectorized UDF for CAST (<STRING> TO DATE WITH FORMAT <STRING>). + * Vectorized UDF for CAST (<STRING> TO DATE WITH FORMAT <STRING>). */ public class CastStringToDateWithFormat extends CastStringToDate { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastStringToTimestampWithFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastStringToTimestampWithFormat.java index 9361e77de8..407249570d 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastStringToTimestampWithFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastStringToTimestampWithFormat.java @@ -26,7 +26,7 @@ import org.apache.hadoop.hive.ql.exec.vector.TimestampColumnVector; import java.nio.charset.StandardCharsets; /** - * Vectorized UDF for CAST (<STRING> TO TIMESTAMP WITH FORMAT <STRING>). + * Vectorized UDF for CAST (<STRING> TO TIMESTAMP WITH FORMAT <STRING>). */ public class CastStringToTimestampWithFormat extends CastStringToTimestamp { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastTimestampToCharWithFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastTimestampToCharWithFormat.java index 5472a7e9de..bcb2182b1f 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastTimestampToCharWithFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastTimestampToCharWithFormat.java @@ -25,7 +25,7 @@ import org.apache.hadoop.hive.ql.exec.vector.TimestampColumnVector; import java.nio.charset.StandardCharsets; /** - * Vectorized UDF for CAST (<TIMESTAMP> TO CHAR(<LENGTH>) WITH FORMAT <STRING>). + * Vectorized UDF for CAST (<TIMESTAMP> TO CHAR(<LENGTH>) WITH FORMAT <STRING>). */ public class CastTimestampToCharWithFormat extends CastTimestampToChar { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastTimestampToStringWithFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastTimestampToStringWithFormat.java index 148995d0cf..bc748d7fe9 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastTimestampToStringWithFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastTimestampToStringWithFormat.java @@ -25,7 +25,7 @@ import org.apache.hadoop.hive.ql.exec.vector.TimestampColumnVector; import java.nio.charset.StandardCharsets; /** - * Vectorized UDF for CAST (<TIMESTAMP> TO STRING WITH FORMAT <STRING>). + * Vectorized UDF for CAST (<TIMESTAMP> TO STRING WITH FORMAT <STRING>). */ public class CastTimestampToStringWithFormat extends CastTimestampToString { private static final long serialVersionUID = 1L; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastTimestampToVarCharWithFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastTimestampToVarCharWithFormat.java index a0569e1023..7062be7bfd 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastTimestampToVarCharWithFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastTimestampToVarCharWithFormat.java @@ -25,7 +25,7 @@ import org.apache.hadoop.hive.ql.exec.vector.TimestampColumnVector; import java.nio.charset.StandardCharsets; /** - * Vectorized UDF for CAST (<TIMESTAMP> TO VARCHAR(<LENGTH>) WITH FORMAT <STRING>). + * Vectorized UDF for CAST (<TIMESTAMP> TO VARCHAR(<LENGTH>) WITH FORMAT <STRING>). */ public class CastTimestampToVarCharWithFormat extends CastTimestampToVarChar { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/AcidInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/AcidInputFormat.java index ea1c6ce088..c23956c759 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/AcidInputFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/AcidInputFormat.java @@ -137,7 +137,7 @@ public interface AcidInputFormat<KEY extends WritableComparable, VALUE> /** * @param minWriteId min writeId of the delta directory * @param maxWriteId max writeId of the delta directory - * @param stmtIds delta dir suffixes when a single txn writes > 1 delta in the same partition + * @param stmtIds delta dir suffixes when a single txn writes > 1 delta in the same partition * @param visibilityTxnId maybe 0, if the dir name didn't have it. txnid:0 is always visible * @param deltaFiles bucketFiles in the directory */ diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/encoded/StreamUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/encoded/StreamUtils.java index 00433c6004..b2323b8653 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/encoded/StreamUtils.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/encoded/StreamUtils.java @@ -37,7 +37,6 @@ public class StreamUtils { * @param streamName - stream name * @param streamBuffer - stream buffer * @return - SettableUncompressedStream - * @throws IOException */ public static SettableUncompressedStream createSettableUncompressedStream(String streamName, ColumnStreamData streamBuffer) { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/log/syslog/SyslogInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/log/syslog/SyslogInputFormat.java index 69b66dd25f..810788104a 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/log/syslog/SyslogInputFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/log/syslog/SyslogInputFormat.java @@ -83,7 +83,7 @@ import com.google.common.annotations.VisibleForTesting; * If a filename is 2019-04-02-21-00_0.log.gz and timeslice is 300s then the file 2019-04-02-21-00_0.log.gz is * expected to have log lines from timestamp 2019:04:02 21:00:00 to 2019:04:02 21:05:00 timestamp. * - Logs table should have 'ts' as timestamp column. - * - Only simple BETWEEN filter predicate is supported for 'ts' column. There cannot be >1 predicates on 'ts' column. + * - Only simple BETWEEN filter predicate is supported for 'ts' column. There cannot be >1 predicates on 'ts' column. */ public class SyslogInputFormat extends TextInputFormat { private static final Logger LOG = LoggerFactory.getLogger(SyslogInputFormat.class); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/log/syslog/SyslogParser.java b/ql/src/java/org/apache/hadoop/hive/ql/log/syslog/SyslogParser.java index 8b700dc5ef..66ed2e53f0 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/log/syslog/SyslogParser.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/log/syslog/SyslogParser.java @@ -44,18 +44,18 @@ import org.apache.hadoop.hive.common.type.Timestamp; * about the way hive logs using syslog format (specifically RFC5424). * * This implementation also parses structured data, returns all parsed fields as map and also un-escapes messages. - * This parser also gracefully handles some corner cases where 'msg' can be empty or line can start with '<' but not + * This parser also gracefully handles some corner cases where 'msg' can be empty or line can start with '<' but not * a valid RFC5424 format etc. * * Assumption: * 1) This parser assumes the linebreaks '\n' in stack traces for example are replaced by '\r' to make single * line message. The reader will do replacement of '\r' with '\n' at the time of read. * 2) This parser assumes structured data values are html escaped. So it will html unescape when parsing structured - * data. (hive writes log lines directly to stderr that look like rfc5424 layout starting with '<' so the expectation + * data. (hive writes log lines directly to stderr that look like rfc5424 layout starting with '<' so the expectation * from log4j2 is to escape those lines using html escaping). - * 3) Read event returns List<Object> conforming to sys.logs table schema in hive. The schema for sys.logs table is + * 3) Read event returns List<Object> conforming to sys.logs table schema in hive. The schema for sys.logs table is * expected to be (facility STRING, severity STRING, version STRING, ts TIMESTAMP, hostname STRING, app_name STRING, - * proc_id STRING, msg_id STRING, structured_data map<STRING,STRING>, msg BINARY, unmatched BINARY) + * proc_id STRING, msg_id STRING, structured_data map<STRING,STRING>, msg BINARY, unmatched BINARY) * 4) Timestamps are in UTC * * This parser is tested with Log4j2's RFC5424 layout generated using the following properties @@ -161,8 +161,7 @@ public class SyslogParser implements Closeable { * Read the next Syslog message from the stream. * * @return a parsed map of object, or null on EOF. - * @throw EOFException if EOF is found in an inappropriate place. - * @throw IOException if the underlying stream fails, or unexpected + * @throws IOException if the underlying stream fails, or unexpected * bytes are seen. */ public List<Object> readEvent() throws IOException { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java b/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java index 48e859b8c8..23c6d85d03 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java @@ -5038,7 +5038,7 @@ private void constructOneLBLocationMap(FileStatus fSta, * @param isSrcLocal true if source is on local file system * @param isAcidIUD true if this is an ACID based Insert/Update/Delete * @param isOverwrite if true, then overwrite if destination file exist, else add a duplicate copy - * @param newFiles if this is non-null, a list of files that were created as a result of this + * @param newFilesStatus if this is non-null, a list of files that were created as a result of this * move will be returned. * @param isManaged if table is managed. * @param isCompactionTable is table used in query-based compaction diff --git a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java index 7d7a01234f..4fcc288050 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java @@ -213,7 +213,7 @@ public interface HiveStorageHandler extends Configurable { * * @param operatorDesc operatorDesc * @param initialProps Map containing initial operator properties - * @return Map<String, String> containing additional operator specific information from storage handler + * @return Map<String, String> containing additional operator specific information from storage handler * OR `initialProps` if the storage handler choose to not provide any such information. */ default Map<String, String> getOperatorDescProperties(OperatorDesc operatorDesc, Map<String, String> initialProps) { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ParallelEdgeFixer.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ParallelEdgeFixer.java index 55d338f2b5..350c3f0cc4 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ParallelEdgeFixer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ParallelEdgeFixer.java @@ -75,7 +75,7 @@ import com.google.common.collect.Lists; * | | | | * | | | +--- | ---+ * | | | | +-----+ | - * | | >>>> | | |RS_T | | + * | | $gt; | | |RS_T | | * | | | | +-----+ | * | | | +--- | ---+ * | | | | diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/SemiJoinReductionMerge.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/SemiJoinReductionMerge.java index 34531591dc..f78e0c3dbc 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/SemiJoinReductionMerge.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/SemiJoinReductionMerge.java @@ -86,21 +86,19 @@ import static org.apache.hadoop.hive.ql.plan.ExprNodeDescUtils.and; * <li>Filter creation from the source relation and broadcast: SOURCE - SEL - GB1 - RS1 - GB2 - RS2</li> * <li>Filter application on the target relation: TS - FIL[in_bloom(col)]</li> * </ul> - * </p> * <p> * An example of the transformation on three single column semi join reducers is shown below. The plan is simplified for * presentation purposes. * <h3>BEFORE:</h3> * <pre> * / SEL[fname] - GB1 - RS1 - GB2 - RS2 \ - * SOURCE - SEL[lname] - GB1 - RS1 - GB2 - RS2 -> TS[Author] - FIL[in_bloom(fname) ^ in_bloom(lname) ^ in_bloom(age)] + * SOURCE - SEL[lname] - GB1 - RS1 - GB2 - RS2 -> TS[Author] - FIL[in_bloom(fname) ^ in_bloom(lname) ^ in_bloom(age)] * \ SEL[age] - GB1 - RS1 - GB2 - RS2 / * </pre> * <h3>AFTER:</h3> * <pre> - * SOURCE - SEL[fname, lname, age] - GB1 - RS1 - GB2 - RS2 -> TS[Author] - FIL[in_bloom(hash(fname,lname,age)] + * SOURCE - SEL[fname, lname, age] - GB1 - RS1 - GB2 - RS2 -> TS[Author] - FIL[in_bloom(hash(fname,lname,age)] * </pre> - * </p> */ public class SemiJoinReductionMerge extends Transform { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/functions/HiveMergeableAggregate.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/functions/HiveMergeableAggregate.java index 041345abda..9abd7f73cd 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/functions/HiveMergeableAggregate.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/functions/HiveMergeableAggregate.java @@ -28,7 +28,7 @@ import org.apache.calcite.sql.type.SqlReturnTypeInference; * Mergeable aggregate. * * A mergeable aggregate is: - * - accepts the same kind as inputs as the output (an X^n -> X function) + * - accepts the same kind as inputs as the output (an X^n -> X function) * * Example: the SUM function is a great example; since SUM of SUM -s is the overall sum. */ diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveAggregateSortLimitRule.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveAggregateSortLimitRule.java index 4e84d37a4e..583a9ee456 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveAggregateSortLimitRule.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveAggregateSortLimitRule.java @@ -30,7 +30,7 @@ import org.apache.hadoop.hive.ql.optimizer.calcite.reloperators.HiveSortLimit; * Rule that adds sorting to GROUP BY col0 LIMIT n in presence of aggregate functions. * Ex.: SELECT id, count(1) FROM t_table GROUP BY id LIMIT 2 * - * Above query has a physical plan like Reducer 2 <- Map 1 (SIMPLE_EDGE) + * Above query has a physical plan like Reducer 2 <- Map 1 (SIMPLE_EDGE) * Both mapper and reducer edges may have multiple Mapper and Reducer instances to enable parallel process of data. * Aggregate function results are calculated in two steps: * 1) first mappers calculate a partial result from the rows processed by each instance. diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveDruidRules.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveDruidRules.java index d7e2508c89..700f604b67 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveDruidRules.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveDruidRules.java @@ -62,7 +62,7 @@ import java.util.TreeSet; * Druid rules with Hive builder factory. * * Simplify this class when upgrading to Calcite 1.26 using - * <a href="https://issues.apache.org/jira/browse/CALCITE-4200"> + * <a href="https://issues.apache.org/jira/browse/CALCITE-4200">CALCITE-4200</a> */ public class HiveDruidRules { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveHepExtractRelNodeRule.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveHepExtractRelNodeRule.java index a645b896b4..12e47843d1 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveHepExtractRelNodeRule.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveHepExtractRelNodeRule.java @@ -29,7 +29,8 @@ import org.apache.calcite.rel.RelNode; /** * The goal of this rule is to extract the RelNode from the * HepRelVertex node so rules do tree traversal can be applied correctly. - * {@see HiveFieldTrimmerRule, HiveAggregateInsertDeleteIncrementalRewritingRule} + * @see HiveFieldTrimmerRule + * @see org.apache.hadoop.hive.ql.optimizer.calcite.rules.views.HiveAggregateInsertDeleteIncrementalRewritingRule */ public class HiveHepExtractRelNodeRule extends RelOptRule { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveProjectSortExchangeTransposeRule.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveProjectSortExchangeTransposeRule.java index d40d5c447c..cd92b86a4c 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveProjectSortExchangeTransposeRule.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveProjectSortExchangeTransposeRule.java @@ -48,7 +48,7 @@ import com.google.common.collect.ImmutableList; * HiveSortExchange * ... * - * => + * => * * HiveSortExchange * HiveProject diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveRewriteToDataSketchesRules.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveRewriteToDataSketchesRules.java index b86423c36d..d97669936e 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveRewriteToDataSketchesRules.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveRewriteToDataSketchesRules.java @@ -67,7 +67,7 @@ import com.google.common.collect.Lists; /** * This rule could rewrite aggregate calls to be calculated using sketch based functions. * - * <br/> + * <br> * Currently it can rewrite: * <ul> * <li>{@code count(distinct(x))} using {@code CountDistinctRewrite} @@ -652,7 +652,7 @@ public final class HiveRewriteToDataSketchesRules { * <pre> * SELECT id, NTILE(4) OVER (ORDER BY id) FROM sketch_input; * ⇒ SELECT id, CASE - * WHEN CEIL(ds_kll_cdf(ds, CAST(id AS FLOAT) )[0]) < 1 + * WHEN CEIL(ds_kll_cdf(ds, CAST(id AS FLOAT) )[0]) < 1 * THEN 1 * ELSE CEIL(ds_kll_cdf(ds, CAST(id AS FLOAT) )[0]) * END @@ -708,7 +708,7 @@ public final class HiveRewriteToDataSketchesRules { * <pre> * SELECT id, RANK() OVER (ORDER BY id) FROM sketch_input; * ⇒ SELECT id, CASE - * WHEN ds_kll_n(ds) < (ceil(ds_kll_rank(ds, CAST(id AS FLOAT) )*ds_kll_n(ds))+1) + * WHEN ds_kll_n(ds) < (ceil(ds_kll_rank(ds, CAST(id AS FLOAT) )*ds_kll_n(ds))+1) * THEN ds_kll_n(ds) * ELSE (ceil(ds_kll_rank(ds, CAST(id AS FLOAT) )*ds_kll_n(ds))+1) * END diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveAggregateInsertDeleteIncrementalRewritingRule.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveAggregateInsertDeleteIncrementalRewritingRule.java index 601cbfc9e9..9005f643e2 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveAggregateInsertDeleteIncrementalRewritingRule.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveAggregateInsertDeleteIncrementalRewritingRule.java @@ -61,7 +61,7 @@ import org.apache.hadoop.hive.ql.optimizer.calcite.rules.HiveHepExtractRelNodeRu * JOIN TAB_B ON (TAB_A.a = TAB_B.z) * WHERE TAB_A.ROW_ID > 5 * GROUP BY a, b) source - * ON (mv.a <=> source.a AND mv.b <=> source.b) + * ON (mv.a <=> source.a AND mv.b <=> source.b) * WHEN MATCHED AND mv.c + source.c <> 0 * THEN UPDATE SET mv.s = mv.s + source.s, mv.c = mv.c + source.c * WHEN MATCHED AND countStar = 0 THEN DELETE @@ -70,13 +70,13 @@ import org.apache.hadoop.hive.ql.optimizer.calcite.rules.HiveHepExtractRelNodeRu * * To be precise, we need to convert it into a MERGE rewritten as: * FROM (select *, true flag from mv) mv right outer join _source_ source - * ON (mv.a <=> source.a AND mv.b <=> source.b) - * INSERT INTO TABLE mv <- (insert new rows into the view) + * ON (mv.a <=> source.a AND mv.b <=> source.b) + * INSERT INTO TABLE mv <- (insert new rows into the view) * SELECT source.a, source.b, s, c * WHERE mv.flag IS NULL - * INSERT INTO TABLE mv <- (update existing rows in the view) + * INSERT INTO TABLE mv <- (update existing rows in the view) * SELECT mv.ROW__ID, source.a, source.b, - * CASE WHEN mv.s IS NULL AND source.s IS NULL THEN NULL <- (use case expression to handle nulls from both sides) + * CASE WHEN mv.s IS NULL AND source.s IS NULL THEN NULL <- (use case expression to handle nulls from both sides) * WHEN mv.s IS NULL THEN source.s * WHEN source.s IS NULL THEN mv.s * ELSE mv.s + source.s END, @@ -86,12 +86,12 @@ import org.apache.hadoop.hive.ql.optimizer.calcite.rules.HiveHepExtractRelNodeRu * ELSE mv.s + source.s END countStar, * WHERE mv.flag AND countStar <> 0 * SORT BY mv.ROW__ID; - * INSERT INTO TABLE mv <- (delete from the views) + * INSERT INTO TABLE mv <- (delete from the views) * SELECT mv.ROW__ID * WHERE mv.flag AND countStar = 0 * SORT BY mv.ROW__ID; * - * {@see CalcitePlanner#fixUpASTAggregateInsertDeleteIncrementalRebuild} + * @see org.apache.hadoop.hive.ql.parse.CalcitePlanner */ public class HiveAggregateInsertDeleteIncrementalRewritingRule extends HiveAggregateIncrementalRewritingRuleBase< HiveAggregateInsertDeleteIncrementalRewritingRule.IncrementalComputePlanWithDeletedRows> { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveAggregateInsertIncrementalRewritingRule.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveAggregateInsertIncrementalRewritingRule.java index 60d444588b..18d7abcafb 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveAggregateInsertIncrementalRewritingRule.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveAggregateInsertIncrementalRewritingRule.java @@ -57,7 +57,7 @@ import java.util.List; * JOIN TAB_B ON (TAB_A.a = TAB_B.z) * WHERE TAB_A.ROW_ID > 5 * GROUP BY a, b) source - * ON (mv.a <=> source.a AND mv.b <=> source.b) + * ON (mv.a <=> source.a AND mv.b <=> source.b) * WHEN MATCHED AND mv.c + source.c <> 0 * THEN UPDATE SET mv.s = mv.s + source.s, mv.c = mv.c + source.c * WHEN NOT MATCHED @@ -65,7 +65,7 @@ import java.util.List; * * To be precise, we need to convert it into a MERGE rewritten as: * FROM (select *, true flag from mv) mv right outer join _source_ source - * ON (mv.a <=> source.a AND mv.b <=> source.b) + * ON (mv.a <=> source.a AND mv.b <=> source.b) * INSERT INTO TABLE mv * SELECT source.a, source.b, s, c * WHERE mv.flag IS NULL @@ -79,7 +79,7 @@ import java.util.List; * WHERE mv.flag * SORT BY mv.ROW__ID; * - * {@see CalcitePlanner#fixUpASTAggregateInsertIncrementalRebuild} + * @see org.apache.hadoop.hive.ql.parse.CalcitePlanner */ public class HiveAggregateInsertIncrementalRewritingRule extends HiveAggregateIncrementalRewritingRuleBase< HiveAggregateIncrementalRewritingRuleBase.IncrementalComputePlan> { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveAggregatePartitionIncrementalRewritingRule.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveAggregatePartitionIncrementalRewritingRule.java index 0d24cde8e6..7e582ed29b 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveAggregatePartitionIncrementalRewritingRule.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveAggregatePartitionIncrementalRewritingRule.java @@ -59,9 +59,9 @@ import static org.apache.hadoop.hive.ql.optimizer.calcite.HiveCalciteUtil.findRe * * SELECT a, b, sum(sumc) FROM ( * SELECT a, b, sumc FROM mat1 - * LEFT SEMI JOIN (SELECT a, b, sum(c) FROM t1 WHERE ROW__ID.writeId > 1 GROUP BY b, a) q ON (mat1.a <=> q.a) + * LEFT SEMI JOIN (SELECT a, b, sum(c) FROM t1 WHERE ROW__ID.writeId > 1 GROUP BY b, a) q ON (mat1.a <=> q.a) * UNION ALL - * SELECT a, b, sum(c) sumc FROM t1 WHERE ROW__ID.writeId > 1 GROUP BY b, a + * SELECT a, b, sum(c) sumc FROM t1 WHERE ROW__ID.writeId > 1 GROUP BY b, a * ) sub * GROUP BY b, a */ diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveJoinInsertDeleteIncrementalRewritingRule.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveJoinInsertDeleteIncrementalRewritingRule.java index 16b3ea19f2..57274e55d8 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveJoinInsertDeleteIncrementalRewritingRule.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/views/HiveJoinInsertDeleteIncrementalRewritingRule.java @@ -44,11 +44,11 @@ import java.util.List; * MULTI INSERT statement instead: one insert branch for inserted rows * and another for inserting deleted rows to delete delta. * Since CBO plan does not contain the INSERT branches we focus on the SELECT part of the plan in this rule. - * See also {@link CalcitePlanner#fixUpASTJoinInsertDeleteIncrementalRebuild(ASTNode)} + * See also {@link CalcitePlanner} * * FROM (select mv.ROW__ID, mv.a, mv.b from mv) mv * RIGHT OUTER JOIN (SELECT _source_.ROW__IS_DELETED,_source_.a, _source_.b FROM _source_) source - * ON (mv.a <=> source.a AND mv.b <=> source.b) + * ON (mv.a <=> source.a AND mv.b <=> source.b) * INSERT INTO TABLE mv_delete_delta * SELECT mv.ROW__ID * WHERE source.ROW__IS__DELETED diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java index 87599815e5..89d6024cc0 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java @@ -532,9 +532,9 @@ public class RexNodeConverter { * </pre> * Or: * <pre> - * (c,d) IN ( (v1,v2), (v3,v4), ...) => (c=v1 && d=v2) || (c=v3 && d=v4) || ... + * (c,d) IN ( (v1,v2), (v3,v4), ...) => (c=v1 && d=v2) || (c=v3 && d=v4) || ... * Input: ((c,d), (v1,v2), (v3,v4), ...) - * Output: (c=v1 && d=v2, c=v3 && d=v4, ...) + * Output: (c=v1 && d=v2, c=v3 && d=v4, ...) * </pre> * * Returns null if the transformation fails, e.g., when non-deterministic diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/topnkey/CommonKeyPrefix.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/topnkey/CommonKeyPrefix.java index 049466050b..f1b6da9568 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/topnkey/CommonKeyPrefix.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/topnkey/CommonKeyPrefix.java @@ -78,10 +78,10 @@ public final class CommonKeyPrefix { * * opKeys: Column[_col0], Column[_col1], Column[_col2], Column[_col3] * parentKeys: Column[KEY._col0], Column[KEY._col1], Column[KEY._col4] - * parentColExprMap: {_col0 -> Column[KEY._col0]}, {_col1 -> Column[KEY._col1]}, {_col4 -> Column[KEY._col4]} + * parentColExprMap: {_col0 -> Column[KEY._col0]}, {_col1 -> Column[KEY._col1]}, {_col4 -> Column[KEY._col4]} * * Column ordering and null ordering is given by a string where each character represents a column order/null order. - * Ex.: a ASC NULLS FIRST, b DESC NULLS LAST, c ASC NULLS LAST -> order="+-+", null order="azz" + * Ex.: a ASC NULLS FIRST, b DESC NULLS LAST, c ASC NULLS LAST -> order="+-+", null order="azz" * * When <code>parentColExprMap</code> is null this method falls back to * {@link #map(List, String, String, List, String, String)}. @@ -90,7 +90,7 @@ public final class CommonKeyPrefix { * @param opOrder operator's key column ordering in {@link String} format * @param opNullOrder operator's key column null ordering in {@link String} format * @param parentKeys {@link List} of {@link ExprNodeDesc}. contains the parent operator's key columns - * @param parentColExprMap {@link Map} of {@link String} -> {@link ExprNodeDesc}. + * @param parentColExprMap {@link Map} of {@link String} -> {@link ExprNodeDesc}. * contains parent operator's key column name {@link ExprNodeDesc} mapping * @param parentOrder parent operator's key column ordering in {@link String} format * @param parentNullOrder parent operator's key column null ordering in {@link String} format diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/topnkey/TopNKeyProcessor.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/topnkey/TopNKeyProcessor.java index 7aa2a54ecd..47fb9cb770 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/topnkey/TopNKeyProcessor.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/topnkey/TopNKeyProcessor.java @@ -39,7 +39,7 @@ import org.slf4j.LoggerFactory; /** * TopNKeyProcessor is a processor for TopNKeyOperator. - * A TopNKeyOperator will be placed before any ReduceSinkOperator which has a topN property >= 0. + * A TopNKeyOperator will be placed before any ReduceSinkOperator which has a topN property >= 0. */ public class TopNKeyProcessor implements SemanticNodeProcessor { private static final Logger LOG = LoggerFactory.getLogger(TopNKeyProcessor.class); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java index 59a1fdf347..6ba9a9075f 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java @@ -2405,7 +2405,6 @@ public class CalcitePlanner extends SemanticAnalyzer { * @param isCollection * @param order * @param rules - * @return HEP program */ protected void generatePartialProgram(HepProgramBuilder programBuilder, boolean isCollection, HepMatchOrder order, RelOptRule... rules) { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/UnparseTranslator.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/UnparseTranslator.java index 0e0b8d6db3..8c5cd852b3 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/UnparseTranslator.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/UnparseTranslator.java @@ -137,7 +137,7 @@ public class UnparseTranslator { /** * Register a translation for an tabName. * - * @param node + * @param tableName * source node (which must be an tabName) to be replaced */ public void addTableNameTranslation(ASTNode tableName, String currentDatabaseName) { @@ -175,7 +175,7 @@ public class UnparseTranslator { /** * Register a translation for an identifier. * - * @param node + * @param identifier * source node (which must be an identifier) to be replaced */ public void addIdentifierTranslation(ASTNode identifier) { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/type/FunctionHelper.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/type/FunctionHelper.java index 3797786109..d4e1dbcdac 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/type/FunctionHelper.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/type/FunctionHelper.java @@ -98,7 +98,7 @@ public interface FunctionHelper { boolean isInFunction(FunctionInfo fi); /** - * returns true if FunctionInfo is a compare function (e.g. '<=') + * returns true if FunctionInfo is a compare function (e.g. '<=') */ boolean isCompareFunction(FunctionInfo fi); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/CompactorThread.java b/ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/CompactorThread.java index 796ef72c19..167b2728f5 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/CompactorThread.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/CompactorThread.java @@ -106,7 +106,7 @@ public abstract class CompactorThread extends Thread implements Configurable { * Get the partition being compacted. * @param ci compaction info returned from the compaction queue * @return metastore partition, or null if there is not partition in this compaction info - * @throws Exception if underlying calls throw, or if the partition name resolves to more than + * @throws MetaException if underlying calls throw, or if the partition name resolves to more than * one partition. */ protected Partition resolvePartition(CompactionInfo ci) throws MetaException { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFRank.java b/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFRank.java index 644e7283c0..a6eb3752b4 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFRank.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFRank.java @@ -274,7 +274,7 @@ public class GenericUDAFRank extends AbstractGenericUDAFResolver { * Calculates the rank of a hypothetical row specified by the arguments of the * function in a group of values specified by the order by clause. * SELECT rank(expression1[, expressionn]*) WITHIN GROUP (ORDER BY col1[, coln]*) - * (the number of rows where col1 < expression1 [and coln < expressionn]*) + 1 + * (the number of rows where col1 < expression1 [and coln < expressionn]*) + 1 */ public static class GenericUDAFHypotheticalSetRankEvaluator extends GenericUDAFEvaluator { public static final String RANK_FIELD = "rank"; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCastFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCastFormat.java index 6e24f9cf0a..36fdae15e7 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCastFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCastFormat.java @@ -53,7 +53,7 @@ import java.io.Serializable; import java.util.Map; /** - * CAST(<value> AS <type> FORMAT <STRING>). + * CAST(<value> AS <type> FORMAT <STRING>). * * Vector expressions: CastDateToCharWithFormat, CastDateToStringWithFormat, * CastDateToVarCharWithFormat, CastTimestampToCharWithFormat, diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/JsonSerDe.java b/serde/src/java/org/apache/hadoop/hive/serde2/JsonSerDe.java index 14d5c6ea78..534246e5f3 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/JsonSerDe.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/JsonSerDe.java @@ -41,10 +41,10 @@ import org.apache.hive.common.util.TimestampParser; /** * Hive SerDe for processing JSON formatted data. This is typically paired with * the TextInputFormat and therefore each line provided to this SerDe must be a - * single, and complete JSON object.<br/> + * single, and complete JSON object.<br> * <h2>Example</h2> * <p> - * {"name="john","age"=30}<br/> + * {"name="john","age"=30}<br> * {"name="sue","age"=32} * </p> */ diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/json/BinaryEncoding.java b/serde/src/java/org/apache/hadoop/hive/serde2/json/BinaryEncoding.java index 02f26b0d74..255562a8d2 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/json/BinaryEncoding.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/json/BinaryEncoding.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.serde2.json; /** - * Enums describing the available String->Bytes encoding available for JSON + * Enums describing the available String->Bytes encoding available for JSON * parsing. This base-64 variant is what most people would think of "the * standard" Base64 encoding for JSON: the specific MIME content transfer * encoding. The Raw String encoding produces an array of bytes by reading the diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/json/HiveJsonReader.java b/serde/src/java/org/apache/hadoop/hive/serde2/json/HiveJsonReader.java index 6e9bb4e226..7ade47b6e5 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/json/HiveJsonReader.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/json/HiveJsonReader.java @@ -71,9 +71,9 @@ import com.google.common.base.Preconditions; /** * This class converts JSON strings into Java or Hive Primitive objects. * - * Support types are:<br/> - * <br/> - * <table border="1"> + * Support types are:<br> + * <br> + * <table border="1" summary=""> * <tr> * <th>JSON Type</th> * <th>Java Type</th> diff --git a/service/src/java/org/apache/hive/service/cli/operation/QueryInfoCache.java b/service/src/java/org/apache/hive/service/cli/operation/QueryInfoCache.java index 59b5cd029b..6dff3477b7 100644 --- a/service/src/java/org/apache/hive/service/cli/operation/QueryInfoCache.java +++ b/service/src/java/org/apache/hive/service/cli/operation/QueryInfoCache.java @@ -78,7 +78,7 @@ public class QueryInfoCache { } /** - * Remove the live operation's query info from the {@liveQueryInfos}, + * Remove the live operation's query info from the {@link #liveQueryInfos}, * and push the query info to the historic query cache if enabled. * @param operation the to remove operation */ diff --git a/shims/common/src/main/java/org/apache/hadoop/hive/shims/HadoopShims.java b/shims/common/src/main/java/org/apache/hadoop/hive/shims/HadoopShims.java index 242990abf0..5696393d6c 100644 --- a/shims/common/src/main/java/org/apache/hadoop/hive/shims/HadoopShims.java +++ b/shims/common/src/main/java/org/apache/hadoop/hive/shims/HadoopShims.java @@ -259,8 +259,6 @@ public interface HadoopShims { * @param filter A filter to use on the files in the directory * @return A list of file status with IDs * @throws IOException An I/O exception of some sort has occurred - * @throws FileNotFoundException If the path is not found in the - * {@code FileSystem} * @throws UnsupportedOperationException the {@code FileSystem} is not a * {@code DistributedFileSystem} */ diff --git a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java index 5e8a81f466..81f6083ce4 100644 --- a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java +++ b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java @@ -4748,7 +4748,7 @@ public class HiveMetaStoreClient implements IMetaStoreClient, AutoCloseable { * if the validWriteIdList is not explicitly passed (as a method argument) to the HMS APIs. * This method returns the ValidWriteIdList based on the VALID_TABLES_WRITEIDS_KEY key. * Since, VALID_TABLES_WRITEIDS_KEY is set during the lock acquisition phase after query compilation - * ( DriverTxnHandler.acquireLocks -> recordValidWriteIds -> setValidWriteIds ), + * ( DriverTxnHandler.acquireLocks -> recordValidWriteIds -> setValidWriteIds ), * this only covers a subset of cases, where we invoke get_* APIs after query compilation, * if the validWriteIdList is not explicitly passed (as a method argument) to the HMS APIs. */ diff --git a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java index 3056a598fb..eaa4542de4 100644 --- a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java +++ b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java @@ -252,9 +252,7 @@ public interface IMetaStoreClient { * @param dbName Name of the database to fetch tables from. * @param tablePattern pattern to match the tables names. * @param requestedFields An int bitmask to indicate the depth of the returned objects - * @param processorCapabilities A list of "capabilities" possessed by the caller, to be matched with table's params - * @param processorId Any string id to identify the caller/client, for logging purposes only. - * @param limit Maximum size of the result set. <=0 indicates no limit + * @param limit Maximum size of the result set. <=0 indicates no limit * @return List of ExtendedTableInfo that match the input arguments. * @throws MetaException Thrown if there is error on fetching from DBMS. * @throws TException Thrown if there is a thrift transport exception. @@ -1689,9 +1687,7 @@ public interface IMetaStoreClient { * @throws MetaException * @throws NoSuchObjectException * @throws TException - * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#create_table(org.apache.hadoop.hive.metastore.api.CreateTableRequest) */ - void createTable(Table tbl) throws AlreadyExistsException, InvalidObjectException, MetaException, NoSuchObjectException, TException; @@ -1703,7 +1699,6 @@ public interface IMetaStoreClient { * @throws NoSuchObjectException * @throws TException */ - void createTable(CreateTableRequest request) throws AlreadyExistsException, InvalidObjectException, MetaException, NoSuchObjectException, TException; @@ -2583,7 +2578,7 @@ public interface IMetaStoreClient { /** * Get the column statistics for a set of columns in a table. This should only be used for * non-partitioned tables. For partitioned tables use - * {@link #getPartitionColumnStatistics(String, String, List, List)}. + * {@link #getPartitionColumnStatistics(String, String, List, List, String)}. * @param dbName database name * @param tableName table name * @param colNames list of column names @@ -2603,7 +2598,7 @@ public interface IMetaStoreClient { /** * Get the column statistics for a set of columns in a table. This should only be used for * non-partitioned tables. For partitioned tables use - * {@link #getPartitionColumnStatistics(String, String, String, List, List)}. + * {@link #getPartitionColumnStatistics(String, String, String, List, List, String)}. * @param catName catalog name * @param dbName database name * @param tableName table name diff --git a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/FileUtils.java b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/FileUtils.java index b8ab28b45c..cf8825ec16 100644 --- a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/FileUtils.java +++ b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/FileUtils.java @@ -353,7 +353,6 @@ public class FileUtils { * @param fs * the file system * @return array of FileStatus - * @throws IOException */ public static List<FileStatus> getFileStatusRecurse(Path base, FileSystem fs) { try { diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ExceptionHandler.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ExceptionHandler.java index ffac60831e..4e4950f99f 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ExceptionHandler.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ExceptionHandler.java @@ -81,8 +81,8 @@ public final class ExceptionHandler { } /** - * Converts the input exception to the target instance of class {@param target}, - * if the input exception is the instance of class {@param source}, throws the + * Converts the input exception to the target instance of class {@code target}, + * if the input exception is the instance of class {@code source}, throws the * converted target exception. */ public <S extends Exception, T extends TException> ExceptionHandler diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreMetadataTransformer.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreMetadataTransformer.java index 6f0bf91df2..4360f09dfa 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreMetadataTransformer.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreMetadataTransformer.java @@ -41,7 +41,7 @@ public interface IMetaStoreMetadataTransformer { * @param processorCapabilities A array of String capabilities received from the data processor * @param processorId String ID used for logging purpose. * @return Map A Map of transformed objects keyed by Table and value is list of required capabilities - * @throws HiveMetaException + * @throws MetaException */ // TODO HiveMetaException or MetaException public Map<Table, List<String>> transform(List<Table> tables, List<String> processorCapabilities, @@ -53,7 +53,7 @@ public interface IMetaStoreMetadataTransformer { * @param processorCapabilities A array of String capabilities received from the data processor * @param processorId String ID used for logging purpose. * @return Map A Map of transformed objects keyed by Partition and value is list of required capabilities - * @throws HiveMetaException + * @throws MetaException */ // TODO HiveMetaException or MetaException public List<Partition> transformPartitions(List<Partition> parts, Table table, List<String> processorCapabilities, @@ -64,7 +64,7 @@ public interface IMetaStoreMetadataTransformer { * @param processorCapabilities A array of String capabilities received from the data processor * @param processorId String ID used for logging purpose. * @return Table An altered Table based on the processor capabilities - * @throws HiveMetaException + * @throws MetaException */ public Table transformCreateTable(Table table, List<String> processorCapabilities, String processorId) throws MetaException; @@ -74,17 +74,17 @@ public interface IMetaStoreMetadataTransformer { * @param processorCapabilities A array of String capabilities received from the data processor * @param processorId String ID used for logging purpose. * @return Database An altered Database based on the processor capabilities - * @throws HiveMetaException + * @throws MetaException */ public Database transformDatabase(Database db, List<String> processorCapabilities, String processorId) throws MetaException; /** - * @param table A table object to be transformed prior to the alteration of a table + * @param oldTable A table object to be transformed prior to the alteration of a table * @param processorCapabilities A array of String capabilities received from the data processor * @param processorId String ID used for logging purpose. * @return Table An altered Table based on the processor capabilities - * @throws HiveMetaException + * @throws MetaException */ public Table transformAlterTable(Table oldTable, Table newTable, List<String> processorCapabilities, String processorId) throws MetaException; diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreEventListener.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreEventListener.java index feb171a384..800ea03b81 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreEventListener.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreEventListener.java @@ -343,7 +343,7 @@ public abstract class MetaStoreEventListener implements Configurable { /** * This will be called to perform acid write operation in a batch. - * @param acidWriteEvent event to be processed + * @param batchAcidWriteEvent event to be processed * @param dbConn jdbc connection to remote meta store db. * @param sqlGenerator helper class to generate db specific sql string. * @throws MetaException diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/RawStore.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/RawStore.java index e886098d45..fac5b836e2 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/RawStore.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/RawStore.java @@ -290,7 +290,7 @@ public interface RawStore extends Configurable { * Get the dataconnector with a given name, if exists. * @param dcName pattern names should match * @return DataConnector object. - * @throws MetaException something went wrong, usually with the database. + * @throws NoSuchObjectException something went wrong, usually with the database. */ DataConnector getDataConnector(String dcName) throws NoSuchObjectException; @@ -542,7 +542,6 @@ public interface RawStore extends Configurable { * Retrieve all materialized views. * @return all materialized views in a catalog * @throws MetaException error querying the RDBMS - * @throws NoSuchObjectException no such database */ List<Table> getAllMaterializedViewObjectsForRewriting(String catName) throws MetaException; @@ -1638,7 +1637,7 @@ public interface RawStore extends Configurable { /** * SQLPrimaryKey represents a single primary key column. * Since a table can have one or more primary keys ( in case of composite primary key ), - * this method returns List<SQLPrimaryKey> + * this method returns List<SQLPrimaryKey> * @param request primary key request * @return list of primary key columns or an empty list if the table does not have a primary key * @throws MetaException error accessing the RDBMS @@ -1667,7 +1666,7 @@ public interface RawStore extends Configurable { /** * SQLForeignKey represents a single foreign key column. * Since a table can have one or more foreign keys ( in case of composite foreign key ), - * this method returns List<SQLForeignKey> + * this method returns List<SQLForeignKey> * @param request ForeignKeysRequest object * @return List of all matching foreign key columns. Note that if more than one foreign key * matches the arguments the results here will be all mixed together into a single list. @@ -1691,7 +1690,7 @@ public interface RawStore extends Configurable { /** * SQLUniqueConstraint represents a single unique constraint column. * Since a table can have one or more unique constraint ( in case of composite unique constraint ), - * this method returns List<SQLUniqueConstraint> + * this method returns List<SQLUniqueConstraint> * @param request UniqueConstraintsRequest object. * @return list of unique constraints * @throws MetaException error access the RDBMS. @@ -1713,7 +1712,7 @@ public interface RawStore extends Configurable { /** * SQLNotNullConstraint represents a single not null constraint column. * Since a table can have one or more not null constraint ( in case of composite not null constraint ), - * this method returns List<SQLNotNullConstraint> + * this method returns List<SQLNotNullConstraint> * @param request NotNullConstraintsRequest object. * @return list of not null constraints * @throws MetaException error accessing the RDBMS. @@ -1735,7 +1734,7 @@ public interface RawStore extends Configurable { /** * SQLDefaultConstraint represents a single default constraint column. * Since a table can have one or more default constraint ( in case of composite default constraint ), - * this method returns List<SQLDefaultConstraint> + * this method returns List<SQLDefaultConstraint> * @param request DefaultConstraintsRequest object. * @return list of default values defined on the table. * @throws MetaException error accessing the RDBMS @@ -1757,7 +1756,7 @@ public interface RawStore extends Configurable { /** * SQLCheckConstraint represents a single check constraint column. * Since a table can have one or more check constraint ( in case of composite check constraint ), - * this method returns List<SQLCheckConstraint> + * this method returns List<SQLCheckConstraint> * @param request CheckConstraintsRequest object. * @return ccheck constraints for this table * @throws MetaException error accessing the RDBMS diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/dataconnector/jdbc/DerbySQLConnectorProvider.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/dataconnector/jdbc/DerbySQLConnectorProvider.java index a6d53752b3..5a8dac3524 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/dataconnector/jdbc/DerbySQLConnectorProvider.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/dataconnector/jdbc/DerbySQLConnectorProvider.java @@ -43,7 +43,7 @@ public class DerbySQLConnectorProvider extends AbstractJDBCConnectorProvider { /** * Returns a list of all table names from the remote database. * @return List A collection of all the table names, null if there are no tables. - * @throws IOException To indicate any failures with executing this API + * @throws MetaException To indicate any failures with executing this API */ @Override protected ResultSet fetchTableNames() throws MetaException { diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnStore.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnStore.java index 71df48993d..9d2ebc7b09 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnStore.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnStore.java @@ -746,7 +746,7 @@ public interface TxnStore extends Configurable { /** * Returns the top ACID metrics from each type {@link CompactionMetricsData.MetricType} - * @oaram limit number of returned records for each type + * @param limit number of returned records for each type * @return list of metrics, always non-null * @throws MetaException */