This is an automated email from the ASF dual-hosted git repository. kgyrtkirk pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/hive.git
The following commit(s) were added to refs/heads/master by this push: new 989e72a HIVE-24390: Spelling fixes - serde (#2802) (Josh Soref reviewed by Zoltan Haindrich) 989e72a is described below commit 989e72a393356c5f91f96d1bab6455a4c75c77a7 Author: Josh Soref <2119212+jso...@users.noreply.github.com> AuthorDate: Tue Nov 23 03:14:51 2021 -0500 HIVE-24390: Spelling fixes - serde (#2802) (Josh Soref reviewed by Zoltan Haindrich) --- .../src/java/org/apache/hadoop/hive/serde2/RandomTypeUtil.java | 6 +++--- .../java/org/apache/hadoop/hive/serde2/SerDeStatsStruct.java | 2 +- serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java | 2 +- .../org/apache/hadoop/hive/serde2/avro/AvroDeserializer.java | 8 ++++---- .../org/apache/hadoop/hive/serde2/avro/AvroSerializer.java | 2 +- .../org/apache/hadoop/hive/serde2/columnar/ColumnarSerDe.java | 2 +- .../org/apache/hadoop/hive/serde2/io/HiveCharWritable.java | 2 +- .../org/apache/hadoop/hive/serde2/json/HiveJsonReader.java | 2 +- .../org/apache/hadoop/hive/serde2/json/HiveJsonWriter.java | 8 ++++---- .../java/org/apache/hadoop/hive/serde2/lazy/LazyFactory.java | 2 +- .../org/apache/hadoop/hive/serde2/lazy/LazyHiveDecimal.java | 2 +- .../java/org/apache/hadoop/hive/serde2/lazy/LazyTimestamp.java | 2 +- .../apache/hadoop/hive/serde2/lazy/fast/StringToDouble.java | 2 +- .../apache/hadoop/hive/serde2/lazybinary/LazyBinaryArray.java | 6 +++--- .../apache/hadoop/hive/serde2/lazybinary/LazyBinarySerDe2.java | 2 +- .../hive/serde2/objectinspector/ListObjectsEqualComparer.java | 2 +- .../hive/serde2/objectinspector/ObjectInspectorConverters.java | 2 +- .../hive/serde2/objectinspector/ObjectInspectorUtils.java | 4 ++-- .../primitive/PrimitiveObjectInspectorUtils.java | 2 +- .../hive/serde2/teradata/TeradataBinaryDataInputStream.java | 2 +- .../hive/serde2/teradata/TeradataBinaryDataOutputStream.java | 2 +- .../hadoop/hive/serde2/teradata/TeradataBinarySerde.java | 4 ++-- .../hadoop/hive/serde2/thrift/TCTLSeparatedProtocol.java | 2 +- .../hadoop/hive/serde2/typeinfo/TimestampLocalTZTypeInfo.java | 2 +- .../org/apache/hadoop/hive/serde2/typeinfo/TypeInfoUtils.java | 6 +++--- .../org/apache/hadoop/hive/serde2/SerdeRandomRowSource.java | 2 +- .../src/test/org/apache/hadoop/hive/serde2/TestJsonSerDe.java | 4 ++-- .../apache/hadoop/hive/serde2/TestTCTLSeparatedProtocol.java | 10 +++++----- .../apache/hadoop/hive/serde2/avro/TestAvroDeserializer.java | 2 +- .../hive/serde2/avro/TestAvroObjectInspectorGenerator.java | 8 ++++---- .../org/apache/hadoop/hive/serde2/avro/TestAvroSerdeUtils.java | 4 ++-- .../hive/serde2/binarysortable/TestBinarySortableFast.java | 2 +- .../org/apache/hadoop/hive/serde2/lazy/TestLazySimpleFast.java | 2 +- .../hadoop/hive/serde2/lazybinary/TestLazyBinaryFast.java | 2 +- .../serde2/objectinspector/TestObjectInspectorConverters.java | 4 ++-- 35 files changed, 59 insertions(+), 59 deletions(-) diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/RandomTypeUtil.java b/serde/src/java/org/apache/hadoop/hive/serde2/RandomTypeUtil.java index ad9de4c..c0e0583 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/RandomTypeUtil.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/RandomTypeUtil.java @@ -131,7 +131,7 @@ public class RandomTypeUtil { public static final long NANOSECONDS_PER_SECOND = TimeUnit.SECONDS.toNanos(1); public static final long MILLISECONDS_PER_SECOND = TimeUnit.SECONDS.toMillis(1); - public static final long NANOSECONDS_PER_MILLISSECOND = TimeUnit.MILLISECONDS.toNanos(1); + public static final long NANOSECONDS_PER_MILLISECOND = TimeUnit.MILLISECONDS.toNanos(1); private static final ThreadLocal<DateFormat> DATE_FORMAT = new ThreadLocal<DateFormat>() { @@ -172,12 +172,12 @@ public class RandomTypeUtil { case 2: // Limit to milliseconds only... optionalNanos = String.format(".%09d", - Integer.valueOf(r.nextInt((int) MILLISECONDS_PER_SECOND)) * NANOSECONDS_PER_MILLISSECOND); + Integer.valueOf(r.nextInt((int) MILLISECONDS_PER_SECOND)) * NANOSECONDS_PER_MILLISECOND); break; case 3: // Limit to below milliseconds only... optionalNanos = String.format(".%09d", - Integer.valueOf(r.nextInt((int) NANOSECONDS_PER_MILLISSECOND))); + Integer.valueOf(r.nextInt((int) NANOSECONDS_PER_MILLISECOND))); break; } String timestampStr = String.format("%04d-%02d-%02d %02d:%02d:%02d%s", diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/SerDeStatsStruct.java b/serde/src/java/org/apache/hadoop/hive/serde2/SerDeStatsStruct.java index cf86e5e..8f0fa9d 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/SerDeStatsStruct.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/SerDeStatsStruct.java @@ -21,7 +21,7 @@ package org.apache.hadoop.hive.serde2; public interface SerDeStatsStruct { /** - * Rerurns the serialized size of the object. + * Returns the serialized size of the object. */ public long getRawDataSerializedSize(); diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java b/serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java index 8156fee..849425e 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java @@ -112,7 +112,7 @@ public final class SerDeUtils { escape.append('t'); break; default: - // Control characeters! According to JSON RFC u0020 + // Control characters! According to JSON RFC u0020 if (c < ' ') { String hex = Integer.toHexString(c); escape.append('\\'); diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroDeserializer.java b/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroDeserializer.java index 60ac3f3..11d6627 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroDeserializer.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroDeserializer.java @@ -158,7 +158,7 @@ class AvroDeserializer { * @param writable Instance of GenericAvroWritable to deserialize * @param readerSchema Schema of the writable to deserialize * @return A list of objects suitable for Hive to work with further - * @throws AvroSerdeException For any exception during deseriliazation + * @throws AvroSerdeException For any exception during deserialization */ public Object deserialize(List<String> columnNames, List<TypeInfo> columnTypes, Writable writable, Schema readerSchema) throws AvroSerdeException { @@ -418,16 +418,16 @@ class AvroDeserializer { // and we would end up doing calculations twice to get the same tag int fsTag = GenericData.get().resolveUnion(fileSchema, datum); // Determine index of value from fileSchema int rsTag = GenericData.get().resolveUnion(recordSchema, datum); // Determine index of value from recordSchema - Object desered = worker(datum, fileSchema == null ? null : fileSchema.getTypes().get(fsTag), + Object desired = worker(datum, fileSchema == null ? null : fileSchema.getTypes().get(fsTag), recordSchema.getTypes().get(rsTag), columnType.getAllUnionObjectTypeInfos().get(rsTag)); - return new StandardUnionObjectInspector.StandardUnion((byte)rsTag, desered); + return new StandardUnionObjectInspector.StandardUnion((byte)rsTag, desired); } private Object deserializeList(Object datum, Schema fileSchema, Schema recordSchema, ListTypeInfo columnType) throws AvroSerdeException { // Need to check the original schema to see if this is actually a Fixed. if(recordSchema.getType().equals(Schema.Type.FIXED)) { - // We're faking out Hive to work through a type system impedence mismatch. + // We're faking out Hive to work through a type system impedance mismatch. // Pull out the backing array and convert to a list. GenericData.Fixed fixed = (GenericData.Fixed) datum; List<Byte> asList = new ArrayList<Byte>(fixed.bytes().length); diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerializer.java b/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerializer.java index eb04701..409ad9d 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerializer.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerializer.java @@ -161,7 +161,7 @@ class AvroSerializer { /** private cache to avoid lots of EnumSymbol creation while serializing. * Two levels because the enum symbol is specific to a schema. - * Object because we want to avoid the overhead of repeated toString calls while maintaining compatability. + * Object because we want to avoid the overhead of repeated toString calls while maintaining compatibility. * Provided there are few enum types per record, and few symbols per enum, memory use should be moderate. * eg 20 types with 50 symbols each as length-10 Strings should be on the order of 100KB per AvroSerializer. */ diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/columnar/ColumnarSerDe.java b/serde/src/java/org/apache/hadoop/hive/serde2/columnar/ColumnarSerDe.java index 5e8277d..fe678e7 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/columnar/ColumnarSerDe.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/columnar/ColumnarSerDe.java @@ -88,7 +88,7 @@ public class ColumnarSerDe extends ColumnarSerDeBase { serdeParams = new LazySerDeParameters(configuration, properties, getClass().getName()); // Create the ObjectInspectors for the fields. Note: Currently - // ColumnarObject uses same ObjectInpector as LazyStruct + // ColumnarObject uses same ObjectInspector as LazyStruct cachedObjectInspector = LazyFactory.createColumnarStructInspector( serdeParams.getColumnNames(), serdeParams.getColumnTypes(), serdeParams); diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/io/HiveCharWritable.java b/serde/src/java/org/apache/hadoop/hive/serde2/io/HiveCharWritable.java index 70f907f..fcd14e5 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/io/HiveCharWritable.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/io/HiveCharWritable.java @@ -26,7 +26,7 @@ import org.apache.hive.common.util.HiveStringUtils; /** * HiveCharWritable. * String values will be padded to full char length. - * Character legnth, comparison, hashCode should ignore trailing spaces. + * Character length, comparison, hashCode should ignore trailing spaces. */ public class HiveCharWritable extends HiveBaseCharWritable implements WritableComparable<HiveCharWritable> { diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/json/HiveJsonReader.java b/serde/src/java/org/apache/hadoop/hive/serde2/json/HiveJsonReader.java index b24ca0c..6e9bb4e 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/json/HiveJsonReader.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/json/HiveJsonReader.java @@ -469,7 +469,7 @@ public class HiveJsonReader { /** * Matches the JSON object's field name with the Hive data type. * - * @param oi The ObjectInsepctor to lookup the matching in + * @param oi The ObjectInspector to lookup the matching in * @param fieldName The name of the field parsed from the JSON text * @return The meta data of regarding this field * @throws SerDeException The SerDe is not configured correctly diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/json/HiveJsonWriter.java b/serde/src/java/org/apache/hadoop/hive/serde2/json/HiveJsonWriter.java index cacc70b..cf5c56b 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/json/HiveJsonWriter.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/json/HiveJsonWriter.java @@ -171,7 +171,7 @@ public class HiveJsonWriter { /** * Visit a vertex in the graph that is a Java Map. * - * @param oi The map's OjectInspector + * @param oi The map's ObjectInspector * @param o The Map object * @return A JsonNode representation of the Map * @throws SerDeException The Map cannot be parsed @@ -199,7 +199,7 @@ public class HiveJsonWriter { /** * Visit a vertex in the graph that is a Java List. * - * @param oi The list's OjectInspector + * @param oi The list's ObjectInspector * @param o The List object * @return A JsonNode representation of the List * @throws SerDeException The List cannot be parsed @@ -225,7 +225,7 @@ public class HiveJsonWriter { * represented as a Java List where the name associated with each element in * the list is stored in the ObjectInspector. * - * @param oi The struct's OjectInspector + * @param oi The struct's ObjectInspector * @param o The List object * @param fieldNames List of names to override the default field names * @return A JsonNode representation of the List @@ -256,7 +256,7 @@ public class HiveJsonWriter { /** * Visit a vertex in the graph that is a union data type. * - * @param oi The union's OjectInspector + * @param oi The union's ObjectInspector * @param o The Union object * @return A JsonNode representation of the union * @throws SerDeException The union cannot be parsed diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyFactory.java b/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyFactory.java index 883ba88..8b9c012 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyFactory.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyFactory.java @@ -84,7 +84,7 @@ public final class LazyFactory { * of the lazy object. * * @param poi PrimitiveObjectInspector - * @param typeBinary a switch to return either a LazyPrimtive class or it's binary + * @param typeBinary a switch to return either a LazyPrimitive class or it's binary * companion * @return LazyPrimitive<? extends ObjectInspector, ? extends Writable> */ diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyHiveDecimal.java b/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyHiveDecimal.java index 2ca023c..093872d 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyHiveDecimal.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyHiveDecimal.java @@ -58,7 +58,7 @@ public class LazyHiveDecimal extends LazyPrimitive<LazyHiveDecimalObjectInspecto } /** - * Initilizes LazyHiveDecimal object by interpreting the input bytes + * Initializes LazyHiveDecimal object by interpreting the input bytes * as a numeric string * * @param bytes diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestamp.java b/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestamp.java index b4f8718..2218942 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestamp.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestamp.java @@ -46,7 +46,7 @@ public class LazyTimestamp extends LazyPrimitive<LazyTimestampObjectInspector, T } /** - * Initilizes LazyTimestamp object by interpreting the input bytes + * Initializes LazyTimestamp object by interpreting the input bytes * as a JDBC timestamp string * * @param bytes diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/lazy/fast/StringToDouble.java b/serde/src/java/org/apache/hadoop/hive/serde2/lazy/fast/StringToDouble.java index 348faac..816a98f 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/lazy/fast/StringToDouble.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/lazy/fast/StringToDouble.java @@ -71,7 +71,7 @@ public class StringToDouble { int c; int exp = 0; /* Exponent read from "EX" field. */ int fracExp = 0; /* Exponent that derives from the fractional - * part. Under normal circumstatnces, it is + * part. Under normal circumstances, it is * the negative of the number of digits in F. * However, if I is very long, the last digits * of I get dropped (otherwise a long I with a diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryArray.java b/serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryArray.java index f741a96..4e67b39 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryArray.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryArray.java @@ -135,9 +135,9 @@ public class LazyBinaryArray extends // adjust arrays adjustArraySize(arraySize); // find out the null-bytes - int arryByteStart = start + vInt.length; - int nullByteCur = arryByteStart; - int nullByteEnd = arryByteStart + (arraySize + 7) / 8; + int arrayByteStart = start + vInt.length; + int nullByteCur = arrayByteStart; + int nullByteEnd = arrayByteStart + (arraySize + 7) / 8; // the begin the real elements int lastElementByteEnd = nullByteEnd; // the list element object inspector diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinarySerDe2.java b/serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinarySerDe2.java index 7a6e166..f04d790 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinarySerDe2.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinarySerDe2.java @@ -65,7 +65,7 @@ import org.apache.hadoop.io.Writable; /** * Subclass of LazyBinarySerDe with faster serialization, initializing a serializer based on the * row columns rather than checking the ObjectInspector category/primitiveType for every value. - * This appears to be around 3x faster than the LazyBinarSerDe serialization. + * This appears to be around 3x faster than the LazyBinarySerDe serialization. */ @SerDeSpec(schemaProps = {serdeConstants.LIST_COLUMNS, serdeConstants.LIST_COLUMN_TYPES}) public class LazyBinarySerDe2 extends LazyBinarySerDe { diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ListObjectsEqualComparer.java b/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ListObjectsEqualComparer.java index a87459b..e90c5f6 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ListObjectsEqualComparer.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ListObjectsEqualComparer.java @@ -32,7 +32,7 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils; * Compare two list of objects. * Two lists are expected to have same types. Type information for every object is * passed when calling Constructor to avoid the step of figuring out types from - * ObjectInspetor and determine how to compare the types when comparing. + * ObjectInspector and determine how to compare the types when comparing. * Also, for string and text elements, it performs slightly better than * using ObjectInspectorUtils.compare() == 0, which instead of calling .compare() * calls .equalTo(), which compares size before byte by byte comparison. diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorConverters.java b/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorConverters.java index 9129177..e7e0473 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorConverters.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorConverters.java @@ -439,7 +439,7 @@ public final class ObjectInspectorConverters { UnionObjectInspector inputOI; SettableUnionObjectInspector outputOI; - // Object inspectors for the tags for the input and output unionss + // Object inspectors for the tags for the input and output unions List<? extends ObjectInspector> inputTagsOIs; List<? extends ObjectInspector> outputTagsOIs; diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorUtils.java b/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorUtils.java index bf547de..747564a 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorUtils.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorUtils.java @@ -631,7 +631,7 @@ public final class ObjectInspectorUtils { /** * Computes the bucket number to which the bucketFields belong to * @param bucketFields the bucketed fields of the row - * @param bucketFieldInspectors the ObjectInpsectors for each of the bucketed fields + * @param bucketFieldInspectors the ObjectInspectors for each of the bucketed fields * @param totalBuckets the number of buckets in the table * @return the bucket number using Murmur hash */ @@ -642,7 +642,7 @@ public final class ObjectInspectorUtils { /** * Computes the bucket number to which the bucketFields belong to * @param bucketFields the bucketed fields of the row - * @param bucketFieldInspectors the ObjectInpsectors for each of the bucketed fields + * @param bucketFieldInspectors the ObjectInspectors for each of the bucketed fields * @param totalBuckets the number of buckets in the table * @return the bucket number */ diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java b/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java index 23c7a0d..c61b4c3 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java @@ -1418,7 +1418,7 @@ public final class PrimitiveObjectInspectorUtils { * Based on the PrimitiveCategory of a type, return the PrimitiveGrouping * that the PrimitiveCategory belongs to (numeric, string, date, etc). * @param primitiveCategory Primitive category of the type - * @return PrimitveGrouping corresponding to the PrimitiveCategory, + * @return PrimitiveGrouping corresponding to the PrimitiveCategory, * or UNKNOWN_GROUP if the type does not match to a grouping. */ public static PrimitiveGrouping getPrimitiveGrouping(PrimitiveCategory primitiveCategory) { diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataInputStream.java b/serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataInputStream.java index 9e19da7..a85b350 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataInputStream.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataInputStream.java @@ -136,7 +136,7 @@ public class TeradataBinaryDataInputStream extends SwappedDataInputStream { * Read CHAR(N). * The representation of char in Teradata binary format is * the byte number to read is based on the [charLength] * [bytePerChar] <- totalLength, - * bytePerChar is decided by the charset: LATAIN charset is 2 bytes per char and UNICODE charset is 3 bytes per char. + * bytePerChar is decided by the charset: LATIN charset is 2 bytes per char and UNICODE charset is 3 bytes per char. * the null char will use space to pad. * * @param totalLength the total length diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataOutputStream.java b/serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataOutputStream.java index 4cd2ac7..c2648c0 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataOutputStream.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataOutputStream.java @@ -165,7 +165,7 @@ public class TeradataBinaryDataOutputStream extends ByteArrayOutputStream { * Write CHAR(N). * The representation of char in Teradata binary format is: * the byte number to read is based on the [charLength] * [bytePerChar] <- totalLength, - * bytePerChar is decided by the charset: LATAIN charset is 2 bytes per char and UNICODE charset is 3 bytes per char. + * bytePerChar is decided by the charset: LATIN charset is 2 bytes per char and UNICODE charset is 3 bytes per char. * the null char will use space to pad. * * @param writable the writable diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinarySerde.java b/serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinarySerde.java index 480d029..543ae14 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinarySerde.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinarySerde.java @@ -412,8 +412,8 @@ import static java.lang.String.format; private Object deserializeField(TeradataBinaryDataInputStream in, TypeInfo type, Object reuse, boolean isNull) throws IOException, ParseException, SerDeException { // isNull: - // In the Teradata Binary file, even the field is null (isNull=true), - // thd data still has some default values to pad the record. + // In the Teradata Binary file, even if the field is null (isNull=true), + // the data still has some default values to pad the record. // In this case, you cannot avoid reading the bytes even it is not used. switch (type.getCategory()) { case PRIMITIVE: diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/thrift/TCTLSeparatedProtocol.java b/serde/src/java/org/apache/hadoop/hive/serde2/thrift/TCTLSeparatedProtocol.java index b04c3e7..d4590fc 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/thrift/TCTLSeparatedProtocol.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/thrift/TCTLSeparatedProtocol.java @@ -182,7 +182,7 @@ public class TCTLSeparatedProtocol extends TProtocol implements protected final TTransport innerTransport; /** - * Strings used to lookup the various configurable paramaters of this + * Strings used to lookup the various configurable parameters of this * protocol. */ public static final String ReturnNullsKey = "separators.return_nulls"; diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TimestampLocalTZTypeInfo.java b/serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TimestampLocalTZTypeInfo.java index a4f3af1..e1f9a26 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TimestampLocalTZTypeInfo.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TimestampLocalTZTypeInfo.java @@ -45,7 +45,7 @@ public class TimestampLocalTZTypeInfo extends PrimitiveTypeInfo { @Override public void setTypeName(String typeName) { - // No need to set type name, it should always be timestamplocaltz + // No need to set type name, it should always be {@link serdeConstants.TIMESTAMPLOCALTZ_TYPE_NAME} return; } diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoUtils.java b/serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoUtils.java index 24cd5d0..b23e3cf 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoUtils.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoUtils.java @@ -712,9 +712,9 @@ public final class TypeInfoUtils { break; } case STRUCT: { - StructTypeInfo strucTypeInfo = (StructTypeInfo) typeInfo; - List<String> fieldNames = strucTypeInfo.getAllStructFieldNames(); - List<TypeInfo> fieldTypeInfos = strucTypeInfo + StructTypeInfo structTypeInfo = (StructTypeInfo) typeInfo; + List<String> fieldNames = structTypeInfo.getAllStructFieldNames(); + List<TypeInfo> fieldTypeInfos = structTypeInfo .getAllStructFieldTypeInfos(); List<ObjectInspector> fieldObjectInspectors = new ArrayList<ObjectInspector>( fieldTypeInfos.size()); diff --git a/serde/src/test/org/apache/hadoop/hive/serde2/SerdeRandomRowSource.java b/serde/src/test/org/apache/hadoop/hive/serde2/SerdeRandomRowSource.java index e77083c..8ed0db8 100644 --- a/serde/src/test/org/apache/hadoop/hive/serde2/SerdeRandomRowSource.java +++ b/serde/src/test/org/apache/hadoop/hive/serde2/SerdeRandomRowSource.java @@ -457,7 +457,7 @@ public class SerdeRandomRowSource { primitiveObjectInspectorList.add(null); break; default: - throw new RuntimeException("Unexpected catagory " + category); + throw new RuntimeException("Unexpected category " + category); } objectInspectorList.add(objectInspector); diff --git a/serde/src/test/org/apache/hadoop/hive/serde2/TestJsonSerDe.java b/serde/src/test/org/apache/hadoop/hive/serde2/TestJsonSerDe.java index 1b24fd9..1135f9f 100644 --- a/serde/src/test/org/apache/hadoop/hive/serde2/TestJsonSerDe.java +++ b/serde/src/test/org/apache/hadoop/hive/serde2/TestJsonSerDe.java @@ -49,7 +49,7 @@ import org.junit.Test; public class TestJsonSerDe { @Test - public void testPrimativeDataTypes() throws Exception { + public void testPrimitiveDataTypes() throws Exception { Properties props = new Properties(); props.setProperty(serdeConstants.LIST_COLUMNS, "name,height,weight,endangered,born"); @@ -168,7 +168,7 @@ public class TestJsonSerDe { /** * Test when a map has a key defined as a numeric value. Technically, JSON * does not support this because each key in a map must be a quoted string. - * Unquoted strings (hence an int value) is allowed by Javascript, but not by + * Unquoted strings (hence an int value) is allowed by JavaScript, but not by * JSON specification. For Hive, the int map key type is stored as a string * and must be converted back into an int type. */ diff --git a/serde/src/test/org/apache/hadoop/hive/serde2/TestTCTLSeparatedProtocol.java b/serde/src/test/org/apache/hadoop/hive/serde2/TestTCTLSeparatedProtocol.java index d5d6566..9c87ecb 100644 --- a/serde/src/test/org/apache/hadoop/hive/serde2/TestTCTLSeparatedProtocol.java +++ b/serde/src/test/org/apache/hadoop/hive/serde2/TestTCTLSeparatedProtocol.java @@ -231,12 +231,12 @@ public class TestTCTLSeparatedProtocol { assertTrue(prot.readString().equals("bye!")); prot.readFieldEnd(); - // shouldl return nulls at end + // should return nulls at end prot.readFieldBegin(); assertNull(prot.readString()); prot.readFieldEnd(); - // shouldl return nulls at end + // should return nulls at end prot.readFieldBegin(); assertNull(prot.readString()); prot.readFieldEnd(); @@ -299,12 +299,12 @@ public class TestTCTLSeparatedProtocol { prot.readListEnd(); prot.readFieldEnd(); - // shouldl return nulls at end + // should return nulls at end prot.readFieldBegin(); assertNull(prot.readString()); prot.readFieldEnd(); - // shouldl return nulls at end + // should return nulls at end prot.readFieldBegin(); assertNull(prot.readString()); prot.readFieldEnd(); @@ -530,7 +530,7 @@ public class TestTCTLSeparatedProtocol { separatedProtocol.initialize(null, new Properties()); try { separatedProtocol.readStructBegin(); - fail("Runtime Exception is expected if the intialization of tokenizer failed."); + fail("Runtime Exception is expected if the initialization of tokenizer failed."); } catch (Exception e) { assertTrue(e.getCause() instanceof TTransportException); } diff --git a/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroDeserializer.java b/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroDeserializer.java index f9ac97a..24800da 100644 --- a/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroDeserializer.java +++ b/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroDeserializer.java @@ -510,7 +510,7 @@ public class TestAvroDeserializer { assertEquals("DALEKS", finalValue); } - @Test // Fixed doesn't exist in Hive. Fixeds go in, lists of bytes go out. + @Test // Fixed doesn't exist in Hive. Fixed go in, lists of bytes go out. public void canDeserializeFixed() throws SerDeException, IOException { Schema s = AvroSerdeUtils.getSchemaFor(TestAvroObjectInspectorGenerator.FIXED_SCHEMA); GenericData.Record record = new GenericData.Record(s); diff --git a/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroObjectInspectorGenerator.java b/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroObjectInspectorGenerator.java index 0029c3f..048d3d9 100644 --- a/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroObjectInspectorGenerator.java +++ b/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroObjectInspectorGenerator.java @@ -162,7 +162,7 @@ public class TestAvroObjectInspectorGenerator { " \"fields\": [\n" + " {\n" + " \"name\":\"baddies\",\n" + - " \"type\":{\"type\":\"enum\",\"name\":\"villians\", \"symbols\": " + + " \"type\":{\"type\":\"enum\",\"name\":\"villains\", \"symbols\": " + "[\"DALEKS\", \"CYBERMEN\", \"SLITHEEN\", \"JAGRAFESS\"]}\n" + " \n" + " \n" + @@ -206,7 +206,7 @@ public class TestAvroObjectInspectorGenerator { " \"fields\": [\n" + " {\n" + " \"name\":\"nullableEnum\",\n" + - " \"type\": [\"null\", {\"type\":\"enum\",\"name\":\"villians\", \"symbols\": " + + " \"type\": [\"null\", {\"type\":\"enum\",\"name\":\"villains\", \"symbols\": " + "[\"DALEKS\", \"CYBERMEN\", \"SLITHEEN\", \"JAGRAFESS\"]}]\n" + " \n" + " \n" + @@ -412,8 +412,8 @@ public class TestAvroObjectInspectorGenerator { /** * Check a given AvroObjectInspectorGenerator to verify that it matches our test * schema's expected map. - * @param aoig should already have been intitialized, may not be null - * @param fieldName name of the contianed column, will always fail if null. + * @param aoig should already have been initialized, may not be null + * @param fieldName name of the contained column, will always fail if null. */ private void verifyMap(final AvroObjectInspectorGenerator aoig, final String fieldName) { // Column names diff --git a/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroSerdeUtils.java b/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroSerdeUtils.java index 4e1ee69..3ef802d 100644 --- a/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroSerdeUtils.java +++ b/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroSerdeUtils.java @@ -48,7 +48,7 @@ public class TestAvroSerdeUtils { " {\"name\":\"mayBeNull\", \"type\":[\"string\", \"null\"]}\n" + " ]\n" + "}"; - // Same union, order reveresed + // Same union, order reversed private final String NULLABLE_UNION2 = "{\n" + " \"type\": \"record\", \n" + " \"name\": \"nullTest\",\n" + @@ -151,7 +151,7 @@ public class TestAvroSerdeUtils { } @Test - public void detemineSchemaTriesToOpenUrl() throws AvroSerdeException, IOException { + public void determineSchemaTriesToOpenUrl() throws AvroSerdeException, IOException { Configuration conf = new Configuration(); Properties props = new Properties(); props.put(AvroTableProperties.SCHEMA_URL.getPropName(), "not:///a.real.url"); diff --git a/serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/TestBinarySortableFast.java b/serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/TestBinarySortableFast.java index 68ba2c2..dbf84da 100644 --- a/serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/TestBinarySortableFast.java +++ b/serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/TestBinarySortableFast.java @@ -338,7 +338,7 @@ public class TestBinarySortableFast { } } if (!VerifyLazy.lazyCompare(typeInfo, complexFieldObj, expectedObject)) { - fail("Comparision failed typeInfo " + typeInfo.toString()); + fail("Comparison failed typeInfo " + typeInfo.toString()); } } } diff --git a/serde/src/test/org/apache/hadoop/hive/serde2/lazy/TestLazySimpleFast.java b/serde/src/test/org/apache/hadoop/hive/serde2/lazy/TestLazySimpleFast.java index 8187419..92b6176 100644 --- a/serde/src/test/org/apache/hadoop/hive/serde2/lazy/TestLazySimpleFast.java +++ b/serde/src/test/org/apache/hadoop/hive/serde2/lazy/TestLazySimpleFast.java @@ -244,7 +244,7 @@ public class TestLazySimpleFast { } } if (!VerifyLazy.lazyCompare(typeInfo, complexFieldObj, expectedObject)) { - fail("Comparision failed typeInfo " + typeInfo.toString()); + fail("Comparison failed typeInfo " + typeInfo.toString()); } } } diff --git a/serde/src/test/org/apache/hadoop/hive/serde2/lazybinary/TestLazyBinaryFast.java b/serde/src/test/org/apache/hadoop/hive/serde2/lazybinary/TestLazyBinaryFast.java index ae23809..fd1f23e 100644 --- a/serde/src/test/org/apache/hadoop/hive/serde2/lazybinary/TestLazyBinaryFast.java +++ b/serde/src/test/org/apache/hadoop/hive/serde2/lazybinary/TestLazyBinaryFast.java @@ -240,7 +240,7 @@ public class TestLazyBinaryFast { } } if (!VerifyLazy.lazyCompare(typeInfo, complexFieldObj, expectedObject)) { - fail("Comparision failed typeInfo " + typeInfo.toString()); + fail("Comparison failed typeInfo " + typeInfo.toString()); } } } diff --git a/serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestObjectInspectorConverters.java b/serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestObjectInspectorConverters.java index f829f94..08a4f2d 100644 --- a/serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestObjectInspectorConverters.java +++ b/serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestObjectInspectorConverters.java @@ -85,7 +85,7 @@ public class TestObjectInspectorConverters { convertText(); // Binary - converBinary(); + convertBinary(); // Union convertUnion(); @@ -185,7 +185,7 @@ private void convertUnion() { assertEquals(expectedObjectExtra, convertedObjectExtra); // we should get back null } -private void converBinary() { +private void convertBinary() { Converter baConverter = ObjectInspectorConverters.getConverter( PrimitiveObjectInspectorFactory.javaStringObjectInspector, PrimitiveObjectInspectorFactory.writableBinaryObjectInspector);