Build failed in Hudson: Hive-trunk-h0.17 #358
See http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/358/changes Changes: [zshao] HIVE-1140. Fix incorrect ambiguous column reference error message. (Paul Yang via zshao) -- Started by timer Building remotely on minerva.apache.org (Ubuntu) Updating http://svn.apache.org/repos/asf/hadoop/hive/trunk U CHANGES.txt A ql/src/test/results/clientnegative/ambiguous_col.q.out A ql/src/test/queries/clientnegative/ambiguous_col.q U ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java At revision 908483 [hive] $ /home/hudson/tools/ant/latest/bin/ant -Dhadoop.version=0.17.2.1 clean package javadoc test Buildfile: build.xml clean: clean: [echo] Cleaning: anttasks [delete] Deleting directory http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks clean: [echo] Cleaning: shims [delete] Deleting directory http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/shims clean: [echo] Cleaning: common clean: [echo] Cleaning: serde clean: [echo] Cleaning: metastore clean: [echo] Cleaning: ql clean: [echo] Cleaning: cli clean: [echo] Cleaning: contrib clean: clean: clean: [echo] Cleaning: hwi clean: [exec] rm -rf http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/odbc http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/service/objs http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/service/fb303/objs http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/metastore/objs clean-online: [delete] Deleting directory http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build clean-offline: jar: create-dirs: [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/shims [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/shims/classes [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/jexl/classes [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/hadoopcore [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/shims/test [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/shims/test/src [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/shims/test/classes compile-ant-tasks: create-dirs: [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/classes [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/test [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/test/src [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/test/classes init: compile: [echo] Compiling: anttasks [javac] Compiling 2 source files to http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/classes [javac] Note: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/ant/src/org/apache/hadoop/hive/ant/QTestGenTask.java uses or overrides a deprecated API. [javac] Note: Recompile with -Xlint:deprecation for details. deploy-ant-tasks: create-dirs: init: compile: [echo] Compiling: anttasks jar: [copy] Copying 1 file to http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/classes/org/apache/hadoop/hive/ant [jar] Building jar: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/hive-anttasks-0.6.0.jar init: compile: ivy-init-dirs: [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/ivy [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/ivy/lib [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/ivy/report [mkdir] Created dir: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/ivy/maven ivy-download: [get] Getting: http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar [get] To: http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/ivy/lib/ivy-2.1.0.jar ivy-probe-antlib: ivy-init-antlib: ivy-init: ivy-retrieve-hadoop-source: [ivy:retrieve] :: Ivy 2.0.0-rc2 - 20081028224207 :: http://ant.apache.org/ivy/ :: :: loading settings :: file =
Build failed in Hudson: Hive-trunk-h0.18 #362
See http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.18/362/changes Changes: [zshao] HIVE-1140. Fix incorrect ambiguous column reference error message. (Paul Yang via zshao) -- [...truncated 2972 lines...] A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPower.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqualOrGreaterThan.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericOp.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitNot.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNot.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPosMod.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNegative.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFMinute.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPDivide.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateDiff.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitXor.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqual.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFConcat.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSecond.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericUnaryOp.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFCeil.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPMod.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFType.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFToBoolean.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPLongDivide.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRegExpExtract.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFFromUnixTime.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateSub.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNotEqual.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFAsin.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFExp.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTF.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFSum.java AUql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIndex.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNull.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSize.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFAverage.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/UDTFCollector.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFConcatWS.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFHash.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStruct.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMin.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMax.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCount.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNotNull.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSplit.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStd.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBridge.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFBridge.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIf.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFUtils.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFEvaluator.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFLocate.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCase.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMap.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFArray.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCoalesce.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDF.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFField.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFExplode.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFElt.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/Collector.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStdSample.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInstr.java A
Build failed in Hudson: Hive-trunk-h0.19 #362
See http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.19/362/changes Changes: [zshao] HIVE-1140. Fix incorrect ambiguous column reference error message. (Paul Yang via zshao) -- [...truncated 2972 lines...] A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPower.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqualOrGreaterThan.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericOp.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitNot.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNot.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPosMod.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNegative.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFMinute.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPDivide.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateDiff.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitXor.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqual.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFConcat.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSecond.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericUnaryOp.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFCeil.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPMod.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFType.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFToBoolean.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPLongDivide.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRegExpExtract.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFFromUnixTime.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateSub.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNotEqual.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFAsin.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFExp.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTF.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFSum.java AUql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIndex.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNull.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSize.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFAverage.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/UDTFCollector.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFConcatWS.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFHash.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStruct.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMin.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMax.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCount.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNotNull.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSplit.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStd.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBridge.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFBridge.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIf.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFUtils.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFEvaluator.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFLocate.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCase.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMap.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFArray.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCoalesce.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDF.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFField.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFExplode.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFElt.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/Collector.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStdSample.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInstr.java A
Build latest hive failed
Hi all, Updated today. See below: Buildfile: D:\hadoop\src\hive\build.xml jar: create-dirs: compile-ant-tasks: create-dirs: init: compile: [echo] Compiling: anttasks [javac] Compiling 1 source file to D:\hadoop\src\hive\build\anttasks\classes [javac] 注意:D:\hadoop\src\hive\ant\src\org\apache\hadoop\hive\ant\QTestGenTask.java 使用或覆盖了已过时的 API。 [javac] 注意:要了解详细信息,请使用 -Xlint:deprecation 重新编译。 deploy-ant-tasks: create-dirs: init: compile: [echo] Compiling: anttasks jar: [jar] Building jar: D:\hadoop\src\hive\build\anttasks\hive-anttasks-0.6.0.jar init: compile: ivy-init-dirs: [mkdir] Created dir: D:\hadoop\src\hive\build\ivy [mkdir] Created dir: D:\hadoop\src\hive\build\ivy\lib [mkdir] Created dir: D:\hadoop\src\hive\build\ivy\report [mkdir] Created dir: D:\hadoop\src\hive\build\ivy\maven ivy-download: [get] Getting: http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar [get] To: D:\hadoop\src\hive\build\ivy\lib\ivy-2.1.0.jar ivy-probe-antlib: ivy-init-antlib: ivy-init: ivy-retrieve-hadoop-source: [ivy:retrieve] :: Ivy 2.0.0-rc2 - 20081028224207 :: http://ant.apache.org/ivy/ :: :: loading settings :: file = D:\hadoop\src\hive\ivy\ivysettings.xml BUILD FAILED D:\hadoop\src\hive\build.xml:148: The following error occurred while executing this line: D:\hadoop\src\hive\build.xml:93: The following error occurred while executing this line: D:\hadoop\src\hive\shims\build.xml:55: The following error occurred while executing this line: D:\hadoop\src\hive\build-common.xml:180: impossible to configure ivy:settings with given file: D:\hadoop\src\hive\ivy\ivysettings.xml : java.text.ParseException: failed to load settings from file:/D:/hadoop/src/hive/ivy/ivysettings.xml: impossible to set defaultTTL to eternal on class org.apache.ivy.core.cache.DefaultRepositoryCacheManager Total time: 12 seconds
Re: Build latest hive failed
Please see the message I sent to this list recently regarding cleaning out your .ant directory for the ivy upgrade. JVS On Feb 10, 2010, at 7:06 AM, jian yi eyj...@gmail.com wrote: Hi all, Updated today. See below: Buildfile: D:\hadoop\src\hive\build.xml jar: create-dirs: compile-ant-tasks: create-dirs: init: compile: [echo] Compiling: anttasks [javac] Compiling 1 source file to D:\hadoop\src\hive\build\anttasks\classes [javac] 注意:D:\hadoop\src\hive\ant\src\org\apache\hadoop\hive\ant \QTestGenTask.java 使用或覆盖了已过时的 API。 [javac] 注意:要了解详细信息,请使用 - Xlint:deprecation 重新编译。 deploy-ant-tasks: create-dirs: init: compile: [echo] Compiling: anttasks jar: [jar] Building jar: D:\hadoop\src\hive\build\anttasks\hive-anttasks-0.6.0.jar init: compile: ivy-init-dirs: [mkdir] Created dir: D:\hadoop\src\hive\build\ivy [mkdir] Created dir: D:\hadoop\src\hive\build\ivy\lib [mkdir] Created dir: D:\hadoop\src\hive\build\ivy\report [mkdir] Created dir: D:\hadoop\src\hive\build\ivy\maven ivy-download: [get] Getting: http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar [get] To: D:\hadoop\src\hive\build\ivy\lib\ivy-2.1.0.jar ivy-probe-antlib: ivy-init-antlib: ivy-init: ivy-retrieve-hadoop-source: [ivy:retrieve] :: Ivy 2.0.0-rc2 - 20081028224207 :: http://ant.apache.org/ivy/ :: :: loading settings :: file = D:\hadoop\src\hive\ivy\ivysettings.xml BUILD FAILED D:\hadoop\src\hive\build.xml:148: The following error occurred while executing this line: D:\hadoop\src\hive\build.xml:93: The following error occurred while executing this line: D:\hadoop\src\hive\shims\build.xml:55: The following error occurred while executing this line: D:\hadoop\src\hive\build-common.xml:180: impossible to configure ivy:settings with given file: D:\hadoop\src\hive\ivy\ivysettings.xml : java.text.ParseException: failed to load settings from file:/D:/hadoop/src/hive/ivy/ivysettings.xml: impossible to set defaultTTL to eternal on class org.apache.ivy.core.cache.DefaultRepositoryCacheManager Total time: 12 seconds
Build failed in Hudson: Hive-trunk-h0.20 #184
See http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.20/184/changes Changes: [zshao] HIVE-1140. Fix incorrect ambiguous column reference error message. (Paul Yang via zshao) -- [...truncated 2972 lines...] A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPower.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqualOrGreaterThan.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericOp.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitNot.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNot.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPosMod.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNegative.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFMinute.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPDivide.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateDiff.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitXor.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqual.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFConcat.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSecond.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericUnaryOp.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFCeil.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPMod.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFType.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFToBoolean.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPLongDivide.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRegExpExtract.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFFromUnixTime.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateSub.java AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNotEqual.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFAsin.java A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFExp.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTF.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFSum.java AUql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIndex.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNull.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSize.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFAverage.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/UDTFCollector.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFConcatWS.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFHash.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStruct.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMin.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMax.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCount.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNotNull.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSplit.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStd.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBridge.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFBridge.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIf.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFUtils.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFEvaluator.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFLocate.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCase.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMap.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFArray.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCoalesce.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDF.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFField.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFExplode.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFElt.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/Collector.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStdSample.java A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInstr.java A
[jira] Commented: (HIVE-1019) java.io.FileNotFoundException: HIVE_PLAN (No such file or directory)
[ https://issues.apache.org/jira/browse/HIVE-1019?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832118#action_12832118 ] Zheng Shao commented on HIVE-1019: -- There are some syntactic fixes needed: 1. Please avoid using import xxx.xxx.* 2. The code-style guideline from HIVE-1148 says we will have 100 chars per line, so we don't need to wrap the line if it's less than 100. 3. There are some unrelated changed about the CombineFileInputFormat java.io.FileNotFoundException: HIVE_PLAN (No such file or directory) Key: HIVE-1019 URL: https://issues.apache.org/jira/browse/HIVE-1019 Project: Hadoop Hive Issue Type: Bug Affects Versions: 0.6.0 Reporter: Bennie Schut Assignee: Bennie Schut Priority: Minor Fix For: 0.6.0 Attachments: HIVE-1019-1.patch, HIVE-1019-2.patch, HIVE-1019-3.patch, HIVE-1019-4.patch, HIVE-1019.patch, stacktrace2.txt I keep getting errors like this: java.io.FileNotFoundException: HIVE_PLAN (No such file or directory) and : java.io.IOException: cannot find dir = hdfs://victoria.ebuddy.com:9000/tmp/hive-dwh/801467596/10002 in partToPartitionInfo! when running multiple threads with roughly similar queries. I have a patch for this which works for me. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-1019) java.io.FileNotFoundException: HIVE_PLAN (No such file or directory)
[ https://issues.apache.org/jira/browse/HIVE-1019?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zheng Shao updated HIVE-1019: - Status: Open (was: Patch Available) java.io.FileNotFoundException: HIVE_PLAN (No such file or directory) Key: HIVE-1019 URL: https://issues.apache.org/jira/browse/HIVE-1019 Project: Hadoop Hive Issue Type: Bug Affects Versions: 0.6.0 Reporter: Bennie Schut Assignee: Bennie Schut Priority: Minor Fix For: 0.6.0 Attachments: HIVE-1019-1.patch, HIVE-1019-2.patch, HIVE-1019-3.patch, HIVE-1019-4.patch, HIVE-1019.patch, stacktrace2.txt I keep getting errors like this: java.io.FileNotFoundException: HIVE_PLAN (No such file or directory) and : java.io.IOException: cannot find dir = hdfs://victoria.ebuddy.com:9000/tmp/hive-dwh/801467596/10002 in partToPartitionInfo! when running multiple threads with roughly similar queries. I have a patch for this which works for me. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-471) A UDF for simple reflection
[ https://issues.apache.org/jira/browse/HIVE-471?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832120#action_12832120 ] Zheng Shao commented on HIVE-471: - Thanks Paul and Edward. Edward, please proceed with the annotations and commit when you get time. A UDF for simple reflection --- Key: HIVE-471 URL: https://issues.apache.org/jira/browse/HIVE-471 Project: Hadoop Hive Issue Type: New Feature Components: Query Processor Affects Versions: 0.6.0 Reporter: Edward Capriolo Assignee: Edward Capriolo Priority: Minor Attachments: hive-471-gen.diff, HIVE-471.1.patch, HIVE-471.2.patch, hive-471.diff There are many methods in java that are static and have no arguments or can be invoked with one simple parameter. More complicated functions will require a UDF but one generic one can work as a poor-mans UDF. {noformat} SELECT reflect(java.lang.String, valueOf, 1), reflect(java.lang.String, isEmpty) FROM src LIMIT 1; {noformat} -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-1149) Optimize CombineHiveFileInputFormat execution speed
Optimize CombineHiveFileInputFormat execution speed --- Key: HIVE-1149 URL: https://issues.apache.org/jira/browse/HIVE-1149 Project: Hadoop Hive Issue Type: Bug Reporter: Zheng Shao When there are a lot of files and a lot of pools, CombineHiveFileInputFormat is pretty slow. One of the culprit is the new URI call in the following function. We should try to get rid of it. {code} protected static PartitionDesc getPartitionDescFromPath( MapString, PartitionDesc pathToPartitionInfo, Path dir) throws IOException { // The format of the keys in pathToPartitionInfo sometimes contains a port // and sometimes doesn't, so we just compare paths. for (Map.EntryString, PartitionDesc entry : pathToPartitionInfo .entrySet()) { try { if (new URI(entry.getKey()).getPath().equals(dir.toUri().getPath())) { return entry.getValue(); } } catch (URISyntaxException e2) { } } throw new IOException(cannot find dir = + dir.toString() + in partToPartitionInfo!); } {code} -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-1149) Optimize CombineHiveFileInputFormat execution speed
[ https://issues.apache.org/jira/browse/HIVE-1149?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zheng Shao updated HIVE-1149: - Priority: Minor (was: Major) Issue Type: Improvement (was: Bug) Optimize CombineHiveFileInputFormat execution speed --- Key: HIVE-1149 URL: https://issues.apache.org/jira/browse/HIVE-1149 Project: Hadoop Hive Issue Type: Improvement Reporter: Zheng Shao Priority: Minor When there are a lot of files and a lot of pools, CombineHiveFileInputFormat is pretty slow. One of the culprit is the new URI call in the following function. We should try to get rid of it. {code} protected static PartitionDesc getPartitionDescFromPath( MapString, PartitionDesc pathToPartitionInfo, Path dir) throws IOException { // The format of the keys in pathToPartitionInfo sometimes contains a port // and sometimes doesn't, so we just compare paths. for (Map.EntryString, PartitionDesc entry : pathToPartitionInfo .entrySet()) { try { if (new URI(entry.getKey()).getPath().equals(dir.toUri().getPath())) { return entry.getValue(); } } catch (URISyntaxException e2) { } } throw new IOException(cannot find dir = + dir.toString() + in partToPartitionInfo!); } {code} -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-259) Add PERCENTILE aggregate function
[ https://issues.apache.org/jira/browse/HIVE-259?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832134#action_12832134 ] Zheng Shao commented on HIVE-259: - Jerome, it seems to me that the best data structure for counting is a HashMap, which allows near-constant-time insertion, find, and insertion. When we terminate we can get the entries and sort them but that cost should be small (it's one-time cost and the number of unique items won't be too big - users should have used round to shrink the number of unique numbers). It seems currently we are paying log(n) cost for each find, and O(n) cost for each insertion. Does that make sense? For sharing the state object, we can just declare the state class as public static. Add PERCENTILE aggregate function - Key: HIVE-259 URL: https://issues.apache.org/jira/browse/HIVE-259 Project: Hadoop Hive Issue Type: New Feature Components: Query Processor Reporter: Venky Iyer Assignee: Jerome Boulon Attachments: HIVE-259.patch Compute atleast 25, 50, 75th percentiles -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-259) Add PERCENTILE aggregate function
[ https://issues.apache.org/jira/browse/HIVE-259?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832139#action_12832139 ] Todd Lipcon commented on HIVE-259: -- Agreed re HashMap. Also, there should be some kind of setting that limits how much RAM gets used up. In a later iteration we could do adaptive histogramming once we hit the limit. In this version we should just throw up our hands and fail with a message that says the user needs to discretize harder. Add PERCENTILE aggregate function - Key: HIVE-259 URL: https://issues.apache.org/jira/browse/HIVE-259 Project: Hadoop Hive Issue Type: New Feature Components: Query Processor Reporter: Venky Iyer Assignee: Jerome Boulon Attachments: HIVE-259.patch Compute atleast 25, 50, 75th percentiles -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-259) Add PERCENTILE aggregate function
[ https://issues.apache.org/jira/browse/HIVE-259?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832146#action_12832146 ] Jerome Boulon commented on HIVE-259: Didn't know that we can use an Hash on the state Object ... Is there any limitation on what can be used on the state object or can we use any java Object? Also how is the state serialized between Map and Reduce? Add PERCENTILE aggregate function - Key: HIVE-259 URL: https://issues.apache.org/jira/browse/HIVE-259 Project: Hadoop Hive Issue Type: New Feature Components: Query Processor Reporter: Venky Iyer Assignee: Jerome Boulon Attachments: HIVE-259.patch Compute atleast 25, 50, 75th percentiles -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-1150) Add comment to explain why we check for dir first in add_partitions().
Add comment to explain why we check for dir first in add_partitions(). -- Key: HIVE-1150 URL: https://issues.apache.org/jira/browse/HIVE-1150 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Reporter: Paul Yang Assignee: Paul Yang Priority: Trivial Fix For: 0.6.0 -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-1150) Add comment to explain why we check for dir first in add_partitions().
[ https://issues.apache.org/jira/browse/HIVE-1150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Paul Yang updated HIVE-1150: Attachment: HIVE-1150.1.patch Add comment to explain why we check for dir first in add_partitions(). -- Key: HIVE-1150 URL: https://issues.apache.org/jira/browse/HIVE-1150 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Reporter: Paul Yang Assignee: Paul Yang Priority: Trivial Fix For: 0.6.0 Attachments: HIVE-1150.1.patch -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-1150) Add comment to explain why we check for dir first in add_partitions().
[ https://issues.apache.org/jira/browse/HIVE-1150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Paul Yang updated HIVE-1150: Status: Patch Available (was: Open) Add comment to explain why we check for dir first in add_partitions(). -- Key: HIVE-1150 URL: https://issues.apache.org/jira/browse/HIVE-1150 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Reporter: Paul Yang Assignee: Paul Yang Priority: Trivial Fix For: 0.6.0 Attachments: HIVE-1150.1.patch -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-1083) allow sub-directories for an external table/partition
[ https://issues.apache.org/jira/browse/HIVE-1083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zheng Shao updated HIVE-1083: - Description: Sometimes users want to define an external table/partition based on all files (recursively) inside a directory. Currently most of the Hadoop InputFormat classes do not support that. We should extract all files recursively in the directory, and add them to the input path of the job. was:Subdirectories should be allowed for tables/partitions. Affects Version/s: 0.6.0 Assignee: Zheng Shao Summary: allow sub-directories for an external table/partition (was: allow sub-directories for a table/partition) allow sub-directories for an external table/partition - Key: HIVE-1083 URL: https://issues.apache.org/jira/browse/HIVE-1083 Project: Hadoop Hive Issue Type: Improvement Components: Query Processor Affects Versions: 0.6.0 Reporter: Namit Jain Assignee: Zheng Shao Fix For: 0.6.0 Sometimes users want to define an external table/partition based on all files (recursively) inside a directory. Currently most of the Hadoop InputFormat classes do not support that. We should extract all files recursively in the directory, and add them to the input path of the job. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-1083) allow sub-directories for an external table/partition
[ https://issues.apache.org/jira/browse/HIVE-1083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zheng Shao updated HIVE-1083: - Labels: inputformat (was: ) Corner cases: C1. We have 4 external tables: abc_recursive, abc, abc_def_recursive, abc_def abc_recursive and abc both points to /abc abc_def and abc_def_recursive both points to /abc/def abc_recursive and abc_def_recursive have set the bit recursive. In ExecDriver, given all tables, we need to find all paths that needs to be added to the input path. In MapOperator, given the current input path, we need to find all the aliases that the current input path corresponds to. allow sub-directories for an external table/partition - Key: HIVE-1083 URL: https://issues.apache.org/jira/browse/HIVE-1083 Project: Hadoop Hive Issue Type: Improvement Components: Query Processor Affects Versions: 0.6.0 Reporter: Namit Jain Assignee: Zheng Shao Fix For: 0.6.0 Sometimes users want to define an external table/partition based on all files (recursively) inside a directory. Currently most of the Hadoop InputFormat classes do not support that. We should extract all files recursively in the directory, and add them to the input path of the job. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-1150) Add comment to explain why we check for dir first in add_partitions().
[ https://issues.apache.org/jira/browse/HIVE-1150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zheng Shao updated HIVE-1150: - Resolution: Fixed Release Note: HIVE-1150. Add comment to explain why we check for dir first in add_partitions(). (Paul Yang via zshao) Hadoop Flags: [Reviewed] Status: Resolved (was: Patch Available) Committed. Thanks Paul! Add comment to explain why we check for dir first in add_partitions(). -- Key: HIVE-1150 URL: https://issues.apache.org/jira/browse/HIVE-1150 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Reporter: Paul Yang Assignee: Paul Yang Priority: Trivial Fix For: 0.6.0 Attachments: HIVE-1150.1.patch -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-1151) Add 'show version' command to Hive CLI
Add 'show version' command to Hive CLI -- Key: HIVE-1151 URL: https://issues.apache.org/jira/browse/HIVE-1151 Project: Hadoop Hive Issue Type: New Feature Components: Clients Affects Versions: 0.6.0 Reporter: Carl Steinbach Assignee: Carl Steinbach At a minimum this command should return the version information obtained from the hive-cli jar. Ideally this command will also return version information obtained from each of the hive jar files present in the CLASSPATH, which will allow us to quickly detect cases where people are using incompatible jars. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-1132) Add metastore API method to get partition by name
[ https://issues.apache.org/jira/browse/HIVE-1132?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Paul Yang updated HIVE-1132: Component/s: Metastore Add metastore API method to get partition by name - Key: HIVE-1132 URL: https://issues.apache.org/jira/browse/HIVE-1132 Project: Hadoop Hive Issue Type: New Feature Components: Metastore Affects Versions: 0.6.0 Reporter: Paul Yang Assignee: Paul Yang Fix For: 0.6.0 Attachments: HIVE-1132.1.patch Currently, get_partition_names returns the partition names in an escaped form ie 'ds=2010-02-03/ts=2010-02-03 18%3A49%3A26/offset=0-3184760670135/instance=nfs/host=nfs'. In this case, the colons have been replaced by %3A. The escaped form is necessary because the partition column values could contain symbols such as '=' or '/' that would interfere with parsing or have some other unwanted effects. See HIVE-883. However, there is no way to directly retrieve the partition using the escaped name because get_partition accepts a ListString that requires the partition column values to be in their original unescaped form. So the proposal is to add get_partition_by_name() that directly accepts the partition name in the escaped form. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-1152) Add metastore API method to drop partition / append partition by name
Add metastore API method to drop partition / append partition by name - Key: HIVE-1152 URL: https://issues.apache.org/jira/browse/HIVE-1152 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Affects Versions: 0.6.0 Reporter: Paul Yang Assignee: Paul Yang Similar to get_partition_by_name, but for drop_partition and append_partition. See related issue. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-1156) Add metastore API method to drop partition / append partition by name
Add metastore API method to drop partition / append partition by name - Key: HIVE-1156 URL: https://issues.apache.org/jira/browse/HIVE-1156 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Affects Versions: 0.6.0 Reporter: Paul Yang Assignee: Paul Yang Similar to get_partition_by_name, but for drop_partition and append_partition. See related issue. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-1154) Add metastore API method to drop partition / append partition by name
Add metastore API method to drop partition / append partition by name - Key: HIVE-1154 URL: https://issues.apache.org/jira/browse/HIVE-1154 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Affects Versions: 0.6.0 Reporter: Paul Yang Assignee: Paul Yang Similar to get_partition_by_name, but for drop_partition and append_partition. See related issue. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-1155) Add metastore API method to drop partition / append partition by name
Add metastore API method to drop partition / append partition by name - Key: HIVE-1155 URL: https://issues.apache.org/jira/browse/HIVE-1155 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Affects Versions: 0.6.0 Reporter: Paul Yang Assignee: Paul Yang Similar to get_partition_by_name, but for drop_partition and append_partition. See related issue. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-1153) Add metastore API method to drop partition / append partition by name
Add metastore API method to drop partition / append partition by name - Key: HIVE-1153 URL: https://issues.apache.org/jira/browse/HIVE-1153 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Affects Versions: 0.6.0 Reporter: Paul Yang Assignee: Paul Yang Similar to get_partition_by_name, but for drop_partition and append_partition. See related issue. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Resolved: (HIVE-1154) Add metastore API method to drop partition / append partition by name
[ https://issues.apache.org/jira/browse/HIVE-1154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Paul Yang resolved HIVE-1154. - Resolution: Duplicate Add metastore API method to drop partition / append partition by name - Key: HIVE-1154 URL: https://issues.apache.org/jira/browse/HIVE-1154 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Affects Versions: 0.6.0 Reporter: Paul Yang Assignee: Paul Yang Similar to get_partition_by_name, but for drop_partition and append_partition. See related issue. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Resolved: (HIVE-1153) Add metastore API method to drop partition / append partition by name
[ https://issues.apache.org/jira/browse/HIVE-1153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Paul Yang resolved HIVE-1153. - Resolution: Duplicate Add metastore API method to drop partition / append partition by name - Key: HIVE-1153 URL: https://issues.apache.org/jira/browse/HIVE-1153 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Affects Versions: 0.6.0 Reporter: Paul Yang Assignee: Paul Yang Similar to get_partition_by_name, but for drop_partition and append_partition. See related issue. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Resolved: (HIVE-1155) Add metastore API method to drop partition / append partition by name
[ https://issues.apache.org/jira/browse/HIVE-1155?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Paul Yang resolved HIVE-1155. - Resolution: Duplicate Add metastore API method to drop partition / append partition by name - Key: HIVE-1155 URL: https://issues.apache.org/jira/browse/HIVE-1155 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Affects Versions: 0.6.0 Reporter: Paul Yang Assignee: Paul Yang Similar to get_partition_by_name, but for drop_partition and append_partition. See related issue. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Resolved: (HIVE-1156) Add metastore API method to drop partition / append partition by name
[ https://issues.apache.org/jira/browse/HIVE-1156?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Paul Yang resolved HIVE-1156. - Resolution: Duplicate Add metastore API method to drop partition / append partition by name - Key: HIVE-1156 URL: https://issues.apache.org/jira/browse/HIVE-1156 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Affects Versions: 0.6.0 Reporter: Paul Yang Assignee: Paul Yang Similar to get_partition_by_name, but for drop_partition and append_partition. See related issue. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-259) Add PERCENTILE aggregate function
[ https://issues.apache.org/jira/browse/HIVE-259?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zheng Shao updated HIVE-259: Attachment: HIVE-259.1.patch Jerome, I did a skeleton of the code to use HashMap. Do you want to start from there and add what is missing? Add PERCENTILE aggregate function - Key: HIVE-259 URL: https://issues.apache.org/jira/browse/HIVE-259 Project: Hadoop Hive Issue Type: New Feature Components: Query Processor Reporter: Venky Iyer Assignee: Jerome Boulon Attachments: HIVE-259.1.patch, HIVE-259.patch Compute atleast 25, 50, 75th percentiles -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-1122) Make ql/metadata/Table and Partition serializable
[ https://issues.apache.org/jira/browse/HIVE-1122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832347#action_12832347 ] Paul Yang commented on HIVE-1122: - Looks good +1 Make ql/metadata/Table and Partition serializable - Key: HIVE-1122 URL: https://issues.apache.org/jira/browse/HIVE-1122 Project: Hadoop Hive Issue Type: Improvement Affects Versions: 0.6.0 Reporter: Zheng Shao Assignee: Zheng Shao Attachments: HIVE-1122.4.patch, HIVE-1122.5.patch, HIVE-1122.pre.1.patch Both Table and Partition are just wrappers of the thrift class TTable and TPartition. However, Table and Partition are not serializable, and there are no guarantee in the code to make sure Table is in sync with the field of TTable. We should make Table and Partition serializable by only serializing the TTable and TPartition fields, and we should hide the TTable details inside the Table class. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-1122) Make ql/metadata/Table and Partition serializable
[ https://issues.apache.org/jira/browse/HIVE-1122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832349#action_12832349 ] He Yongqiang commented on HIVE-1122: Will test and commit Make ql/metadata/Table and Partition serializable - Key: HIVE-1122 URL: https://issues.apache.org/jira/browse/HIVE-1122 Project: Hadoop Hive Issue Type: Improvement Affects Versions: 0.6.0 Reporter: Zheng Shao Assignee: Zheng Shao Attachments: HIVE-1122.4.patch, HIVE-1122.5.patch, HIVE-1122.pre.1.patch Both Table and Partition are just wrappers of the thrift class TTable and TPartition. However, Table and Partition are not serializable, and there are no guarantee in the code to make sure Table is in sync with the field of TTable. We should make Table and Partition serializable by only serializing the TTable and TPartition fields, and we should hide the TTable details inside the Table class. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-917) Bucketed Map Join
[ https://issues.apache.org/jira/browse/HIVE-917?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832359#action_12832359 ] Namit Jain commented on HIVE-917: - Can you generate the patch again - it is not applying cleanly Bucketed Map Join - Key: HIVE-917 URL: https://issues.apache.org/jira/browse/HIVE-917 Project: Hadoop Hive Issue Type: New Feature Reporter: Zheng Shao Assignee: He Yongqiang Attachments: hive-917-2010-2-3.patch, hive-917-2010-2-8.patch Hive already have support for map-join. Map-join treats the big table as job input, and in each mapper, it loads all data from a small table. In case the big table is already bucketed on the join key, we don't have to load the whole small table in each of the mappers. This will greatly alleviate the memory pressure, and make map-join work with medium-sized tables. There are 4 steps we can improve: S0. This is what the user can already do now: create a new bucketed table and insert all data from the small table to it; Submit BUCKETNUM jobs, each doing a map-side join of bigtable TABLEPARTITION(BUCKET i OUT OF NBUCKETS) with smallbucketedtable TABLEPARTITION(BUCKET i OUT OF NBUCKETS). S1. Change the code so that when map-join is loading the small table, we automatically drop the rows with the keys that are NOT in the same bucket as the big table. This should alleviate the problem on memory, but we might still have thousands of mappers reading the whole of the small table. S2. Let's say the user already bucketed the small table on the join key into exactly the same number of buckets (or a factor of the buckets of the big table), then map-join can choose to load only the buckets that are useful. S3. Add a new hint (e.g. /*+ MAPBUCKETJOIN(a) */), so that Hive automatically does S2, without the need of asking the user to create temporary bucketed table for the small table. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-1157) UDFs can't be loaded via add jar when jar is on HDFS
UDFs can't be loaded via add jar when jar is on HDFS -- Key: HIVE-1157 URL: https://issues.apache.org/jira/browse/HIVE-1157 Project: Hadoop Hive Issue Type: Improvement Components: Query Processor Reporter: Philip Zeyliger Priority: Minor As discussed on the mailing list, it would be nice if you could use UDFs that are on jars on HDFS. The proposed implementation would be for add jar to recognize that the target file is on HDFS, copy it locally, and load it into the classpath. {quote} Hi folks, I have a quick question about UDF support in Hive. I'm on the 0.5 branch. Can you use a UDF where the jar which contains the function is on HDFS, and not on the local filesystem. Specifically, the following does not seem to work: # This is Hive 0.5, from svn $bin/hive Hive history file=/tmp/philip/hive_job_log_philip_201002081541_370227273.txt hive add jar hdfs://localhost/FooTest.jar; Added hdfs://localhost/FooTest.jar to class path hive create temporary function cube as 'com.cloudera.FooTestUDF'; FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.FunctionTask Does this work for other people? I could probably fix it by changing add jar to download remote jars locally, when necessary (to load them into the classpath), or update URLClassLoader (or whatever is underneath there) to read directly from HDFS, which seems a bit more fragile. But I wanted to make sure that my interpretation of what's going on is right before I have at it. Thanks, -- Philip {quote} {quote} Yes that's correct. I prefer to download the jars in add jar. Zheng {quote} -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-1122) Make ql/metadata/Table and Partition serializable
[ https://issues.apache.org/jira/browse/HIVE-1122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] He Yongqiang updated HIVE-1122: --- Resolution: Fixed Release Note: HIVE-1122. Make ql/metadata/Table and Partition serializable Status: Resolved (was: Patch Available) Committed. Thanks Zheng! Make ql/metadata/Table and Partition serializable - Key: HIVE-1122 URL: https://issues.apache.org/jira/browse/HIVE-1122 Project: Hadoop Hive Issue Type: Improvement Affects Versions: 0.6.0 Reporter: Zheng Shao Assignee: Zheng Shao Attachments: HIVE-1122.4.patch, HIVE-1122.5.patch, HIVE-1122.pre.1.patch Both Table and Partition are just wrappers of the thrift class TTable and TPartition. However, Table and Partition are not serializable, and there are no guarantee in the code to make sure Table is in sync with the field of TTable. We should make Table and Partition serializable by only serializing the TTable and TPartition fields, and we should hide the TTable details inside the Table class. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-1157) UDFs can't be loaded via add jar when jar is on HDFS
[ https://issues.apache.org/jira/browse/HIVE-1157?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832381#action_12832381 ] Edward Capriolo commented on HIVE-1157: --- Removing local file dependencies is much cleaner. UDFs can't be loaded via add jar when jar is on HDFS -- Key: HIVE-1157 URL: https://issues.apache.org/jira/browse/HIVE-1157 Project: Hadoop Hive Issue Type: Improvement Components: Query Processor Reporter: Philip Zeyliger Priority: Minor As discussed on the mailing list, it would be nice if you could use UDFs that are on jars on HDFS. The proposed implementation would be for add jar to recognize that the target file is on HDFS, copy it locally, and load it into the classpath. {quote} Hi folks, I have a quick question about UDF support in Hive. I'm on the 0.5 branch. Can you use a UDF where the jar which contains the function is on HDFS, and not on the local filesystem. Specifically, the following does not seem to work: # This is Hive 0.5, from svn $bin/hive Hive history file=/tmp/philip/hive_job_log_philip_201002081541_370227273.txt hive add jar hdfs://localhost/FooTest.jar; Added hdfs://localhost/FooTest.jar to class path hive create temporary function cube as 'com.cloudera.FooTestUDF'; FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.FunctionTask Does this work for other people? I could probably fix it by changing add jar to download remote jars locally, when necessary (to load them into the classpath), or update URLClassLoader (or whatever is underneath there) to read directly from HDFS, which seems a bit more fragile. But I wanted to make sure that my interpretation of what's going on is right before I have at it. Thanks, -- Philip {quote} {quote} Yes that's correct. I prefer to download the jars in add jar. Zheng {quote} -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-1152) Add metastore API method to drop partition / append partition by name
[ https://issues.apache.org/jira/browse/HIVE-1152?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Paul Yang updated HIVE-1152: Status: Patch Available (was: Open) Add metastore API method to drop partition / append partition by name - Key: HIVE-1152 URL: https://issues.apache.org/jira/browse/HIVE-1152 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Affects Versions: 0.6.0 Reporter: Paul Yang Assignee: Paul Yang Attachments: HIVE-1152.1.patch Similar to get_partition_by_name, but for drop_partition and append_partition. See related issue. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-1152) Add metastore API method to drop partition / append partition by name
[ https://issues.apache.org/jira/browse/HIVE-1152?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Paul Yang updated HIVE-1152: Attachment: HIVE-1152.1.patch Add metastore API method to drop partition / append partition by name - Key: HIVE-1152 URL: https://issues.apache.org/jira/browse/HIVE-1152 Project: Hadoop Hive Issue Type: Improvement Components: Metastore Affects Versions: 0.6.0 Reporter: Paul Yang Assignee: Paul Yang Attachments: HIVE-1152.1.patch Similar to get_partition_by_name, but for drop_partition and append_partition. See related issue. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-1157) UDFs can't be loaded via add jar when jar is on HDFS
[ https://issues.apache.org/jira/browse/HIVE-1157?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832389#action_12832389 ] Philip Zeyliger commented on HIVE-1157: --- Edward, I'm not sure what you mean. -- Philip UDFs can't be loaded via add jar when jar is on HDFS -- Key: HIVE-1157 URL: https://issues.apache.org/jira/browse/HIVE-1157 Project: Hadoop Hive Issue Type: Improvement Components: Query Processor Reporter: Philip Zeyliger Priority: Minor As discussed on the mailing list, it would be nice if you could use UDFs that are on jars on HDFS. The proposed implementation would be for add jar to recognize that the target file is on HDFS, copy it locally, and load it into the classpath. {quote} Hi folks, I have a quick question about UDF support in Hive. I'm on the 0.5 branch. Can you use a UDF where the jar which contains the function is on HDFS, and not on the local filesystem. Specifically, the following does not seem to work: # This is Hive 0.5, from svn $bin/hive Hive history file=/tmp/philip/hive_job_log_philip_201002081541_370227273.txt hive add jar hdfs://localhost/FooTest.jar; Added hdfs://localhost/FooTest.jar to class path hive create temporary function cube as 'com.cloudera.FooTestUDF'; FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.FunctionTask Does this work for other people? I could probably fix it by changing add jar to download remote jars locally, when necessary (to load them into the classpath), or update URLClassLoader (or whatever is underneath there) to read directly from HDFS, which seems a bit more fragile. But I wanted to make sure that my interpretation of what's going on is right before I have at it. Thanks, -- Philip {quote} {quote} Yes that's correct. I prefer to download the jars in add jar. Zheng {quote} -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.