Build failed in Hudson: Hive-trunk-h0.17 #358

2010-02-10 Thread Apache Hudson Server
See http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/358/changes

Changes:

[zshao] HIVE-1140. Fix incorrect ambiguous column reference error message. 
(Paul Yang via zshao)

--
Started by timer
Building remotely on minerva.apache.org (Ubuntu)
Updating http://svn.apache.org/repos/asf/hadoop/hive/trunk
U CHANGES.txt
A ql/src/test/results/clientnegative/ambiguous_col.q.out
A ql/src/test/queries/clientnegative/ambiguous_col.q
U ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java
At revision 908483
[hive] $ /home/hudson/tools/ant/latest/bin/ant -Dhadoop.version=0.17.2.1 clean 
package javadoc test
Buildfile: build.xml

clean:

clean:
 [echo] Cleaning: anttasks
   [delete] Deleting directory 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks

clean:
 [echo] Cleaning: shims
   [delete] Deleting directory 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/shims

clean:
 [echo] Cleaning: common

clean:
 [echo] Cleaning: serde

clean:
 [echo] Cleaning: metastore

clean:
 [echo] Cleaning: ql

clean:
 [echo] Cleaning: cli

clean:
 [echo] Cleaning: contrib

clean:

clean:

clean:
 [echo] Cleaning: hwi

clean:
 [exec] rm -rf 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/odbc 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/service/objs
 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/service/fb303/objs
 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/metastore/objs

clean-online:
   [delete] Deleting directory 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build

clean-offline:

jar:

create-dirs:
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/shims
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/shims/classes
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/jexl/classes
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/hadoopcore
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/shims/test
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/shims/test/src
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/shims/test/classes

compile-ant-tasks:

create-dirs:
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/classes
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/test
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/test/src
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/test/classes

init:

compile:
 [echo] Compiling: anttasks
[javac] Compiling 2 source files to 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/classes
[javac] Note: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/ant/src/org/apache/hadoop/hive/ant/QTestGenTask.java
 uses or overrides a deprecated API.
[javac] Note: Recompile with -Xlint:deprecation for details.

deploy-ant-tasks:

create-dirs:

init:

compile:
 [echo] Compiling: anttasks

jar:
 [copy] Copying 1 file to 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/classes/org/apache/hadoop/hive/ant
  [jar] Building jar: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/anttasks/hive-anttasks-0.6.0.jar

init:

compile:

ivy-init-dirs:
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/ivy
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/ivy/lib
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/ivy/report
[mkdir] Created dir: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/ivy/maven

ivy-download:
  [get] Getting: 
http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar
  [get] To: 
http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/ws/hive/build/ivy/lib/ivy-2.1.0.jar

ivy-probe-antlib:

ivy-init-antlib:

ivy-init:

ivy-retrieve-hadoop-source:
[ivy:retrieve] :: Ivy 2.0.0-rc2 - 20081028224207 :: http://ant.apache.org/ivy/ 
::
:: loading settings :: file = 

Build failed in Hudson: Hive-trunk-h0.18 #362

2010-02-10 Thread Apache Hudson Server
See http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.18/362/changes

Changes:

[zshao] HIVE-1140. Fix incorrect ambiguous column reference error message. 
(Paul Yang via zshao)

--
[...truncated 2972 lines...]
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPower.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqualOrGreaterThan.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericOp.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitNot.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNot.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPosMod.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNegative.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFMinute.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPDivide.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateDiff.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitXor.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqual.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFConcat.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSecond.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericUnaryOp.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFCeil.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPMod.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFType.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFToBoolean.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPLongDivide.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRegExpExtract.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFFromUnixTime.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateSub.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNotEqual.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFAsin.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFExp.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTF.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFSum.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIndex.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNull.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSize.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFAverage.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/UDTFCollector.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFConcatWS.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFHash.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStruct.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMin.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMax.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCount.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNotNull.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSplit.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStd.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBridge.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFBridge.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIf.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFUtils.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFEvaluator.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFLocate.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCase.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMap.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFArray.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCoalesce.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDF.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFField.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFExplode.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFElt.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/Collector.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStdSample.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInstr.java
A 

Build failed in Hudson: Hive-trunk-h0.19 #362

2010-02-10 Thread Apache Hudson Server
See http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.19/362/changes

Changes:

[zshao] HIVE-1140. Fix incorrect ambiguous column reference error message. 
(Paul Yang via zshao)

--
[...truncated 2972 lines...]
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPower.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqualOrGreaterThan.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericOp.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitNot.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNot.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPosMod.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNegative.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFMinute.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPDivide.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateDiff.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitXor.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqual.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFConcat.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSecond.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericUnaryOp.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFCeil.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPMod.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFType.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFToBoolean.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPLongDivide.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRegExpExtract.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFFromUnixTime.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateSub.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNotEqual.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFAsin.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFExp.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTF.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFSum.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIndex.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNull.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSize.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFAverage.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/UDTFCollector.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFConcatWS.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFHash.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStruct.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMin.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMax.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCount.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNotNull.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSplit.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStd.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBridge.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFBridge.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIf.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFUtils.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFEvaluator.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFLocate.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCase.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMap.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFArray.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCoalesce.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDF.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFField.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFExplode.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFElt.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/Collector.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStdSample.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInstr.java
A 

Build latest hive failed

2010-02-10 Thread jian yi
Hi all,

Updated today. See below:



Buildfile: D:\hadoop\src\hive\build.xml

jar:

create-dirs:

compile-ant-tasks:

create-dirs:

init:

compile:
[echo] Compiling: anttasks
   [javac] Compiling 1 source file to
D:\hadoop\src\hive\build\anttasks\classes
   [javac]
注意:D:\hadoop\src\hive\ant\src\org\apache\hadoop\hive\ant\QTestGenTask.java
使用或覆盖了已过时的 API。
   [javac] 注意:要了解详细信息,请使用 -Xlint:deprecation 重新编译。

deploy-ant-tasks:

create-dirs:

init:

compile:
[echo] Compiling: anttasks

jar:
 [jar] Building jar:
D:\hadoop\src\hive\build\anttasks\hive-anttasks-0.6.0.jar

init:

compile:

ivy-init-dirs:
   [mkdir] Created dir: D:\hadoop\src\hive\build\ivy
   [mkdir] Created dir: D:\hadoop\src\hive\build\ivy\lib
   [mkdir] Created dir: D:\hadoop\src\hive\build\ivy\report
   [mkdir] Created dir: D:\hadoop\src\hive\build\ivy\maven

ivy-download:
 [get] Getting:
http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar
 [get] To: D:\hadoop\src\hive\build\ivy\lib\ivy-2.1.0.jar

ivy-probe-antlib:

ivy-init-antlib:

ivy-init:

ivy-retrieve-hadoop-source:
[ivy:retrieve] :: Ivy 2.0.0-rc2 - 20081028224207 ::
http://ant.apache.org/ivy/ ::
:: loading settings :: file = D:\hadoop\src\hive\ivy\ivysettings.xml

BUILD FAILED
D:\hadoop\src\hive\build.xml:148: The following error occurred while
executing this line:
D:\hadoop\src\hive\build.xml:93: The following error occurred while
executing this line:
D:\hadoop\src\hive\shims\build.xml:55: The following error occurred while
executing this line:
D:\hadoop\src\hive\build-common.xml:180: impossible to configure
ivy:settings with given file: D:\hadoop\src\hive\ivy\ivysettings.xml :
java.text.ParseException: failed to load settings from
file:/D:/hadoop/src/hive/ivy/ivysettings.xml: impossible to set defaultTTL
to eternal on class org.apache.ivy.core.cache.DefaultRepositoryCacheManager

Total time: 12 seconds


Re: Build latest hive failed

2010-02-10 Thread John Sichi
Please see the message I sent to this list recently regarding cleaning  
out your .ant directory for the ivy upgrade.

JVS

On Feb 10, 2010, at 7:06 AM, jian yi eyj...@gmail.com wrote:

 Hi all,

 Updated today. See below:



 Buildfile: D:\hadoop\src\hive\build.xml

 jar:

 create-dirs:

 compile-ant-tasks:

 create-dirs:

 init:

 compile:
[echo] Compiling: anttasks
   [javac] Compiling 1 source file to
 D:\hadoop\src\hive\build\anttasks\classes
   [javac]
 注意:D:\hadoop\src\hive\ant\src\org\apache\hadoop\hive\ant 
 \QTestGenTask.java
 使用或覆盖了已过时的 API。
   [javac] 注意:要了解详细信息,请使用 - 
 Xlint:deprecation 重新编译。

 deploy-ant-tasks:

 create-dirs:

 init:

 compile:
[echo] Compiling: anttasks

 jar:
 [jar] Building jar:
 D:\hadoop\src\hive\build\anttasks\hive-anttasks-0.6.0.jar

 init:

 compile:

 ivy-init-dirs:
   [mkdir] Created dir: D:\hadoop\src\hive\build\ivy
   [mkdir] Created dir: D:\hadoop\src\hive\build\ivy\lib
   [mkdir] Created dir: D:\hadoop\src\hive\build\ivy\report
   [mkdir] Created dir: D:\hadoop\src\hive\build\ivy\maven

 ivy-download:
 [get] Getting:
 http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar
 [get] To: D:\hadoop\src\hive\build\ivy\lib\ivy-2.1.0.jar

 ivy-probe-antlib:

 ivy-init-antlib:

 ivy-init:

 ivy-retrieve-hadoop-source:
 [ivy:retrieve] :: Ivy 2.0.0-rc2 - 20081028224207 ::
 http://ant.apache.org/ivy/ ::
 :: loading settings :: file = D:\hadoop\src\hive\ivy\ivysettings.xml

 BUILD FAILED
 D:\hadoop\src\hive\build.xml:148: The following error occurred while
 executing this line:
 D:\hadoop\src\hive\build.xml:93: The following error occurred while
 executing this line:
 D:\hadoop\src\hive\shims\build.xml:55: The following error occurred  
 while
 executing this line:
 D:\hadoop\src\hive\build-common.xml:180: impossible to configure
 ivy:settings with given file: D:\hadoop\src\hive\ivy\ivysettings.xml :
 java.text.ParseException: failed to load settings from
 file:/D:/hadoop/src/hive/ivy/ivysettings.xml: impossible to set  
 defaultTTL
 to eternal on class  
 org.apache.ivy.core.cache.DefaultRepositoryCacheManager

 Total time: 12 seconds


Build failed in Hudson: Hive-trunk-h0.20 #184

2010-02-10 Thread Apache Hudson Server
See http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.20/184/changes

Changes:

[zshao] HIVE-1140. Fix incorrect ambiguous column reference error message. 
(Paul Yang via zshao)

--
[...truncated 2972 lines...]
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPower.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqualOrGreaterThan.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericOp.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitNot.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNot.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPosMod.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNegative.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFMinute.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPDivide.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateDiff.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitXor.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqual.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFConcat.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSecond.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericUnaryOp.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFCeil.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPMod.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFType.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFToBoolean.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPLongDivide.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRegExpExtract.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFFromUnixTime.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateSub.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNotEqual.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFAsin.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/UDFExp.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTF.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFSum.java
AUql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIndex.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNull.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSize.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFAverage.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/UDTFCollector.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFConcatWS.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFHash.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStruct.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMin.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMax.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCount.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNotNull.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSplit.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStd.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBridge.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFBridge.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIf.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFUtils.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFEvaluator.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFLocate.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCase.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMap.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFArray.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCoalesce.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDF.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFField.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFExplode.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFElt.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/Collector.java
A 
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStdSample.java
A ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInstr.java
A 

[jira] Commented: (HIVE-1019) java.io.FileNotFoundException: HIVE_PLAN (No such file or directory)

2010-02-10 Thread Zheng Shao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1019?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832118#action_12832118
 ] 

Zheng Shao commented on HIVE-1019:
--

There are some syntactic fixes needed:
1. Please avoid using import xxx.xxx.*
2. The code-style guideline from HIVE-1148 says we will have 100 chars per 
line, so we don't need to wrap the line if it's less than 100.
3. There are some unrelated changed about the CombineFileInputFormat


 java.io.FileNotFoundException: HIVE_PLAN (No such file or directory)
 

 Key: HIVE-1019
 URL: https://issues.apache.org/jira/browse/HIVE-1019
 Project: Hadoop Hive
  Issue Type: Bug
Affects Versions: 0.6.0
Reporter: Bennie Schut
Assignee: Bennie Schut
Priority: Minor
 Fix For: 0.6.0

 Attachments: HIVE-1019-1.patch, HIVE-1019-2.patch, HIVE-1019-3.patch, 
 HIVE-1019-4.patch, HIVE-1019.patch, stacktrace2.txt


 I keep getting errors like this:
 java.io.FileNotFoundException: HIVE_PLAN (No such file or directory)
 and :
 java.io.IOException: cannot find dir = 
 hdfs://victoria.ebuddy.com:9000/tmp/hive-dwh/801467596/10002 in 
 partToPartitionInfo!
 when running multiple threads with roughly similar queries.
 I have a patch for this which works for me.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1019) java.io.FileNotFoundException: HIVE_PLAN (No such file or directory)

2010-02-10 Thread Zheng Shao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1019?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zheng Shao updated HIVE-1019:
-

Status: Open  (was: Patch Available)

 java.io.FileNotFoundException: HIVE_PLAN (No such file or directory)
 

 Key: HIVE-1019
 URL: https://issues.apache.org/jira/browse/HIVE-1019
 Project: Hadoop Hive
  Issue Type: Bug
Affects Versions: 0.6.0
Reporter: Bennie Schut
Assignee: Bennie Schut
Priority: Minor
 Fix For: 0.6.0

 Attachments: HIVE-1019-1.patch, HIVE-1019-2.patch, HIVE-1019-3.patch, 
 HIVE-1019-4.patch, HIVE-1019.patch, stacktrace2.txt


 I keep getting errors like this:
 java.io.FileNotFoundException: HIVE_PLAN (No such file or directory)
 and :
 java.io.IOException: cannot find dir = 
 hdfs://victoria.ebuddy.com:9000/tmp/hive-dwh/801467596/10002 in 
 partToPartitionInfo!
 when running multiple threads with roughly similar queries.
 I have a patch for this which works for me.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-471) A UDF for simple reflection

2010-02-10 Thread Zheng Shao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-471?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832120#action_12832120
 ] 

Zheng Shao commented on HIVE-471:
-

Thanks Paul and Edward.
Edward, please proceed with the annotations and commit when you get time.


 A UDF for simple reflection
 ---

 Key: HIVE-471
 URL: https://issues.apache.org/jira/browse/HIVE-471
 Project: Hadoop Hive
  Issue Type: New Feature
  Components: Query Processor
Affects Versions: 0.6.0
Reporter: Edward Capriolo
Assignee: Edward Capriolo
Priority: Minor
 Attachments: hive-471-gen.diff, HIVE-471.1.patch, HIVE-471.2.patch, 
 hive-471.diff


 There are many methods in java that are static and have no arguments or can 
 be invoked with one simple parameter. More complicated functions will require 
 a UDF but one generic one can work as a poor-mans UDF.
 {noformat}
 SELECT reflect(java.lang.String, valueOf, 1), reflect(java.lang.String, 
 isEmpty)
 FROM src LIMIT 1;
 {noformat}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1149) Optimize CombineHiveFileInputFormat execution speed

2010-02-10 Thread Zheng Shao (JIRA)
Optimize CombineHiveFileInputFormat execution speed
---

 Key: HIVE-1149
 URL: https://issues.apache.org/jira/browse/HIVE-1149
 Project: Hadoop Hive
  Issue Type: Bug
Reporter: Zheng Shao


When there are a lot of files and a lot of pools, CombineHiveFileInputFormat is 
pretty slow.
One of the culprit is the new URI call in the following function. We should 
try to get rid of it.

{code}
  protected static PartitionDesc getPartitionDescFromPath(
  MapString, PartitionDesc pathToPartitionInfo, Path dir) throws 
IOException {
// The format of the keys in pathToPartitionInfo sometimes contains a port
// and sometimes doesn't, so we just compare paths.
for (Map.EntryString, PartitionDesc entry : pathToPartitionInfo
.entrySet()) {
  try {
if (new URI(entry.getKey()).getPath().equals(dir.toUri().getPath())) {
  return entry.getValue();
}
  } catch (URISyntaxException e2) {
  }
}
throw new IOException(cannot find dir =  + dir.toString()
+  in partToPartitionInfo!);
  }
{code}


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1149) Optimize CombineHiveFileInputFormat execution speed

2010-02-10 Thread Zheng Shao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1149?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zheng Shao updated HIVE-1149:
-

  Priority: Minor  (was: Major)
Issue Type: Improvement  (was: Bug)

 Optimize CombineHiveFileInputFormat execution speed
 ---

 Key: HIVE-1149
 URL: https://issues.apache.org/jira/browse/HIVE-1149
 Project: Hadoop Hive
  Issue Type: Improvement
Reporter: Zheng Shao
Priority: Minor

 When there are a lot of files and a lot of pools, CombineHiveFileInputFormat 
 is pretty slow.
 One of the culprit is the new URI call in the following function. We should 
 try to get rid of it.
 {code}
   protected static PartitionDesc getPartitionDescFromPath(
   MapString, PartitionDesc pathToPartitionInfo, Path dir) throws 
 IOException {
 // The format of the keys in pathToPartitionInfo sometimes contains a port
 // and sometimes doesn't, so we just compare paths.
 for (Map.EntryString, PartitionDesc entry : pathToPartitionInfo
 .entrySet()) {
   try {
 if (new URI(entry.getKey()).getPath().equals(dir.toUri().getPath())) {
   return entry.getValue();
 }
   } catch (URISyntaxException e2) {
   }
 }
 throw new IOException(cannot find dir =  + dir.toString()
 +  in partToPartitionInfo!);
   }
 {code}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-259) Add PERCENTILE aggregate function

2010-02-10 Thread Zheng Shao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-259?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832134#action_12832134
 ] 

Zheng Shao commented on HIVE-259:
-

Jerome, it seems to me that the best data structure for counting is a HashMap, 
which allows near-constant-time insertion, find, and insertion. When we 
terminate we can get the entries and sort them but that cost should be small 
(it's one-time cost and the number of unique items won't be too big - users 
should have used round to shrink the number of unique numbers).

It seems currently we are paying log(n) cost for each find, and O(n) cost for 
each insertion.

Does that make sense?

For sharing the state object, we can just declare the state class as public 
static.


 Add PERCENTILE aggregate function
 -

 Key: HIVE-259
 URL: https://issues.apache.org/jira/browse/HIVE-259
 Project: Hadoop Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Venky Iyer
Assignee: Jerome Boulon
 Attachments: HIVE-259.patch


 Compute atleast 25, 50, 75th percentiles

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-259) Add PERCENTILE aggregate function

2010-02-10 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-259?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832139#action_12832139
 ] 

Todd Lipcon commented on HIVE-259:
--

Agreed re HashMap. Also, there should be some kind of setting that limits how 
much RAM gets used up. In a later iteration we could do adaptive histogramming 
once we hit the limit. In this version we should just throw up our hands and 
fail with a message that says the user needs to discretize harder.

 Add PERCENTILE aggregate function
 -

 Key: HIVE-259
 URL: https://issues.apache.org/jira/browse/HIVE-259
 Project: Hadoop Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Venky Iyer
Assignee: Jerome Boulon
 Attachments: HIVE-259.patch


 Compute atleast 25, 50, 75th percentiles

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-259) Add PERCENTILE aggregate function

2010-02-10 Thread Jerome Boulon (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-259?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832146#action_12832146
 ] 

Jerome Boulon commented on HIVE-259:


Didn't know that we can use an Hash on the state Object ...
Is there any limitation on what can be used on the state object or can we use 
any java Object? 
Also how is the state serialized between Map and Reduce?

 Add PERCENTILE aggregate function
 -

 Key: HIVE-259
 URL: https://issues.apache.org/jira/browse/HIVE-259
 Project: Hadoop Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Venky Iyer
Assignee: Jerome Boulon
 Attachments: HIVE-259.patch


 Compute atleast 25, 50, 75th percentiles

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1150) Add comment to explain why we check for dir first in add_partitions().

2010-02-10 Thread Paul Yang (JIRA)
Add comment to explain why we check for dir first in add_partitions().
--

 Key: HIVE-1150
 URL: https://issues.apache.org/jira/browse/HIVE-1150
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Reporter: Paul Yang
Assignee: Paul Yang
Priority: Trivial
 Fix For: 0.6.0




-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1150) Add comment to explain why we check for dir first in add_partitions().

2010-02-10 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang updated HIVE-1150:


Attachment: HIVE-1150.1.patch

 Add comment to explain why we check for dir first in add_partitions().
 --

 Key: HIVE-1150
 URL: https://issues.apache.org/jira/browse/HIVE-1150
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Reporter: Paul Yang
Assignee: Paul Yang
Priority: Trivial
 Fix For: 0.6.0

 Attachments: HIVE-1150.1.patch




-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1150) Add comment to explain why we check for dir first in add_partitions().

2010-02-10 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang updated HIVE-1150:


Status: Patch Available  (was: Open)

 Add comment to explain why we check for dir first in add_partitions().
 --

 Key: HIVE-1150
 URL: https://issues.apache.org/jira/browse/HIVE-1150
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Reporter: Paul Yang
Assignee: Paul Yang
Priority: Trivial
 Fix For: 0.6.0

 Attachments: HIVE-1150.1.patch




-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1083) allow sub-directories for an external table/partition

2010-02-10 Thread Zheng Shao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zheng Shao updated HIVE-1083:
-

  Description: 
Sometimes users want to define an external table/partition based on all files 
(recursively) inside a directory.

Currently most of the Hadoop InputFormat classes do not support that. We should 
extract all files recursively in the directory, and add them to the input path 
of the job.



  was:Subdirectories should be allowed for tables/partitions.

Affects Version/s: 0.6.0
 Assignee: Zheng Shao
  Summary: allow sub-directories for an external table/partition  
(was: allow sub-directories for a table/partition)

 allow sub-directories for an external table/partition
 -

 Key: HIVE-1083
 URL: https://issues.apache.org/jira/browse/HIVE-1083
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Query Processor
Affects Versions: 0.6.0
Reporter: Namit Jain
Assignee: Zheng Shao
 Fix For: 0.6.0


 Sometimes users want to define an external table/partition based on all files 
 (recursively) inside a directory.
 Currently most of the Hadoop InputFormat classes do not support that. We 
 should extract all files recursively in the directory, and add them to the 
 input path of the job.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1083) allow sub-directories for an external table/partition

2010-02-10 Thread Zheng Shao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zheng Shao updated HIVE-1083:
-

Labels: inputformat  (was: )

Corner cases:
C1. We have 4 external tables: abc_recursive, abc, abc_def_recursive, abc_def
abc_recursive and abc both points to /abc
abc_def and abc_def_recursive both points to /abc/def
abc_recursive and abc_def_recursive have set the bit recursive.

In ExecDriver, given all tables, we need to find all paths that needs to be 
added to the input path.
In MapOperator, given the current input path, we need to find all the aliases 
that the current input path corresponds to.


 allow sub-directories for an external table/partition
 -

 Key: HIVE-1083
 URL: https://issues.apache.org/jira/browse/HIVE-1083
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Query Processor
Affects Versions: 0.6.0
Reporter: Namit Jain
Assignee: Zheng Shao
 Fix For: 0.6.0


 Sometimes users want to define an external table/partition based on all files 
 (recursively) inside a directory.
 Currently most of the Hadoop InputFormat classes do not support that. We 
 should extract all files recursively in the directory, and add them to the 
 input path of the job.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1150) Add comment to explain why we check for dir first in add_partitions().

2010-02-10 Thread Zheng Shao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zheng Shao updated HIVE-1150:
-

  Resolution: Fixed
Release Note: HIVE-1150. Add comment to explain why we check for dir first 
in add_partitions(). (Paul Yang via zshao)
Hadoop Flags: [Reviewed]
  Status: Resolved  (was: Patch Available)

Committed. Thanks Paul!

 Add comment to explain why we check for dir first in add_partitions().
 --

 Key: HIVE-1150
 URL: https://issues.apache.org/jira/browse/HIVE-1150
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Reporter: Paul Yang
Assignee: Paul Yang
Priority: Trivial
 Fix For: 0.6.0

 Attachments: HIVE-1150.1.patch




-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1151) Add 'show version' command to Hive CLI

2010-02-10 Thread Carl Steinbach (JIRA)
Add 'show version' command to Hive CLI
--

 Key: HIVE-1151
 URL: https://issues.apache.org/jira/browse/HIVE-1151
 Project: Hadoop Hive
  Issue Type: New Feature
  Components: Clients
Affects Versions: 0.6.0
Reporter: Carl Steinbach
Assignee: Carl Steinbach


At a minimum this command should return the version information obtained
from the hive-cli jar. Ideally this command will also return version information
obtained from each of the hive jar files present in the CLASSPATH, which
will allow us to quickly detect cases where people are using incompatible
jars.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1132) Add metastore API method to get partition by name

2010-02-10 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1132?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang updated HIVE-1132:


Component/s: Metastore

 Add metastore API method to get partition by name
 -

 Key: HIVE-1132
 URL: https://issues.apache.org/jira/browse/HIVE-1132
 Project: Hadoop Hive
  Issue Type: New Feature
  Components: Metastore
Affects Versions: 0.6.0
Reporter: Paul Yang
Assignee: Paul Yang
 Fix For: 0.6.0

 Attachments: HIVE-1132.1.patch


 Currently, get_partition_names returns the partition names in an escaped form 
  ie 'ds=2010-02-03/ts=2010-02-03 
 18%3A49%3A26/offset=0-3184760670135/instance=nfs/host=nfs'. In this case, the 
 colons have been replaced by %3A. The escaped form is necessary because the 
 partition column values could contain symbols such as '=' or '/' that would 
 interfere with parsing or have some other unwanted effects. See HIVE-883.
 However, there is no way to directly retrieve the partition using the escaped 
 name because get_partition accepts a ListString that requires the partition 
 column values to be in their original unescaped form. So the proposal is to 
 add get_partition_by_name() that directly accepts the partition name in the 
 escaped form.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1152) Add metastore API method to drop partition / append partition by name

2010-02-10 Thread Paul Yang (JIRA)
Add metastore API method to drop partition / append partition by name
-

 Key: HIVE-1152
 URL: https://issues.apache.org/jira/browse/HIVE-1152
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Affects Versions: 0.6.0
Reporter: Paul Yang
Assignee: Paul Yang


Similar to get_partition_by_name, but for drop_partition and append_partition. 
See related issue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1156) Add metastore API method to drop partition / append partition by name

2010-02-10 Thread Paul Yang (JIRA)
Add metastore API method to drop partition / append partition by name
-

 Key: HIVE-1156
 URL: https://issues.apache.org/jira/browse/HIVE-1156
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Affects Versions: 0.6.0
Reporter: Paul Yang
Assignee: Paul Yang


Similar to get_partition_by_name, but for drop_partition and append_partition. 
See related issue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1154) Add metastore API method to drop partition / append partition by name

2010-02-10 Thread Paul Yang (JIRA)
Add metastore API method to drop partition / append partition by name
-

 Key: HIVE-1154
 URL: https://issues.apache.org/jira/browse/HIVE-1154
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Affects Versions: 0.6.0
Reporter: Paul Yang
Assignee: Paul Yang


Similar to get_partition_by_name, but for drop_partition and append_partition. 
See related issue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1155) Add metastore API method to drop partition / append partition by name

2010-02-10 Thread Paul Yang (JIRA)
Add metastore API method to drop partition / append partition by name
-

 Key: HIVE-1155
 URL: https://issues.apache.org/jira/browse/HIVE-1155
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Affects Versions: 0.6.0
Reporter: Paul Yang
Assignee: Paul Yang


Similar to get_partition_by_name, but for drop_partition and append_partition. 
See related issue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1153) Add metastore API method to drop partition / append partition by name

2010-02-10 Thread Paul Yang (JIRA)
Add metastore API method to drop partition / append partition by name
-

 Key: HIVE-1153
 URL: https://issues.apache.org/jira/browse/HIVE-1153
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Affects Versions: 0.6.0
Reporter: Paul Yang
Assignee: Paul Yang


Similar to get_partition_by_name, but for drop_partition and append_partition. 
See related issue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HIVE-1154) Add metastore API method to drop partition / append partition by name

2010-02-10 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang resolved HIVE-1154.
-

Resolution: Duplicate

 Add metastore API method to drop partition / append partition by name
 -

 Key: HIVE-1154
 URL: https://issues.apache.org/jira/browse/HIVE-1154
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Affects Versions: 0.6.0
Reporter: Paul Yang
Assignee: Paul Yang

 Similar to get_partition_by_name, but for drop_partition and 
 append_partition. See related issue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HIVE-1153) Add metastore API method to drop partition / append partition by name

2010-02-10 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang resolved HIVE-1153.
-

Resolution: Duplicate

 Add metastore API method to drop partition / append partition by name
 -

 Key: HIVE-1153
 URL: https://issues.apache.org/jira/browse/HIVE-1153
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Affects Versions: 0.6.0
Reporter: Paul Yang
Assignee: Paul Yang

 Similar to get_partition_by_name, but for drop_partition and 
 append_partition. See related issue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HIVE-1155) Add metastore API method to drop partition / append partition by name

2010-02-10 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1155?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang resolved HIVE-1155.
-

Resolution: Duplicate

 Add metastore API method to drop partition / append partition by name
 -

 Key: HIVE-1155
 URL: https://issues.apache.org/jira/browse/HIVE-1155
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Affects Versions: 0.6.0
Reporter: Paul Yang
Assignee: Paul Yang

 Similar to get_partition_by_name, but for drop_partition and 
 append_partition. See related issue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HIVE-1156) Add metastore API method to drop partition / append partition by name

2010-02-10 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1156?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang resolved HIVE-1156.
-

Resolution: Duplicate

 Add metastore API method to drop partition / append partition by name
 -

 Key: HIVE-1156
 URL: https://issues.apache.org/jira/browse/HIVE-1156
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Affects Versions: 0.6.0
Reporter: Paul Yang
Assignee: Paul Yang

 Similar to get_partition_by_name, but for drop_partition and 
 append_partition. See related issue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-259) Add PERCENTILE aggregate function

2010-02-10 Thread Zheng Shao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-259?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zheng Shao updated HIVE-259:


Attachment: HIVE-259.1.patch

Jerome, I did a skeleton of the code to use HashMap. Do you want to start from 
there and add what is missing?


 Add PERCENTILE aggregate function
 -

 Key: HIVE-259
 URL: https://issues.apache.org/jira/browse/HIVE-259
 Project: Hadoop Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Venky Iyer
Assignee: Jerome Boulon
 Attachments: HIVE-259.1.patch, HIVE-259.patch


 Compute atleast 25, 50, 75th percentiles

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1122) Make ql/metadata/Table and Partition serializable

2010-02-10 Thread Paul Yang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832347#action_12832347
 ] 

Paul Yang commented on HIVE-1122:
-

Looks good +1

 Make ql/metadata/Table and Partition serializable
 -

 Key: HIVE-1122
 URL: https://issues.apache.org/jira/browse/HIVE-1122
 Project: Hadoop Hive
  Issue Type: Improvement
Affects Versions: 0.6.0
Reporter: Zheng Shao
Assignee: Zheng Shao
 Attachments: HIVE-1122.4.patch, HIVE-1122.5.patch, 
 HIVE-1122.pre.1.patch


 Both Table and Partition are just wrappers of the thrift class TTable and 
 TPartition.
 However, Table and Partition are not serializable, and there are no guarantee 
 in the code to make sure Table is in sync with the field of TTable.
 We should make Table and Partition serializable by only serializing the 
 TTable and TPartition fields, and we should hide the TTable details inside 
 the Table class.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1122) Make ql/metadata/Table and Partition serializable

2010-02-10 Thread He Yongqiang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832349#action_12832349
 ] 

He Yongqiang commented on HIVE-1122:


Will test and commit

 Make ql/metadata/Table and Partition serializable
 -

 Key: HIVE-1122
 URL: https://issues.apache.org/jira/browse/HIVE-1122
 Project: Hadoop Hive
  Issue Type: Improvement
Affects Versions: 0.6.0
Reporter: Zheng Shao
Assignee: Zheng Shao
 Attachments: HIVE-1122.4.patch, HIVE-1122.5.patch, 
 HIVE-1122.pre.1.patch


 Both Table and Partition are just wrappers of the thrift class TTable and 
 TPartition.
 However, Table and Partition are not serializable, and there are no guarantee 
 in the code to make sure Table is in sync with the field of TTable.
 We should make Table and Partition serializable by only serializing the 
 TTable and TPartition fields, and we should hide the TTable details inside 
 the Table class.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-917) Bucketed Map Join

2010-02-10 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-917?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832359#action_12832359
 ] 

Namit Jain commented on HIVE-917:
-

Can you generate the patch again - it is not applying cleanly

 Bucketed Map Join
 -

 Key: HIVE-917
 URL: https://issues.apache.org/jira/browse/HIVE-917
 Project: Hadoop Hive
  Issue Type: New Feature
Reporter: Zheng Shao
Assignee: He Yongqiang
 Attachments: hive-917-2010-2-3.patch, hive-917-2010-2-8.patch


 Hive already have support for map-join. Map-join treats the big table as job 
 input, and in each mapper, it loads all data from a small table.
 In case the big table is already bucketed on the join key, we don't have to 
 load the whole small table in each of the mappers. This will greatly 
 alleviate the memory pressure, and make map-join work with medium-sized 
 tables.
 There are 4 steps we can improve:
 S0. This is what the user can already do now: create a new bucketed table and 
 insert all data from the small table to it; Submit BUCKETNUM jobs, each doing 
 a map-side join of bigtable TABLEPARTITION(BUCKET i OUT OF NBUCKETS) with 
 smallbucketedtable TABLEPARTITION(BUCKET i OUT OF NBUCKETS).
 S1. Change the code so that when map-join is loading the small table, we 
 automatically drop the rows with the keys that are NOT in the same bucket as 
 the big table. This should alleviate the problem on memory, but we might 
 still have thousands of mappers reading the whole of the small table.
 S2. Let's say the user already bucketed the small table on the join key into 
 exactly the same number of buckets (or a factor of the buckets of the big 
 table), then map-join can choose to load only the buckets that are useful.
 S3. Add a new hint (e.g. /*+ MAPBUCKETJOIN(a) */), so that Hive automatically 
 does S2, without the need of asking the user to create temporary bucketed 
 table for the small table.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1157) UDFs can't be loaded via add jar when jar is on HDFS

2010-02-10 Thread Philip Zeyliger (JIRA)
UDFs can't be loaded via add jar when jar is on HDFS
--

 Key: HIVE-1157
 URL: https://issues.apache.org/jira/browse/HIVE-1157
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Philip Zeyliger
Priority: Minor


As discussed on the mailing list, it would be nice if you could use UDFs that 
are on jars on HDFS.  The proposed implementation would be for add jar to 
recognize that the target file is on HDFS, copy it locally, and load it into 
the classpath.

{quote}
Hi folks,

I have a quick question about UDF support in Hive.  I'm on the 0.5 branch.  Can 
you use a UDF where the jar which contains the function is on HDFS, and not on 
the local filesystem.  Specifically, the following does not seem to work:

# This is Hive 0.5, from svn
$bin/hive  
Hive history file=/tmp/philip/hive_job_log_philip_201002081541_370227273.txt
hive add jar hdfs://localhost/FooTest.jar; 
 
Added hdfs://localhost/FooTest.jar to class path
hive create temporary function cube as 'com.cloudera.FooTestUDF';  
  
FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.FunctionTask

Does this work for other people?  I could probably fix it by changing add jar 
to download remote jars locally, when necessary (to load them into the 
classpath), or update URLClassLoader (or whatever is underneath there) to read 
directly from HDFS, which seems a bit more fragile.  But I wanted to make sure 
that my interpretation of what's going on is right before I have at it.

Thanks,

-- Philip
{quote}

{quote}
Yes that's correct. I prefer to download the jars in add jar.

Zheng
{quote}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1122) Make ql/metadata/Table and Partition serializable

2010-02-10 Thread He Yongqiang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

He Yongqiang updated HIVE-1122:
---

  Resolution: Fixed
Release Note: HIVE-1122. Make ql/metadata/Table and Partition serializable
  Status: Resolved  (was: Patch Available)

Committed. Thanks Zheng!

 Make ql/metadata/Table and Partition serializable
 -

 Key: HIVE-1122
 URL: https://issues.apache.org/jira/browse/HIVE-1122
 Project: Hadoop Hive
  Issue Type: Improvement
Affects Versions: 0.6.0
Reporter: Zheng Shao
Assignee: Zheng Shao
 Attachments: HIVE-1122.4.patch, HIVE-1122.5.patch, 
 HIVE-1122.pre.1.patch


 Both Table and Partition are just wrappers of the thrift class TTable and 
 TPartition.
 However, Table and Partition are not serializable, and there are no guarantee 
 in the code to make sure Table is in sync with the field of TTable.
 We should make Table and Partition serializable by only serializing the 
 TTable and TPartition fields, and we should hide the TTable details inside 
 the Table class.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1157) UDFs can't be loaded via add jar when jar is on HDFS

2010-02-10 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1157?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832381#action_12832381
 ] 

Edward Capriolo commented on HIVE-1157:
---

Removing local file dependencies is much cleaner.

 UDFs can't be loaded via add jar when jar is on HDFS
 --

 Key: HIVE-1157
 URL: https://issues.apache.org/jira/browse/HIVE-1157
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Philip Zeyliger
Priority: Minor

 As discussed on the mailing list, it would be nice if you could use UDFs that 
 are on jars on HDFS.  The proposed implementation would be for add jar to 
 recognize that the target file is on HDFS, copy it locally, and load it into 
 the classpath.
 {quote}
 Hi folks,
 I have a quick question about UDF support in Hive.  I'm on the 0.5 branch.  
 Can you use a UDF where the jar which contains the function is on HDFS, and 
 not on the local filesystem.  Specifically, the following does not seem to 
 work:
 # This is Hive 0.5, from svn
 $bin/hive  
 Hive history file=/tmp/philip/hive_job_log_philip_201002081541_370227273.txt
 hive add jar hdfs://localhost/FooTest.jar;   

 Added hdfs://localhost/FooTest.jar to class path
 hive create temporary function cube as 'com.cloudera.FooTestUDF';
 
 FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.FunctionTask
 Does this work for other people?  I could probably fix it by changing add 
 jar to download remote jars locally, when necessary (to load them into the 
 classpath), or update URLClassLoader (or whatever is underneath there) to 
 read directly from HDFS, which seems a bit more fragile.  But I wanted to 
 make sure that my interpretation of what's going on is right before I have at 
 it.
 Thanks,
 -- Philip
 {quote}
 {quote}
 Yes that's correct. I prefer to download the jars in add jar.
 Zheng
 {quote}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1152) Add metastore API method to drop partition / append partition by name

2010-02-10 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1152?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang updated HIVE-1152:


Status: Patch Available  (was: Open)

 Add metastore API method to drop partition / append partition by name
 -

 Key: HIVE-1152
 URL: https://issues.apache.org/jira/browse/HIVE-1152
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Affects Versions: 0.6.0
Reporter: Paul Yang
Assignee: Paul Yang
 Attachments: HIVE-1152.1.patch


 Similar to get_partition_by_name, but for drop_partition and 
 append_partition. See related issue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1152) Add metastore API method to drop partition / append partition by name

2010-02-10 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1152?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang updated HIVE-1152:


Attachment: HIVE-1152.1.patch

 Add metastore API method to drop partition / append partition by name
 -

 Key: HIVE-1152
 URL: https://issues.apache.org/jira/browse/HIVE-1152
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Metastore
Affects Versions: 0.6.0
Reporter: Paul Yang
Assignee: Paul Yang
 Attachments: HIVE-1152.1.patch


 Similar to get_partition_by_name, but for drop_partition and 
 append_partition. See related issue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1157) UDFs can't be loaded via add jar when jar is on HDFS

2010-02-10 Thread Philip Zeyliger (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1157?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12832389#action_12832389
 ] 

Philip Zeyliger commented on HIVE-1157:
---

Edward,

I'm not sure what you mean.

-- Philip

 UDFs can't be loaded via add jar when jar is on HDFS
 --

 Key: HIVE-1157
 URL: https://issues.apache.org/jira/browse/HIVE-1157
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Philip Zeyliger
Priority: Minor

 As discussed on the mailing list, it would be nice if you could use UDFs that 
 are on jars on HDFS.  The proposed implementation would be for add jar to 
 recognize that the target file is on HDFS, copy it locally, and load it into 
 the classpath.
 {quote}
 Hi folks,
 I have a quick question about UDF support in Hive.  I'm on the 0.5 branch.  
 Can you use a UDF where the jar which contains the function is on HDFS, and 
 not on the local filesystem.  Specifically, the following does not seem to 
 work:
 # This is Hive 0.5, from svn
 $bin/hive  
 Hive history file=/tmp/philip/hive_job_log_philip_201002081541_370227273.txt
 hive add jar hdfs://localhost/FooTest.jar;   

 Added hdfs://localhost/FooTest.jar to class path
 hive create temporary function cube as 'com.cloudera.FooTestUDF';
 
 FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.FunctionTask
 Does this work for other people?  I could probably fix it by changing add 
 jar to download remote jars locally, when necessary (to load them into the 
 classpath), or update URLClassLoader (or whatever is underneath there) to 
 read directly from HDFS, which seems a bit more fragile.  But I wanted to 
 make sure that my interpretation of what's going on is right before I have at 
 it.
 Thanks,
 -- Philip
 {quote}
 {quote}
 Yes that's correct. I prefer to download the jars in add jar.
 Zheng
 {quote}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.