Just tested analyze table with a trunk build (from yesterday, oct 4th). tried various variations (with or without partitions) of it, but regardless of what I try, I either get: -- analyze table normalized compute statistics;
FAILED: Error in semantic analysis: Table is partitioned and partition specification is needed -- Fair enough if it is not supported, but specifying no partitions seems to be supported according to the docs at http://wiki.apache.org/hadoop/Hive/StatsDev ? -- analyze table normalized partition(intdate) compute statistics; FAILED: Error in semantic analysis: line 1:36 Dynamic partition cannot be the parent of a static partition intdate -- ok, I may understand this (or maybe not :)) may be good to add some notes about it on the wiki though Then the OOM: analyze table normalized partition(intdate,country,logtype,service,hostname,filedate,filedate_ext) compute statistics; Exception in thread "main" java.lang.OutOfMemoryError: Java heap space at java.util.zip.InflaterInputStream.<init>(InflaterInputStream.java:71) at java.util.zip.ZipFile$1.<init>(ZipFile.java:212) at java.util.zip.ZipFile.getInputStream(ZipFile.java:212) at java.util.zip.ZipFile.getInputStream(ZipFile.java:180) at java.util.jar.JarFile.getManifestFromReference(JarFile.java:167) at java.util.jar.JarFile.getManifest(JarFile.java:148) at sun.misc.URLClassPath$JarLoader$2.getManifest(URLClassPath.java:696) at java.net.URLClassLoader.defineClass(URLClassLoader.java:228) at java.net.URLClassLoader.access$000(URLClassLoader.java:58) at java.net.URLClassLoader$1.run(URLClassLoader.java:197) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:190) at java.lang.ClassLoader.loadClass(ClassLoader.java:307) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) at java.lang.ClassLoader.loadClass(ClassLoader.java:248) at org.datanucleus.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:262) at org.datanucleus.jdo.state.JDOStateManagerImpl.isLoaded(JDOStateManagerImpl.java:2020) at org.apache.hadoop.hive.metastore.model.MStorageDescriptor.jdoGetsortCols(MStorageDescriptor.java) at org.apache.hadoop.hive.metastore.model.MStorageDescriptor.getSortCols(MStorageDescriptor.java:206) at org.apache.hadoop.hive.metastore.ObjectStore.convertToStorageDescriptor(ObjectStore.java:759) at org.apache.hadoop.hive.metastore.ObjectStore.convertToPart(ObjectStore.java:859) at org.apache.hadoop.hive.metastore.ObjectStore.convertToParts(ObjectStore.java:896) at org.apache.hadoop.hive.metastore.ObjectStore.getPartitions(ObjectStore.java:886) at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$21.run(HiveMetaStore.java:1333) at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$21.run(HiveMetaStore.java:1330) at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.executeWithRetry(HiveMetaStore.java:234) at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.get_partitions(HiveMetaStore.java:1330) at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.get_partitions_ps(HiveMetaStore.java:1760) at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.listPartitions(HiveMetaStoreClient.java:515) at org.apache.hadoop.hive.ql.metadata.Hive.getPartitions(Hive.java:1267) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.setupStats(SemanticAnalyzer.java:5793) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genTablePlan(SemanticAnalyzer.java:5603) the actual stack is different from each execution of analyze. Another version: Exception in thread "main" java.lang.OutOfMemoryError: Java heap space at java.util.Arrays.copyOf(Arrays.java:2882) at java.lang.AbstractStringBuilder.expandCapacity(AbstractStringBuilder.java:100) at java.lang.AbstractStringBuilder.append(AbstractStringBuilder.java:597) at java.lang.StringBuilder.append(StringBuilder.java:212) at org.datanucleus.JDOClassLoaderResolver.newCacheKey(JDOClassLoaderResolver.java:382) at org.datanucleus.JDOClassLoaderResolver.classForName(JDOClassLoaderResolver.java:173) at org.datanucleus.JDOClassLoaderResolver.classForName(JDOClassLoaderResolver.java:412) at org.datanucleus.store.mapped.mapping.EmbeddedMapping.getJavaType(EmbeddedMapping.java:574) at org.datanucleus.store.mapped.mapping.EmbeddedMapping.getObject(EmbeddedMapping.java:455) at org.datanucleus.store.mapped.scostore.ListStoreIterator.<init>(ListStoreIterator.java:94) at org.datanucleus.store.rdbms.scostore.RDBMSListStoreIterator.<init>(RDBMSListStoreIterator.java:41) at org.datanucleus.store.rdbms.scostore.RDBMSJoinListStore.listIterator(RDBMSJoinListStore.java:158) at org.datanucleus.store.mapped.scostore.AbstractListStore.listIterator(AbstractListStore.java:84) at org.datanucleus.store.mapped.scostore.AbstractListStore.iterator(AbstractListStore.java:74) at org.datanucleus.store.types.sco.backed.List.loadFromStore(List.java:241) at org.datanucleus.store.types.sco.backed.List.iterator(List.java:494) at org.apache.hadoop.hive.metastore.ObjectStore.convertToFieldSchemas(ObjectStore.java:706) at org.apache.hadoop.hive.metastore.ObjectStore.convertToStorageDescriptor(ObjectStore.java:759) at org.apache.hadoop.hive.metastore.ObjectStore.convertToPart(ObjectStore.java:859) at org.apache.hadoop.hive.metastore.ObjectStore.convertToParts(ObjectStore.java:896) at org.apache.hadoop.hive.metastore.ObjectStore.getPartitions(ObjectStore.java:886) at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$21.run(HiveMetaStore.java:1333) at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$21.run(HiveMetaStore.java:1330) at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.executeWithRetry(HiveMetaStore.java:234) at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.get_partitions(HiveMetaStore.java:1330) at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.get_partitions_ps(HiveMetaStore.java:1760) at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.listPartitions(HiveMetaStoreClient.java:515) at org.apache.hadoop.hive.ql.metadata.Hive.getPartitions(Hive.java:1267) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.setupStats(SemanticAnalyzer.java:5793) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genTablePlan(SemanticAnalyzer.java:5603) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:5834) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:6432) Makes no difference if I limit this to a single partition or not or any other variation of the partition specification. It is a sequence file based table, dynamic and static partitions as well as compression. Best regards, Terje