[ https://issues.apache.org/jira/browse/HIVE-26759?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Akshat Mathur updated HIVE-26759: --------------------------------- Fix Version/s: 4.0.0-alpha-2 Resolution: Fixed Status: Resolved (was: Patch Available) > ERROR: column "CC_START" does not exist, when Postgres is used as Hive > metastore > -------------------------------------------------------------------------------- > > Key: HIVE-26759 > URL: https://issues.apache.org/jira/browse/HIVE-26759 > Project: Hive > Issue Type: Bug > Components: Metastore > Affects Versions: 4.0.0-alpha-2 > Reporter: Akshat Mathur > Assignee: Akshat Mathur > Priority: Major > Labels: pull-request-available > Fix For: 4.0.0-alpha-2 > > Time Spent: 1h 40m > Remaining Estimate: 0h > > This error is coming when Postgres is used as Hive Metastore. > hive-site.xml > > {code:java} > <?xml version="1.0"?> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > <configuration> > <!-- UPSTREAM --> > <property> > <name>hive.server2.logging.operation.level</name> > <value>NONE</value> > </property> > <property> > <name>hive.log4j.file</name> > <value>hive-log4j.properties</value> > </property> > <property> > <name>metastore.log4j.file</name> > <value>metastore-log4j.properties</value> > </property> > <!-- Intellij --> > <property> > <name>hive.jar.path</name> > > <value>/Users/am/Desktop/work/upstream/hive/ql/target/hive-exec-4.0.0-SNAPSHOT.jar</value> > <description>The location of hive_cli.jar that is used when > submitting jobs in a separate jvm.</description> > </property> > <property> > <name>hive.hadoop.classpath</name> > > <value>/Users/am/Desktop/work/upstream/hive/ql/target/hive-exec-4.0.0-SNAPSHOT.jar</value> > </property> > <property> > <name>hive.metastore.local</name> > <value>false</value> > </property> > <property> > <name>hive.metastore.uris</name> > <value>thrift://localhost:9083</value> > </property> > <property> > <name>hive.metastore.warehouse.dir</name> > <value>/Users/am/Desktop/work/hivestuff/warehouse</value> > </property> > <property> > <name>hive.server2.metrics.enabled</name> > <value>true</value> > </property> > > <property> > <name>spark.eventLog.enabled</name> > <value>true</value> > </property> > <property> > <name>spark.eventLog.dir</name> > <value>/tmp/hive</value> > </property> > <!-- Intellij --> > > <property> > <name>metastore.metastore.event.db.notification.api.auth</name> > <value>false</value> > </property> > <property> > <name>hive.metastore.schema.verification</name> > <value>false</value> > </property> > <property> > <name>datanucleus.autoCreateTables</name> > <value>true</value> > </property> > > <property> > <name>hive.exec.scratchdir</name> > <values>/tmp/hive-${user.name}</values> > </property> > <property> > <name>javax.jdo.option.ConnectionURL</name> > <value>jdbc:postgresql://localhost:5432/hive_metastore</value> > <description>JDBC connect string for a JDBC > metastore</description> > </property> > <property> > <name>javax.jdo.option.ConnectionDriverName</name> > <value>org.postgresql.Driver</value> > </property> > <property> > <name>javax.jdo.option.ConnectionUserName</name> > <value>hive</value> > </property> > <property> > <name>javax.jdo.option.ConnectionPassword</name> > <value>hive</value> > </property> > <property> > <name>datanucleus.schema.autoCreateAll</name> > <value>true</value> > </property> > <property> > <name>hive.server2.enable.doAs</name> > <value>false</value> > <description></description> > </property> > <property> > <name>hive.server2.enable.impersonation</name> > <value>false</value> > <description></description> > </property> > <property> > <name>dfs.namenode.acls.enabled</name> > <value>false</value> > </property> > <!-- FAIR SCHEDULER --> > <!-- These following lines are needed to use ACID features --> > <!-- BEGIN --> > <!-- > <property> > <name>hive.enforce.bucketing</name> > <value>true</value> > </property> > <property> > <name>hive.support.concurrency</name> > <value>true</value> > </property> > <property> > <name>hive.exec.dynamic.partition.mode</name> > <value>nonstrict</value> > </property> > <property> > <name>hive.txn.manager</name> > <value>org.apache.hadoop.hive.ql.lockmgr.DbTxnManager</value> > </property> > <property> > <name>hive.lock.manager</name> > <value>org.apache.hadoop.hive.ql.lockmgr.DbLockManager</value> > </property> > <property> > <name>hive.compactor.initiator.on</name> > <value>true</value> > </property> > <property> > <name>hive.compactor.worker.threads</name> > <value>2</value> > </property> > --> > <!-- END --> > <property> > <name>hive.server2.webui.explain.output</name> > <value>true</value> > </property> > <property> > <name>hive.server2.webui.show.graph</name> > <value>true</value> > </property> > <property> > <name>hive.server2.webui.show.stats</name> > <value>true</value> > </property> > <property> > <name>hive.server2.webui.max.graph.size</name> > <value>40</value> > </property> > <!-- ACID --> > <property> > <name>hive.txn.manager</name> > <value>org.apache.hadoop.hive.ql.lockmgr.DbTxnManager</value> > </property> > <property> > <name>hive.compactor.initiator.on</name> > <value>true</value> > </property> > <property> > <name>hive.compactor.worker.threads</name> > <value>3</value> > </property> > <property> > <name>metastore.compactor.worker.threads</name> > <value>4</value> > </property> > <property> > <name>hive.support.concurrency</name> > <value>true</value> > </property> > <property> > <name>hive.exec.dynamic.partition.mode</name> > <value>nonstrict</value> > </property> > <property> > <name>hive.lock.manager</name> > <value>org.apache.hadoop.hive.ql.lockmgr.DbLockManager</value> > </property> > <property> > <name>hive.compactor.crud.query.based</name> > <value>true</value> > </property> > <property> > <name>hive.metastore.runworker.in</name> > <value>hs2</value> > </property> > <!-- Random --> > <!-- <property> > <name>hive.users.in.admin.role</name> > <value>karencoppage</value> > </property> --> > <!--Timestamp--> > <!-- <property> > <name>hive.parquet.write.int64.timestamp</name> > <value>true</value> > </property> > --> > <!--for WebUI explain plan--> > <!-- <property> > <name>hive.server2.webui.max.historic.queries</name> > <value>40</value> > </property> --> > <!-- <property> > <name></name> > <value></value> > </property> > --> > </configuration> > {code} > > > Following is the stack trace when HMS service is started: > {code:java} > [Thread-5] ERROR org.apache.hadoop.hive.ql.txn.compactor.Initiator - > Initiator loop caught unexpected exception this time through the loop > org.apache.hadoop.hive.metastore.api.MetaException: Unable to select from > transaction database org.postgresql.util.PSQLException: ERROR: column > "CC_START" does not exist > Position: 1215 > at > org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2676) > at > org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2366) > at > org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:356) > at org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:490) > at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:408) > at > org.postgresql.jdbc.PgPreparedStatement.executeWithFlags(PgPreparedStatement.java:181) > at > org.postgresql.jdbc.PgPreparedStatement.executeQuery(PgPreparedStatement.java:133) > at > com.zaxxer.hikari.pool.ProxyPreparedStatement.executeQuery(ProxyPreparedStatement.java:52) > at > com.zaxxer.hikari.pool.HikariProxyPreparedStatement.executeQuery(HikariProxyPreparedStatement.java) > at > org.apache.hadoop.hive.metastore.txn.TxnHandler.showCompact(TxnHandler.java:3894) > at > org.apache.hadoop.hive.ql.txn.compactor.Initiator.run(Initiator.java:154) > at > org.apache.hadoop.hive.metastore.txn.TxnHandler.showCompact(TxnHandler.java:3946) > ~[classes/:?] > at > org.apache.hadoop.hive.ql.txn.compactor.Initiator.run(Initiator.java:154) > ~[hive-exec-4.0.0-SNAPSHOT.jar:4.0.0-SNAPSHOT] {code} > This error disappears when derby is configured as HMS. -- This message was sent by Atlassian Jira (v8.20.10#820010)