Never mind.  Figured out where that was.

Thanks.



On Wed, Oct 23, 2013 at 2:27 PM, SF Hadoop <sfhad...@gmail.com> wrote:

> Where is package.jdo located?  The one that you changed?
>
> Thanks.
>
>
>
> On Wed, Oct 23, 2013 at 1:22 PM, Timothy Potter <thelabd...@gmail.com>wrote:
>
>> I updated package.jdo to use COMMENT instead of FCOMMENT (which is
>> something I had to do for HCatalog a long while back) ... may not be the
>> "right" solution but worked for me.
>>
>> Cheers,
>> Tim
>>
>>
>> On Wed, Oct 23, 2013 at 2:09 PM, SF Hadoop <sfhad...@gmail.com> wrote:
>>
>>> Has anyone come up with further information on this issue?  I am
>>> experiencing the same thing.
>>>
>>> Hive is set to auto-create if not exist yet it still fails.  I cannot
>>> create *any* table at all.
>>>
>>> Any help is appreciated.
>>>
>>>
>>>
>>>
>>> On Sat, Oct 19, 2013 at 11:56 PM, Jov <am...@amutu.com> wrote:
>>>
>>>> can you confirm the script content?there may be bug,you can open a
>>>> issue.
>>>>
>>>> jov
>>>>
>>>> On Oct 20, 2013 1:11 PM, "Zhang Xiaoyu" <zhangxiaoyu...@gmail.com>
>>>> wrote:
>>>> >
>>>> > Hi, Jov,
>>>> > Thanks. I understand turn on those two properties resolve the
>>>> problems. But I run the hive 0.12 scheme script. I assume it should create
>>>> all required tables.
>>>> >
>>>> > Johnny
>>>> >
>>>> >
>>>> > On Sat, Oct 19, 2013 at 7:44 PM, Jov <am...@amutu.com> wrote:
>>>> >>
>>>> >> jov
>>>> >>
>>>> >>
>>>> >> On Oct 20, 2013 8:07 AM, "Zhang Xiaoyu" <zhangxiaoyu...@gmail.com>
>>>> wrote:
>>>> >> >
>>>> >> > Hi, all,
>>>> >> > When I using Hive 0.12 with MySQL metastore. I set those
>>>> properties in hive-site.xml.
>>>> >> > datanucleus.autoCreateSchema = false
>>>> >> > datanucleus.autoCreateTables= false
>>>> >> you should set these properties to true,then hive will auto add new
>>>> column.
>>>> >>
>>>> >> >
>>>> >> > In beeline, "show tables" is fine, but create a new table got
>>>> below exception, any idea? Since I create the metastore table by the hive
>>>> 0.12 scheme script, it shouldn't complain about the missing columns in
>>>> metastore tables.
>>>> >> >
>>>> >> > Thanks,
>>>> >> > Johnny
>>>> >> >
>>>> >> > -------------------------
>>>> >> >
>>>> >> > FAILED: Error in metadata:
>>>> MetaException(message:javax.jdo.JDODataStoreException: Add request failed :
>>>> INSERT INTO `COLUMNS_V2`
>>>> (`CD_ID`,`FCOMMENT`,`COLUMN_NAME`,`TYPE_NAME`,`INTEGER_IDX`) VALUES
>>>> (?,?,?,?,?)
>>>> >> > at
>>>> org.datanucleus.api.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:422)
>>>> >> > at
>>>> org.datanucleus.api.jdo.JDOPersistenceManager.jdoMakePersistent(JDOPersistenceManager.java:745)
>>>> >> > at
>>>> org.datanucleus.api.jdo.JDOPersistenceManager.makePersistent(JDOPersistenceManager.java:765)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.ObjectStore.createTable(ObjectStore.java:638)
>>>> >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>> >> > at
>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>> >> > at
>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>> >> > at java.lang.reflect.Method.invoke(Method.java:601)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.RetryingRawStore.invoke(RetryingRawStore.java:111)
>>>> >> > at sun.proxy.$Proxy6.createTable(Unknown Source)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_core(HiveMetaStore.java:1081)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_with_environment_context(HiveMetaStore.java:1114)
>>>> >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>> >> > at
>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>> >> > at
>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>> >> > at java.lang.reflect.Method.invoke(Method.java:601)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:102)
>>>> >> > at sun.proxy.$Proxy8.create_table_with_environment_context(Unknown
>>>> Source)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:464)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:453)
>>>> >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>> >> > at
>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>> >> > at
>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>> >> > at java.lang.reflect.Method.invoke(Method.java:601)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:74)
>>>> >> > at sun.proxy.$Proxy10.createTable(Unknown Source)
>>>> >> > at
>>>> org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:593)
>>>> >> > at
>>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:3784)
>>>> >> > at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:256)
>>>> >> > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:144)
>>>> >> > at
>>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57)
>>>> >> > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1355)
>>>> >> > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1139)
>>>> >> > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:945)
>>>> >> > at
>>>> org.apache.hive.service.cli.operation.SQLOperation.run(SQLOperation.java:95)
>>>> >> > at
>>>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:193)
>>>> >> > at
>>>> org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:148)
>>>> >> > at
>>>> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:203)
>>>> >> > at
>>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1133)
>>>> >> > at
>>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1118)
>>>> >> > at
>>>> org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>>>> >> > at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>>>> >> > at
>>>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:40)
>>>> >> > at
>>>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:37)
>>>> >> > at java.security.AccessController.doPrivileged(Native Method)
>>>> >> > at javax.security.auth.Subject.doAs(Subject.java:415)
>>>> >> > at
>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232)
>>>> >> > at
>>>> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:526)
>>>> >> > at
>>>> org.apache.hive.service.auth.TUGIContainingProcessor.process(TUGIContainingProcessor.java:37)
>>>> >> > at
>>>> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>>>> >> > at
>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>> >> > at
>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>> >> > at java.lang.Thread.run(Thread.java:722)
>>>> >> > NestedThrowablesStackTrace:
>>>> >> > com.mysql.jdbc.exceptions.jdbc4.MySQLSyntaxErrorException: Unknown
>>>> column 'FCOMMENT' in 'field list'
>>>> >> this column is new in 0.12,will be auto added if you set those
>>>> settings to true
>>>> >>
>>>> >> > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>>>> Method)
>>>> >> > at
>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
>>>> >> > at
>>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>>>> >> > at java.lang.reflect.Constructor.newInstance(Constructor.java:525)
>>>> >> > at com.mysql.jdbc.Util.handleNewInstance(Util.java:411)
>>>> >> > at com.mysql.jdbc.Util.getInstance(Util.java:386)
>>>> >> > at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:1052)
>>>> >> > at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:3597)
>>>> >> > at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:3529)
>>>> >> > at com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:1990)
>>>> >> > at com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2151)
>>>> >> > at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2625)
>>>> >> > at
>>>> com.mysql.jdbc.PreparedStatement.executeInternal(PreparedStatement.java:2119)
>>>> >> > at
>>>> com.mysql.jdbc.PreparedStatement.executeUpdate(PreparedStatement.java:2415)
>>>> >> > at
>>>> com.mysql.jdbc.PreparedStatement.executeUpdate(PreparedStatement.java:2333)
>>>> >> > at
>>>> com.mysql.jdbc.PreparedStatement.executeUpdate(PreparedStatement.java:2318)
>>>> >> > at
>>>> org.apache.commons.dbcp.DelegatingPreparedStatement.executeUpdate(DelegatingPreparedStatement.java:105)
>>>> >> > at
>>>> org.apache.commons.dbcp.DelegatingPreparedStatement.executeUpdate(DelegatingPreparedStatement.java:105)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.SQLController.executeStatementUpdate(SQLController.java:407)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.scostore.RDBMSJoinListStore.internalAdd(RDBMSJoinListStore.java:313)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.scostore.AbstractListStore.addAll(AbstractListStore.java:136)
>>>> >> > at
>>>> org.datanucleus.store.mapped.mapping.CollectionMapping.postInsert(CollectionMapping.java:134)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.request.InsertRequest.execute(InsertRequest.java:517)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertTable(RDBMSPersistenceHandler.java:163)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertObject(RDBMSPersistenceHandler.java:139)
>>>> >> > at
>>>> org.datanucleus.state.JDOStateManager.internalMakePersistent(JDOStateManager.java:2371)
>>>> >> > at
>>>> org.datanucleus.state.JDOStateManager.makePersistent(JDOStateManager.java:2347)
>>>> >> > at
>>>> org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1798)
>>>> >> > at
>>>> org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1886)
>>>> >> > at
>>>> org.datanucleus.store.mapped.mapping.PersistableMapping.setObjectAsValue(PersistableMapping.java:665)
>>>> >> > at
>>>> org.datanucleus.store.mapped.mapping.PersistableMapping.setObject(PersistableMapping.java:425)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.fieldmanager.ParameterSetter.storeObjectField(ParameterSetter.java:197)
>>>> >> > at
>>>> org.datanucleus.state.AbstractStateManager.providedObjectField(AbstractStateManager.java:1452)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.model.MStorageDescriptor.jdoProvideField(MStorageDescriptor.java)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.model.MStorageDescriptor.jdoProvideFields(MStorageDescriptor.java)
>>>> >> > at
>>>> org.datanucleus.state.AbstractStateManager.provideFields(AbstractStateManager.java:1520)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.request.InsertRequest.execute(InsertRequest.java:288)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertTable(RDBMSPersistenceHandler.java:163)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertObject(RDBMSPersistenceHandler.java:139)
>>>> >> > at
>>>> org.datanucleus.state.JDOStateManager.internalMakePersistent(JDOStateManager.java:2371)
>>>> >> > at
>>>> org.datanucleus.state.JDOStateManager.makePersistent(JDOStateManager.java:2347)
>>>> >> > at
>>>> org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1798)
>>>> >> > at
>>>> org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1886)
>>>> >> > at
>>>> org.datanucleus.store.mapped.mapping.PersistableMapping.setObjectAsValue(PersistableMapping.java:665)
>>>> >> > at
>>>> org.datanucleus.store.mapped.mapping.PersistableMapping.setObject(PersistableMapping.java:425)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.fieldmanager.ParameterSetter.storeObjectField(ParameterSetter.java:197)
>>>> >> > at
>>>> org.datanucleus.state.AbstractStateManager.providedObjectField(AbstractStateManager.java:1452)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.model.MTable.jdoProvideField(MTable.java)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.model.MTable.jdoProvideFields(MTable.java)
>>>> >> > at
>>>> org.datanucleus.state.AbstractStateManager.provideFields(AbstractStateManager.java:1520)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.request.InsertRequest.execute(InsertRequest.java:288)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertTable(RDBMSPersistenceHandler.java:163)
>>>> >> > at
>>>> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertObject(RDBMSPersistenceHandler.java:139)
>>>> >> > at
>>>> org.datanucleus.state.JDOStateManager.internalMakePersistent(JDOStateManager.java:2371)
>>>> >> > at
>>>> org.datanucleus.state.JDOStateManager.makePersistent(JDOStateManager.java:2347)
>>>> >> > at
>>>> org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1798)
>>>> >> > at
>>>> org.datanucleus.ObjectManagerImpl.persistObjectWork(ObjectManagerImpl.java:1647)
>>>> >> > at
>>>> org.datanucleus.ObjectManagerImpl.persistObject(ObjectManagerImpl.java:1504)
>>>> >> > at
>>>> org.datanucleus.MultithreadedObjectManager.persistObject(MultithreadedObjectManager.java:298)
>>>> >> > at
>>>> org.datanucleus.api.jdo.JDOPersistenceManager.jdoMakePersistent(JDOPersistenceManager.java:740)
>>>> >> > at
>>>> org.datanucleus.api.jdo.JDOPersistenceManager.makePersistent(JDOPersistenceManager.java:765)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.ObjectStore.createTable(ObjectStore.java:638)
>>>> >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>> >> > at
>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>> >> > at
>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>> >> > at java.lang.reflect.Method.invoke(Method.java:601)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.RetryingRawStore.invoke(RetryingRawStore.java:111)
>>>> >> > at sun.proxy.$Proxy6.createTable(Unknown Source)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_core(HiveMetaStore.java:1081)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_with_environment_context(HiveMetaStore.java:1114)
>>>> >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>> >> > at
>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>> >> > at
>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>> >> > at java.lang.reflect.Method.invoke(Method.java:601)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:102)
>>>> >> > at sun.proxy.$Proxy8.create_table_with_environment_context(Unknown
>>>> Source)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:464)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:453)
>>>> >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>> >> > at
>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>> >> > at
>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>> >> > at java.lang.reflect.Method.invoke(Method.java:601)
>>>> >> > at
>>>> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:74)
>>>> >> > at sun.proxy.$Proxy10.createTable(Unknown Source)
>>>> >> > at
>>>> org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:593)
>>>> >> > at
>>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:3784)
>>>> >> > at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:256)
>>>> >> > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:144)
>>>> >> > at
>>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57)
>>>> >> > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1355)
>>>> >> > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1139)
>>>> >> > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:945)
>>>> >> > at
>>>> org.apache.hive.service.cli.operation.SQLOperation.run(SQLOperation.java:95)
>>>> >> > at
>>>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:193)
>>>> >> > at
>>>> org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:148)
>>>> >> > at
>>>> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:203)
>>>> >> > at
>>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1133)
>>>> >> > at
>>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1118)
>>>> >> > at
>>>> org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>>>> >> > at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>>>> >> > at
>>>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:40)
>>>> >> > at
>>>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:37)
>>>> >> > at java.security.AccessController.doPrivileged(Native Method)
>>>> >> > at javax.security.auth.Subject.doAs(Subject.java:415)
>>>> >> > at
>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232)
>>>> >> > at
>>>> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:526)
>>>> >> > at
>>>> org.apache.hive.service.auth.TUGIContainingProcessor.process(TUGIContainingProcessor.java:37)
>>>> >> > at
>>>> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>>>> >> > at
>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>> >> > at
>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>> >> > at java.lang.Thread.run(Thread.java:722)
>>>> >> > )
>>>> >> > FAILED: Execution Error, return code 1 from
>>>> org.apache.hadoop.hive.ql.exec.DDLTask
>>>> >> > org.apache.hive.service.cli.HiveSQLException: Error while
>>>> processing statement: FAILED: Execution Error, return code 1 from
>>>> org.apache.hadoop.hive.ql.exec.DDLTask
>>>> >> > at
>>>> org.apache.hive.service.cli.operation.SQLOperation.run(SQLOperation.java:97)
>>>> >> > at
>>>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:193)
>>>> >> > at
>>>> org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:148)
>>>> >> > at
>>>> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:203)
>>>> >> > at
>>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1133)
>>>> >> > at
>>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1118)
>>>> >> > at
>>>> org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>>>> >> > at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>>>> >> > at
>>>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:40)
>>>> >> > at
>>>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:37)
>>>> >> > at java.security.AccessController.doPrivileged(Native Method)
>>>> >> > at javax.security.auth.Subject.doAs(Subject.java:415)
>>>> >> > at
>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232)
>>>> >> > at
>>>> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:526)
>>>> >> > at
>>>> org.apache.hive.service.auth.TUGIContainingProcessor.process(TUGIContainingProcessor.java:37)
>>>> >> > at
>>>> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>>>> >> > at
>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>> >> > at
>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>> >> > at java.lang.Thread.run(Thread.java:722)
>>>> >
>>>> >
>>>>
>>>>
>>>
>>
>

Reply via email to