Never mind. Figured out where that was. Thanks.
On Wed, Oct 23, 2013 at 2:27 PM, SF Hadoop <sfhad...@gmail.com> wrote: > Where is package.jdo located? The one that you changed? > > Thanks. > > > > On Wed, Oct 23, 2013 at 1:22 PM, Timothy Potter <thelabd...@gmail.com>wrote: > >> I updated package.jdo to use COMMENT instead of FCOMMENT (which is >> something I had to do for HCatalog a long while back) ... may not be the >> "right" solution but worked for me. >> >> Cheers, >> Tim >> >> >> On Wed, Oct 23, 2013 at 2:09 PM, SF Hadoop <sfhad...@gmail.com> wrote: >> >>> Has anyone come up with further information on this issue? I am >>> experiencing the same thing. >>> >>> Hive is set to auto-create if not exist yet it still fails. I cannot >>> create *any* table at all. >>> >>> Any help is appreciated. >>> >>> >>> >>> >>> On Sat, Oct 19, 2013 at 11:56 PM, Jov <am...@amutu.com> wrote: >>> >>>> can you confirm the script content?there may be bug,you can open a >>>> issue. >>>> >>>> jov >>>> >>>> On Oct 20, 2013 1:11 PM, "Zhang Xiaoyu" <zhangxiaoyu...@gmail.com> >>>> wrote: >>>> > >>>> > Hi, Jov, >>>> > Thanks. I understand turn on those two properties resolve the >>>> problems. But I run the hive 0.12 scheme script. I assume it should create >>>> all required tables. >>>> > >>>> > Johnny >>>> > >>>> > >>>> > On Sat, Oct 19, 2013 at 7:44 PM, Jov <am...@amutu.com> wrote: >>>> >> >>>> >> jov >>>> >> >>>> >> >>>> >> On Oct 20, 2013 8:07 AM, "Zhang Xiaoyu" <zhangxiaoyu...@gmail.com> >>>> wrote: >>>> >> > >>>> >> > Hi, all, >>>> >> > When I using Hive 0.12 with MySQL metastore. I set those >>>> properties in hive-site.xml. >>>> >> > datanucleus.autoCreateSchema = false >>>> >> > datanucleus.autoCreateTables= false >>>> >> you should set these properties to true,then hive will auto add new >>>> column. >>>> >> >>>> >> > >>>> >> > In beeline, "show tables" is fine, but create a new table got >>>> below exception, any idea? Since I create the metastore table by the hive >>>> 0.12 scheme script, it shouldn't complain about the missing columns in >>>> metastore tables. >>>> >> > >>>> >> > Thanks, >>>> >> > Johnny >>>> >> > >>>> >> > ------------------------- >>>> >> > >>>> >> > FAILED: Error in metadata: >>>> MetaException(message:javax.jdo.JDODataStoreException: Add request failed : >>>> INSERT INTO `COLUMNS_V2` >>>> (`CD_ID`,`FCOMMENT`,`COLUMN_NAME`,`TYPE_NAME`,`INTEGER_IDX`) VALUES >>>> (?,?,?,?,?) >>>> >> > at >>>> org.datanucleus.api.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:422) >>>> >> > at >>>> org.datanucleus.api.jdo.JDOPersistenceManager.jdoMakePersistent(JDOPersistenceManager.java:745) >>>> >> > at >>>> org.datanucleus.api.jdo.JDOPersistenceManager.makePersistent(JDOPersistenceManager.java:765) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.ObjectStore.createTable(ObjectStore.java:638) >>>> >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> >> > at >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>> >> > at >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> >> > at java.lang.reflect.Method.invoke(Method.java:601) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.RetryingRawStore.invoke(RetryingRawStore.java:111) >>>> >> > at sun.proxy.$Proxy6.createTable(Unknown Source) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_core(HiveMetaStore.java:1081) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_with_environment_context(HiveMetaStore.java:1114) >>>> >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> >> > at >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>> >> > at >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> >> > at java.lang.reflect.Method.invoke(Method.java:601) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:102) >>>> >> > at sun.proxy.$Proxy8.create_table_with_environment_context(Unknown >>>> Source) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:464) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:453) >>>> >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> >> > at >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>> >> > at >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> >> > at java.lang.reflect.Method.invoke(Method.java:601) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:74) >>>> >> > at sun.proxy.$Proxy10.createTable(Unknown Source) >>>> >> > at >>>> org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:593) >>>> >> > at >>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:3784) >>>> >> > at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:256) >>>> >> > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:144) >>>> >> > at >>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57) >>>> >> > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1355) >>>> >> > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1139) >>>> >> > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:945) >>>> >> > at >>>> org.apache.hive.service.cli.operation.SQLOperation.run(SQLOperation.java:95) >>>> >> > at >>>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:193) >>>> >> > at >>>> org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:148) >>>> >> > at >>>> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:203) >>>> >> > at >>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1133) >>>> >> > at >>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1118) >>>> >> > at >>>> org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) >>>> >> > at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39) >>>> >> > at >>>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:40) >>>> >> > at >>>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:37) >>>> >> > at java.security.AccessController.doPrivileged(Native Method) >>>> >> > at javax.security.auth.Subject.doAs(Subject.java:415) >>>> >> > at >>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232) >>>> >> > at >>>> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:526) >>>> >> > at >>>> org.apache.hive.service.auth.TUGIContainingProcessor.process(TUGIContainingProcessor.java:37) >>>> >> > at >>>> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206) >>>> >> > at >>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>> >> > at >>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>> >> > at java.lang.Thread.run(Thread.java:722) >>>> >> > NestedThrowablesStackTrace: >>>> >> > com.mysql.jdbc.exceptions.jdbc4.MySQLSyntaxErrorException: Unknown >>>> column 'FCOMMENT' in 'field list' >>>> >> this column is new in 0.12,will be auto added if you set those >>>> settings to true >>>> >> >>>> >> > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >>>> Method) >>>> >> > at >>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) >>>> >> > at >>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >>>> >> > at java.lang.reflect.Constructor.newInstance(Constructor.java:525) >>>> >> > at com.mysql.jdbc.Util.handleNewInstance(Util.java:411) >>>> >> > at com.mysql.jdbc.Util.getInstance(Util.java:386) >>>> >> > at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:1052) >>>> >> > at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:3597) >>>> >> > at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:3529) >>>> >> > at com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:1990) >>>> >> > at com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2151) >>>> >> > at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2625) >>>> >> > at >>>> com.mysql.jdbc.PreparedStatement.executeInternal(PreparedStatement.java:2119) >>>> >> > at >>>> com.mysql.jdbc.PreparedStatement.executeUpdate(PreparedStatement.java:2415) >>>> >> > at >>>> com.mysql.jdbc.PreparedStatement.executeUpdate(PreparedStatement.java:2333) >>>> >> > at >>>> com.mysql.jdbc.PreparedStatement.executeUpdate(PreparedStatement.java:2318) >>>> >> > at >>>> org.apache.commons.dbcp.DelegatingPreparedStatement.executeUpdate(DelegatingPreparedStatement.java:105) >>>> >> > at >>>> org.apache.commons.dbcp.DelegatingPreparedStatement.executeUpdate(DelegatingPreparedStatement.java:105) >>>> >> > at >>>> org.datanucleus.store.rdbms.SQLController.executeStatementUpdate(SQLController.java:407) >>>> >> > at >>>> org.datanucleus.store.rdbms.scostore.RDBMSJoinListStore.internalAdd(RDBMSJoinListStore.java:313) >>>> >> > at >>>> org.datanucleus.store.rdbms.scostore.AbstractListStore.addAll(AbstractListStore.java:136) >>>> >> > at >>>> org.datanucleus.store.mapped.mapping.CollectionMapping.postInsert(CollectionMapping.java:134) >>>> >> > at >>>> org.datanucleus.store.rdbms.request.InsertRequest.execute(InsertRequest.java:517) >>>> >> > at >>>> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertTable(RDBMSPersistenceHandler.java:163) >>>> >> > at >>>> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertObject(RDBMSPersistenceHandler.java:139) >>>> >> > at >>>> org.datanucleus.state.JDOStateManager.internalMakePersistent(JDOStateManager.java:2371) >>>> >> > at >>>> org.datanucleus.state.JDOStateManager.makePersistent(JDOStateManager.java:2347) >>>> >> > at >>>> org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1798) >>>> >> > at >>>> org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1886) >>>> >> > at >>>> org.datanucleus.store.mapped.mapping.PersistableMapping.setObjectAsValue(PersistableMapping.java:665) >>>> >> > at >>>> org.datanucleus.store.mapped.mapping.PersistableMapping.setObject(PersistableMapping.java:425) >>>> >> > at >>>> org.datanucleus.store.rdbms.fieldmanager.ParameterSetter.storeObjectField(ParameterSetter.java:197) >>>> >> > at >>>> org.datanucleus.state.AbstractStateManager.providedObjectField(AbstractStateManager.java:1452) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.model.MStorageDescriptor.jdoProvideField(MStorageDescriptor.java) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.model.MStorageDescriptor.jdoProvideFields(MStorageDescriptor.java) >>>> >> > at >>>> org.datanucleus.state.AbstractStateManager.provideFields(AbstractStateManager.java:1520) >>>> >> > at >>>> org.datanucleus.store.rdbms.request.InsertRequest.execute(InsertRequest.java:288) >>>> >> > at >>>> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertTable(RDBMSPersistenceHandler.java:163) >>>> >> > at >>>> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertObject(RDBMSPersistenceHandler.java:139) >>>> >> > at >>>> org.datanucleus.state.JDOStateManager.internalMakePersistent(JDOStateManager.java:2371) >>>> >> > at >>>> org.datanucleus.state.JDOStateManager.makePersistent(JDOStateManager.java:2347) >>>> >> > at >>>> org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1798) >>>> >> > at >>>> org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1886) >>>> >> > at >>>> org.datanucleus.store.mapped.mapping.PersistableMapping.setObjectAsValue(PersistableMapping.java:665) >>>> >> > at >>>> org.datanucleus.store.mapped.mapping.PersistableMapping.setObject(PersistableMapping.java:425) >>>> >> > at >>>> org.datanucleus.store.rdbms.fieldmanager.ParameterSetter.storeObjectField(ParameterSetter.java:197) >>>> >> > at >>>> org.datanucleus.state.AbstractStateManager.providedObjectField(AbstractStateManager.java:1452) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.model.MTable.jdoProvideField(MTable.java) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.model.MTable.jdoProvideFields(MTable.java) >>>> >> > at >>>> org.datanucleus.state.AbstractStateManager.provideFields(AbstractStateManager.java:1520) >>>> >> > at >>>> org.datanucleus.store.rdbms.request.InsertRequest.execute(InsertRequest.java:288) >>>> >> > at >>>> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertTable(RDBMSPersistenceHandler.java:163) >>>> >> > at >>>> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertObject(RDBMSPersistenceHandler.java:139) >>>> >> > at >>>> org.datanucleus.state.JDOStateManager.internalMakePersistent(JDOStateManager.java:2371) >>>> >> > at >>>> org.datanucleus.state.JDOStateManager.makePersistent(JDOStateManager.java:2347) >>>> >> > at >>>> org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1798) >>>> >> > at >>>> org.datanucleus.ObjectManagerImpl.persistObjectWork(ObjectManagerImpl.java:1647) >>>> >> > at >>>> org.datanucleus.ObjectManagerImpl.persistObject(ObjectManagerImpl.java:1504) >>>> >> > at >>>> org.datanucleus.MultithreadedObjectManager.persistObject(MultithreadedObjectManager.java:298) >>>> >> > at >>>> org.datanucleus.api.jdo.JDOPersistenceManager.jdoMakePersistent(JDOPersistenceManager.java:740) >>>> >> > at >>>> org.datanucleus.api.jdo.JDOPersistenceManager.makePersistent(JDOPersistenceManager.java:765) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.ObjectStore.createTable(ObjectStore.java:638) >>>> >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> >> > at >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>> >> > at >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> >> > at java.lang.reflect.Method.invoke(Method.java:601) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.RetryingRawStore.invoke(RetryingRawStore.java:111) >>>> >> > at sun.proxy.$Proxy6.createTable(Unknown Source) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_core(HiveMetaStore.java:1081) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_with_environment_context(HiveMetaStore.java:1114) >>>> >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> >> > at >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>> >> > at >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> >> > at java.lang.reflect.Method.invoke(Method.java:601) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:102) >>>> >> > at sun.proxy.$Proxy8.create_table_with_environment_context(Unknown >>>> Source) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:464) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:453) >>>> >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> >> > at >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>> >> > at >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> >> > at java.lang.reflect.Method.invoke(Method.java:601) >>>> >> > at >>>> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:74) >>>> >> > at sun.proxy.$Proxy10.createTable(Unknown Source) >>>> >> > at >>>> org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:593) >>>> >> > at >>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:3784) >>>> >> > at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:256) >>>> >> > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:144) >>>> >> > at >>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57) >>>> >> > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1355) >>>> >> > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1139) >>>> >> > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:945) >>>> >> > at >>>> org.apache.hive.service.cli.operation.SQLOperation.run(SQLOperation.java:95) >>>> >> > at >>>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:193) >>>> >> > at >>>> org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:148) >>>> >> > at >>>> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:203) >>>> >> > at >>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1133) >>>> >> > at >>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1118) >>>> >> > at >>>> org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) >>>> >> > at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39) >>>> >> > at >>>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:40) >>>> >> > at >>>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:37) >>>> >> > at java.security.AccessController.doPrivileged(Native Method) >>>> >> > at javax.security.auth.Subject.doAs(Subject.java:415) >>>> >> > at >>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232) >>>> >> > at >>>> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:526) >>>> >> > at >>>> org.apache.hive.service.auth.TUGIContainingProcessor.process(TUGIContainingProcessor.java:37) >>>> >> > at >>>> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206) >>>> >> > at >>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>> >> > at >>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>> >> > at java.lang.Thread.run(Thread.java:722) >>>> >> > ) >>>> >> > FAILED: Execution Error, return code 1 from >>>> org.apache.hadoop.hive.ql.exec.DDLTask >>>> >> > org.apache.hive.service.cli.HiveSQLException: Error while >>>> processing statement: FAILED: Execution Error, return code 1 from >>>> org.apache.hadoop.hive.ql.exec.DDLTask >>>> >> > at >>>> org.apache.hive.service.cli.operation.SQLOperation.run(SQLOperation.java:97) >>>> >> > at >>>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:193) >>>> >> > at >>>> org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:148) >>>> >> > at >>>> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:203) >>>> >> > at >>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1133) >>>> >> > at >>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1118) >>>> >> > at >>>> org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) >>>> >> > at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39) >>>> >> > at >>>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:40) >>>> >> > at >>>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:37) >>>> >> > at java.security.AccessController.doPrivileged(Native Method) >>>> >> > at javax.security.auth.Subject.doAs(Subject.java:415) >>>> >> > at >>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232) >>>> >> > at >>>> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:526) >>>> >> > at >>>> org.apache.hive.service.auth.TUGIContainingProcessor.process(TUGIContainingProcessor.java:37) >>>> >> > at >>>> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206) >>>> >> > at >>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>> >> > at >>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>> >> > at java.lang.Thread.run(Thread.java:722) >>>> > >>>> > >>>> >>>> >>> >> >