[ 
https://issues.apache.org/jira/browse/HIVE-1373?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872962#action_12872962
 ] 

Ashish Thusoo commented on HIVE-1373:
-------------------------------------

1 copy is anyway done from lib to dist/lib for these jars. If we go directly to 
ivy we would copy things from the ivy cache to dist/lib. So the number of 
copies in the build process
would remain the same, no? There is of course the first time overhead of 
downloading these jars from their repos to the ivy cache.

> Missing connection pool plugin in Eclipse classpath
> ---------------------------------------------------
>
>                 Key: HIVE-1373
>                 URL: https://issues.apache.org/jira/browse/HIVE-1373
>             Project: Hadoop Hive
>          Issue Type: Bug
>          Components: Build Infrastructure
>         Environment: Eclipse, Linux
>            Reporter: Vinithra Varadharajan
>            Assignee: Vinithra Varadharajan
>            Priority: Minor
>         Attachments: HIVE-1373.patch
>
>
> In a recent checkin, connection pool dependency was introduced but eclipse 
> .classpath file was not updated.  This causes launch configurations from 
> within Eclipse to fail.
> {code}
> hive> show tables;
> show tables;
> 10/05/26 14:59:46 INFO parse.ParseDriver: Parsing command: show tables
> 10/05/26 14:59:46 INFO parse.ParseDriver: Parse Completed
> 10/05/26 14:59:46 INFO ql.Driver: Semantic Analysis Completed
> 10/05/26 14:59:46 INFO ql.Driver: Returning Hive schema: 
> Schema(fieldSchemas:[FieldSchema(name:tab_name, type:string, comment:from 
> deserializer)], properties:null)
> 10/05/26 14:59:46 INFO ql.Driver: query plan = 
> file:/tmp/vinithra/hive_2010-05-26_14-59-46_058_1636674338194744357/queryplan.xml
> 10/05/26 14:59:46 INFO ql.Driver: Starting command: show tables
> 10/05/26 14:59:46 INFO metastore.HiveMetaStore: 0: Opening raw store with 
> implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
> 10/05/26 14:59:46 INFO metastore.ObjectStore: ObjectStore, initialize called
> FAILED: Error in metadata: javax.jdo.JDOFatalInternalException: Error 
> creating transactional connection factory
> NestedThrowables:
> java.lang.reflect.InvocationTargetException
> 10/05/26 14:59:47 ERROR exec.DDLTask: FAILED: Error in metadata: 
> javax.jdo.JDOFatalInternalException: Error creating transactional connection 
> factory
> NestedThrowables:
> java.lang.reflect.InvocationTargetException
> org.apache.hadoop.hive.ql.metadata.HiveException: 
> javax.jdo.JDOFatalInternalException: Error creating transactional connection 
> factory
> NestedThrowables:
> java.lang.reflect.InvocationTargetException
>       at org.apache.hadoop.hive.ql.metadata.Hive.getTablesForDb(Hive.java:491)
>       at 
> org.apache.hadoop.hive.ql.metadata.Hive.getTablesByPattern(Hive.java:472)
>       at org.apache.hadoop.hive.ql.metadata.Hive.getAllTables(Hive.java:458)
>       at org.apache.hadoop.hive.ql.exec.DDLTask.showTables(DDLTask.java:504)
>       at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:176)
>       at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:107)
>       at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:55)
>       at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:631)
>       at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:504)
>       at org.apache.hadoop.hive.ql.Driver.run(Driver.java:382)
>       at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:138)
>       at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:197)
>       at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:303)
> Caused by: javax.jdo.JDOFatalInternalException: Error creating transactional 
> connection factory
> NestedThrowables:
> java.lang.reflect.InvocationTargetException
>       at 
> org.datanucleus.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:395)
>       at 
> org.datanucleus.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:547)
>       at 
> org.datanucleus.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:175)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>       at java.lang.reflect.Method.invoke(Method.java:597)
>       at javax.jdo.JDOHelper$16.run(JDOHelper.java:1956)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.jdo.JDOHelper.invoke(JDOHelper.java:1951)
>       at 
> javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1159)
>       at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:803)
>       at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:698)
>       at 
> org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:191)
>       at 
> org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:208)
>       at 
> org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:153)
>       at 
> org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:128)
>       at 
> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:62)
>       at 
> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:117)
>       at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:276)
>       at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.executeWithRetry(HiveMetaStore.java:228)
>       at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:374)
>       at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:166)
>       at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.<init>(HiveMetaStore.java:131)
>       at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:83)
>       at 
> org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:1077)
>       at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:1087)
>       at org.apache.hadoop.hive.ql.metadata.Hive.getTablesForDb(Hive.java:489)
>       ... 12 more
> Caused by: java.lang.reflect.InvocationTargetException
>       at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>       at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
>       at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>       at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>       at 
> org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:576)
>       at 
> org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:324)
>       at 
> org.datanucleus.store.AbstractStoreManager.<init>(AbstractStoreManager.java:190)
>       at 
> org.datanucleus.store.mapped.MappedStoreManager.<init>(MappedStoreManager.java:139)
>       at 
> org.datanucleus.store.rdbms.RDBMSManager.<init>(RDBMSManager.java:265)
>       at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>       at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
>       at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>       at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>       at 
> org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:576)
>       at 
> org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:300)
>       at 
> org.datanucleus.store.FederationManager.initialiseStoreManager(FederationManager.java:106)
>       at 
> org.datanucleus.store.FederationManager.<init>(FederationManager.java:68)
>       at 
> org.datanucleus.ObjectManagerFactoryImpl.initialiseStoreManager(ObjectManagerFactoryImpl.java:152)
>       at 
> org.datanucleus.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:529)
>       ... 38 more
> Caused by: org.datanucleus.exceptions.NucleusException: Attempt to invoke the 
> "DBCP" plugin to create a ConnectionPool gave an error : The connection pool 
> plugin of type "DBCP" was not found in the CLASSPATH!
>       at 
> org.datanucleus.store.rdbms.ConnectionFactoryImpl.initDataSourceTx(ConnectionFactoryImpl.java:169)
>       at 
> org.datanucleus.store.rdbms.ConnectionFactoryImpl.<init>(ConnectionFactoryImpl.java:91)
>       ... 57 more
> Caused by: org.datanucleus.exceptions.NucleusUserException: The connection 
> pool plugin of type "DBCP" was not found in the CLASSPATH!
>       at 
> org.datanucleus.store.rdbms.ConnectionFactoryImpl.initDataSourceTx(ConnectionFactoryImpl.java:143)
>       ... 58 more
> FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.DDLTask
> 10/05/26 14:59:47 ERROR ql.Driver: FAILED: Execution Error, return code 1 
> from org.apache.hadoop.hive.ql.exec.DDLTask
> hive> 
> {code}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to