[ https://issues.apache.org/jira/browse/HIVE-1373?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12876250#action_12876250 ]
Edward Capriolo commented on HIVE-1373: --------------------------------------- {quote} 1 copy is anyway done from lib to dist/lib for these jars. If we go directly to ivy we would copy things from the ivy cache to dist/lib. So the number of copies in the build process would remain the same, no? There is of course the first time overhead of downloading these jars from their repos to the ivy cache. {quote} I follow what you are thinking. Currently the code I did takes specifc jars from metastore ivy dowloads. We could probably have ivy download directly to build/lib. I just think we should watch to make sure many unneeded jars do not appear. > Missing connection pool plugin in Eclipse classpath > --------------------------------------------------- > > Key: HIVE-1373 > URL: https://issues.apache.org/jira/browse/HIVE-1373 > Project: Hadoop Hive > Issue Type: Bug > Components: Build Infrastructure > Environment: Eclipse, Linux > Reporter: Vinithra Varadharajan > Assignee: Vinithra Varadharajan > Attachments: HIVE-1373.patch > > > In a recent checkin, connection pool dependency was introduced but eclipse > .classpath file was not updated. This causes launch configurations from > within Eclipse to fail. > {code} > hive> show tables; > show tables; > 10/05/26 14:59:46 INFO parse.ParseDriver: Parsing command: show tables > 10/05/26 14:59:46 INFO parse.ParseDriver: Parse Completed > 10/05/26 14:59:46 INFO ql.Driver: Semantic Analysis Completed > 10/05/26 14:59:46 INFO ql.Driver: Returning Hive schema: > Schema(fieldSchemas:[FieldSchema(name:tab_name, type:string, comment:from > deserializer)], properties:null) > 10/05/26 14:59:46 INFO ql.Driver: query plan = > file:/tmp/vinithra/hive_2010-05-26_14-59-46_058_1636674338194744357/queryplan.xml > 10/05/26 14:59:46 INFO ql.Driver: Starting command: show tables > 10/05/26 14:59:46 INFO metastore.HiveMetaStore: 0: Opening raw store with > implemenation class:org.apache.hadoop.hive.metastore.ObjectStore > 10/05/26 14:59:46 INFO metastore.ObjectStore: ObjectStore, initialize called > FAILED: Error in metadata: javax.jdo.JDOFatalInternalException: Error > creating transactional connection factory > NestedThrowables: > java.lang.reflect.InvocationTargetException > 10/05/26 14:59:47 ERROR exec.DDLTask: FAILED: Error in metadata: > javax.jdo.JDOFatalInternalException: Error creating transactional connection > factory > NestedThrowables: > java.lang.reflect.InvocationTargetException > org.apache.hadoop.hive.ql.metadata.HiveException: > javax.jdo.JDOFatalInternalException: Error creating transactional connection > factory > NestedThrowables: > java.lang.reflect.InvocationTargetException > at org.apache.hadoop.hive.ql.metadata.Hive.getTablesForDb(Hive.java:491) > at > org.apache.hadoop.hive.ql.metadata.Hive.getTablesByPattern(Hive.java:472) > at org.apache.hadoop.hive.ql.metadata.Hive.getAllTables(Hive.java:458) > at org.apache.hadoop.hive.ql.exec.DDLTask.showTables(DDLTask.java:504) > at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:176) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:107) > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:55) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:631) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:504) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:382) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:138) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:197) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:303) > Caused by: javax.jdo.JDOFatalInternalException: Error creating transactional > connection factory > NestedThrowables: > java.lang.reflect.InvocationTargetException > at > org.datanucleus.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:395) > at > org.datanucleus.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:547) > at > org.datanucleus.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:175) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > at java.lang.reflect.Method.invoke(Method.java:597) > at javax.jdo.JDOHelper$16.run(JDOHelper.java:1956) > at java.security.AccessController.doPrivileged(Native Method) > at javax.jdo.JDOHelper.invoke(JDOHelper.java:1951) > at > javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1159) > at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:803) > at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:698) > at > org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:191) > at > org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:208) > at > org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:153) > at > org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:128) > at > org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:62) > at > org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:117) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:276) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.executeWithRetry(HiveMetaStore.java:228) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:374) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:166) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.<init>(HiveMetaStore.java:131) > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:83) > at > org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:1077) > at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:1087) > at org.apache.hadoop.hive.ql.metadata.Hive.getTablesForDb(Hive.java:489) > ... 12 more > Caused by: java.lang.reflect.InvocationTargetException > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) > at java.lang.reflect.Constructor.newInstance(Constructor.java:513) > at > org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:576) > at > org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:324) > at > org.datanucleus.store.AbstractStoreManager.<init>(AbstractStoreManager.java:190) > at > org.datanucleus.store.mapped.MappedStoreManager.<init>(MappedStoreManager.java:139) > at > org.datanucleus.store.rdbms.RDBMSManager.<init>(RDBMSManager.java:265) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) > at java.lang.reflect.Constructor.newInstance(Constructor.java:513) > at > org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:576) > at > org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:300) > at > org.datanucleus.store.FederationManager.initialiseStoreManager(FederationManager.java:106) > at > org.datanucleus.store.FederationManager.<init>(FederationManager.java:68) > at > org.datanucleus.ObjectManagerFactoryImpl.initialiseStoreManager(ObjectManagerFactoryImpl.java:152) > at > org.datanucleus.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:529) > ... 38 more > Caused by: org.datanucleus.exceptions.NucleusException: Attempt to invoke the > "DBCP" plugin to create a ConnectionPool gave an error : The connection pool > plugin of type "DBCP" was not found in the CLASSPATH! > at > org.datanucleus.store.rdbms.ConnectionFactoryImpl.initDataSourceTx(ConnectionFactoryImpl.java:169) > at > org.datanucleus.store.rdbms.ConnectionFactoryImpl.<init>(ConnectionFactoryImpl.java:91) > ... 57 more > Caused by: org.datanucleus.exceptions.NucleusUserException: The connection > pool plugin of type "DBCP" was not found in the CLASSPATH! > at > org.datanucleus.store.rdbms.ConnectionFactoryImpl.initDataSourceTx(ConnectionFactoryImpl.java:143) > ... 58 more > FAILED: Execution Error, return code 1 from > org.apache.hadoop.hive.ql.exec.DDLTask > 10/05/26 14:59:47 ERROR ql.Driver: FAILED: Execution Error, return code 1 > from org.apache.hadoop.hive.ql.exec.DDLTask > hive> > {code} -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.