[ https://issues.apache.org/jira/browse/HIVE-1536?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12898655#action_12898655 ]
Sean Flatley commented on HIVE-1536: ------------------------------------ The log entries can be found at the end of this comment. I did find in the Eclipse console an error entry reporting that JAVA_HOME is not set, so it does appear to be configuration: [junit] PREHOOK: query: select a,b,c,d,f as e,f*2 from testHiveJdbcDriverTable limit 1 [junit] PREHOOK: type: QUERY [junit] PREHOOK: Input: defa...@testhivejdbcdrivertable [junit] PREHOOK: Output: file:/tmp/sean/hive_2010-08-14_17-32-24_783_1792704952460259887/-mr-10000 [junit] Total MapReduce jobs = 1 [junit] Launching Job 1 out of 1 [junit] Number of reduce tasks is set to 0 since there's no reduce operator [junit] Error: JAVA_HOME is not set. [junit] FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MapRedTask I do have JAVA_HOME set and exported in my .bashrc file. Here is the hive.log file entries. 2010-08-14 17:32:24,929 INFO ql.Driver (Driver.java:execute(425)) - Starting command: select a,b,c,d,f as e,f*2 from testHiveJdbcDriverTable limit 1 2010-08-14 17:32:24,935 ERROR SessionState (SessionState.java:printError(277)) - PREHOOK: query: select a,b,c,d,f as e,f*2 from testHiveJdbcDriverTable limit 1 2010-08-14 17:32:24,937 ERROR SessionState (SessionState.java:printError(277)) - PREHOOK: type: QUERY 2010-08-14 17:32:24,938 ERROR SessionState (SessionState.java:printError(277)) - PREHOOK: Input: defa...@testhivejdbcdrivertable 2010-08-14 17:32:24,939 ERROR SessionState (SessionState.java:printError(277)) - PREHOOK: Output: file:/tmp/sean/hive_2010-08-14_17-32-24_783_1792704952460259887/-mr-10000 2010-08-14 17:32:24,940 INFO ql.Driver (SessionState.java:printInfo(268)) - Total MapReduce jobs = 1 2010-08-14 17:32:24,942 INFO ql.Driver (SessionState.java:printInfo(268)) - Launching Job 1 out of 1 2010-08-14 17:32:24,947 INFO exec.MapRedTask (SessionState.java:printInfo(268)) - Number of reduce tasks is set to 0 since there's no reduce operator 2010-08-14 17:32:24,967 INFO exec.MapRedTask (MapRedTask.java:execute(152)) - Generating plan file file:/tmp/sean/hive_2010-08-14_17-32-24_783_1792704952460259887/-local-10002/plan.xml 2010-08-14 17:32:25,588 INFO exec.MapRedTask (MapRedTask.java:execute(173)) - Executing: /home/sean/projects/hive/build/hadoopcore/hadoop-0.20.0/bin/hadoop jar /home/sean/projects/hive/build/ql/hive-exec-0.7.0.jar org.apache.hadoop.hive.ql.exec.ExecDriver -libjars file:///home/sean/projects/hive/build/jdbc/test/test-udfs.jar -plan file:/tmp/sean/hive_2010-08-14_17-32-24_783_1792704952460259887/-local-10002/plan.xml -nolog -jobconf datanucleus.connectionPoolingType=DBCP -jobconf hive.exec.script.allow.partial.consumption=false -jobconf hive.query.id=sean_20100814173232_f49d776c-7273-4995-8cc3-f7aceabdbde3 -jobconf hive.hwi.listen.port=9999 -jobconf hive.map.aggr=true -jobconf hive.map.aggr.hash.min.reduction=0.5 -jobconf datanucleus.plugin.pluginRegistryBundleCheck=LOG -jobconf hive.exec.reducers.bytes.per.reducer=1000000000 -jobconf hive.optimize.cp=true -jobconf hive.exec.dynamic.partition.mode=strict -jobconf hive.merge.size.smallfiles.avgsize=16000000 -jobconf datanucleus.cache.level2.type=SOFT -jobconf hive.exec.max.created.files=100000 -jobconf hive.script.serde=org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -jobconf hive.fileformat.check=true -jobconf hive.exec.max.dynamic.partitions.pernode=100 -jobconf hive.enforce.sorting=false -jobconf hive.optimize.ppd=true -jobconf hive.optimize.groupby=true -jobconf hive.enforce.bucketing=false -jobconf javax.jdo.option.ConnectionUserName=APP -jobconf hive.mapred.reduce.tasks.speculative.execution=true -jobconf mapred.job.name=select+a%2Cb%2Cc%2Cd%2Cf+as+e%2Cf*2+from+testHiveJ...1%28Stage-1%29 -jobconf javax.jdo.option.DetachAllOnCommit=true -jobconf hive.mapred.local.mem=0 -jobconf datanucleus.cache.level2=false -jobconf hive.session.id=sean_201008141732 -jobconf fs.pfile.impl=org.apache.hadoop.fs.ProxyLocalFileSystem -jobconf hive.script.operator.id.env.var=HIVE_SCRIPT_OPERATOR_ID -jobconf hive.archive.har.parentdir.settable=false -jobconf hadoop.job.ugi=sean%2Csean%2Cadm%2Cdialout%2Ccdrom%2Cplugdev%2Clpadmin%2Cadmin%2Csambashare -jobconf test.src.dir=file%3A%2F%2F%24%7Bbuild.dir%7D%2Fsrc%2Ftest -jobconf hive.metastore.server.max.threads=100000 -jobconf hive.udtf.auto.progress=false -jobconf hive.hwi.war.file=lib%2Fhive-hwi-%40VERSION%40.war -jobconf datanucleus.validateTables=false -jobconf hive.exec.compress.output=false -jobconf hive.test.mode.prefix=test_ -jobconf hive.mapjoin.bucket.cache.size=100 -jobconf test.log.dir=%24%7Bbuild.dir%7D%2Ftest%2Flogs -jobconf test.data.files=%24%7Buser.dir%7D%2F..%2Fdata%2Ffiles -jobconf datanucleus.validateConstraints=false -jobconf hive.metastore.server.tcp.keepalive=true -jobconf mapred.reduce.tasks=-1 -jobconf hive.query.string=select+a%2Cb%2Cc%2Cd%2Cf+as+e%2Cf*2+from+testHiveJdbcDriverTable+limit+1 -jobconf hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat -jobconf hive.task.progress=false -jobconf hive.jar.path=%24%7Bbuild.dir.hive%7D%2Fql%2Fhive-exec-%24%7Bversion%7D.jar -jobconf hive.metastore.ds.retry.interval=1000 -jobconf javax.jdo.option.ConnectionDriverName=org.apache.derby.jdbc.EmbeddedDriver -jobconf hive.skewjoin.mapjoin.map.tasks=10000 -jobconf hive.mapjoin.maxsize=100000 -jobconf hive.archive.enabled=false -jobconf hive.aux.jars.path=file%3A%2F%2F%2Fhome%2Fsean%2Fprojects%2Fhive%2Fbuild%2Fjdbc%2Ftest%2Ftest-udfs.jar -jobconf hive.exec.dynamic.partition=false -jobconf hive.exec.pre.hooks=org.apache.hadoop.hive.ql.hooks.PreExecutePrinter%2C+org.apache.hadoop.hive.ql.hooks.EnforceReadOnlyTables -jobconf hive.optimize.skewjoin=false -jobconf hive.groupby.mapaggr.checkinterval=100000 -jobconf hive.test.mode=false -jobconf hive.exec.parallel=false -jobconf hive.exec.counters.pull.interval=1000 -jobconf hive.default.fileformat=TextFile -jobconf hive.exec.max.dynamic.partitions=1000 -jobconf fs.har.impl=org.apache.hadoop.hive.shims.HiveHarFileSystem -jobconf hive.test.mode.samplefreq=32 -jobconf hive.metastore.ds.retry.attempts=1 -jobconf javax.jdo.option.NonTransactionalRead=true -jobconf hive.script.auto.progress=false -jobconf hive.merge.mapredfiles=false -jobconf javax.jdo.option.ConnectionURL=jdbc%3Aderby%3A%3BdatabaseName%3D..%2Fbuild%2Ftest%2Fjunit_metastore_db%3Bcreate%3Dtrue -jobconf hive.exec.compress.intermediate=false -jobconf hive.metastore.rawstore.impl=org.apache.hadoop.hive.metastore.ObjectStore -jobconf hive.map.aggr.hash.percentmemory=0.5 -jobconf hive.hwi.listen.host=0.0.0.0 -jobconf datanucleus.transactionIsolation=read-committed -jobconf hive.merge.size.per.task=256000000 -jobconf datanucleus.autoCreateSchema=true -jobconf hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.PostExecutePrinter -jobconf hive.groupby.skewindata=false -jobconf hive.metastore.local=true -jobconf hive.skewjoin.mapjoin.min.split=33554432 -jobconf hadoop.tmp.dir=%24%7Bbuild.dir.hive%7D%2Ftest%2Fhadoop-%24%7Buser.name%7D -jobconf hive.mapred.mode=nonstrict -jobconf hive.optimize.pruner=true -jobconf hive.skewjoin.key=100000 -jobconf hive.default.partition.name=__HIVE_DEFAULT_PARTITION__ -jobconf hive.hbase.wal.enabled=true -jobconf datanucleus.validateColumns=false -jobconf datanucleus.identifierFactory=datanucleus -jobconf hive.querylog.location=%24%7Bbuild.dir%7D%2Ftmp -jobconf hive.optimize.reducededuplication=true -jobconf hive.exec.reducers.max=999 -jobconf javax.jdo.PersistenceManagerFactoryClass=org.datanucleus.jdo.JDOPersistenceManagerFactory -jobconf hive.heartbeat.interval=1000 -jobconf hive.join.cache.size=25000 -jobconf hive.metastore.warehouse.dir=%24%7Btest.warehouse.dir%7D -jobconf datanucleus.autoStartMechanismMode=checked -jobconf javax.jdo.option.ConnectionPassword=mine -jobconf hive.metastore.connect.retries=5 -jobconf hive.exec.mode.local.auto=false -jobconf hive.mapjoin.cache.numrows=25000 -jobconf hive.exec.parallel.thread.number=8 -jobconf datanucleus.storeManagerType=rdbms -jobconf hive.script.recordreader=org.apache.hadoop.hive.ql.exec.TextRecordReader -jobconf hive.exec.scratchdir=%24%7Bbuild.dir%7D%2Fscratchdir -jobconf hive.metastore.metadb.dir=file%3A%2F%2F%24%7Bbuild.dir%7D%2Ftest%2Fdata%2Fmetadb%2F -jobconf hive.metastore.server.min.threads=200 -jobconf hive.script.recordwriter=org.apache.hadoop.hive.ql.exec.TextRecordWriter -jobconf hive.merge.mapfiles=true -jobconf hive.exec.script.maxerrsize=100000 -jobconf test.query.file1=file%3A%2F%2F%24%7Buser.dir%7D%2F..%2Fql%2Fsrc%2Ftest%2Forg%2Fapache%2Fhadoop%2Fhive%2Fql%2Finput2.q -jobconf hive.join.emit.interval=1000 -jobconf hive.added.jars.path= -jobconf mapred.system.dir=%2Fhome%2Fsean%2Fprojects%2Fhive%2Fbuild%2Ftest%2Fhadoop-sean%2Fmapred%2Fsystem%2F1753935500 -jobconf mapred.local.dir=%2Fhome%2Fsean%2Fprojects%2Fhive%2Fbuild%2Ftest%2Fhadoop-sean%2Fmapred%2Flocal%2F-2115807459 2010-08-14 17:32:25,633 ERROR exec.MapRedTask (MapRedTask.java:execute(224)) - Execution failed with exit status: 1 2010-08-14 17:32:25,635 ERROR ql.Driver (SessionState.java:printError(277)) - FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MapRedTask > Add support for JDBC PreparedStatements > --------------------------------------- > > Key: HIVE-1536 > URL: https://issues.apache.org/jira/browse/HIVE-1536 > Project: Hadoop Hive > Issue Type: Improvement > Components: Drivers > Reporter: Sean Flatley > > As a result of a Sprint which had us using Pentaho Data Integration with the > Hive database we have updated the driver. Many PreparedStatement methods > have been implemented. A patch will be attached tomorrow with a summary of > changes. > Note: A checkout of Hive/trunk was performed and the TestJdbcDriver test > cased was run. This was done before any modifications were made to the > checked out project. The testResultSetMetaData failed: > java.sql.SQLException: Query returned non-zero code: 9, cause: FAILED: > Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MapRedTask > at > org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:189) > at > org.apache.hadoop.hive.jdbc.TestJdbcDriver.testResultSetMetaData(TestJdbcDriver.java:530) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > at java.lang.reflect.Method.invoke(Method.java:597) > at junit.framework.TestCase.runTest(TestCase.java:154) > at junit.framework.TestCase.runBare(TestCase.java:127) > at junit.framework.TestResult$1.protect(TestResult.java:106) > at junit.framework.TestResult.runProtected(TestResult.java:124) > at junit.framework.TestResult.run(TestResult.java:109) > at junit.framework.TestCase.run(TestCase.java:118) > at junit.framework.TestSuite.runTest(TestSuite.java:208) > at junit.framework.TestSuite.run(TestSuite.java:203) > at > org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:420) > at > org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:911) > at > org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:768) > A co-worker did the same and the tests passed. Both environments were Ubuntu > and Hadoop version 0.20.2. > Tests added to the TestJdbcDriver by us were successful. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.