I'm not using Yarn, But I'll give that a shot.. 

Is hive 11 dependent on Yarn? 

JC


On Aug 23, 2013, at 1:19 PM, Justin Workman <justinjwork...@gmail.com> wrote:

> This looks similar to the problem you get when you don't set 
> yarn.resourcemanager.address. 
> 
> Sent from my iPhone
> 
> On Aug 23, 2013, at 11:12 AM, Jim Colestock <j...@ramblingredneck.com> wrote:
> 
>> Hello Justin, 
>> 
>> Thanks so much for the write up..  I'm compiling now.. I did find one error 
>> in your instructions: 
>> 
>> patch -p0 > HIVE-4619.D10971.1.path
>> 
>> I believe it should be: 
>> 
>> patch -p0 < HIVE-4619.D10971.1.patch
>> 
>> 
>> But alas, I still doesn't work for me.  I managed to get it to compile after 
>> using maven to install a zookeeper dependency.. But when I try to do 
>> anything in hive that requires an M/R job, I get the following: 
>> 
>> 2013-08-23 13:07:07,865 INFO org.apache.hadoop.mapred.TaskLogsTruncater: 
>> Initializing logs' truncater with mapRetainSize=-1 and reduceRetainSize=-1
>> 2013-08-23 13:07:07,867 WARN org.apache.hadoop.mapred.Child: Error running 
>> child
>> java.lang.RuntimeException: java.lang.NullPointerException
>>  at 
>> org.apache.hadoop.hive.ql.exec.Utilities.getMapRedWork(Utilities.java:230)
>>  at 
>> org.apache.hadoop.hive.ql.io.HiveInputFormat.init(HiveInputFormat.java:255)
>>  at 
>> org.apache.hadoop.hive.ql.io.HiveInputFormat.pushProjectionsAndFilters(HiveInputFormat.java:381)
>>  at 
>> org.apache.hadoop.hive.ql.io.HiveInputFormat.pushProjectionsAndFilters(HiveInputFormat.java:374)
>>  at 
>> org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getRecordReader(CombineHiveInputFormat.java:540)
>>  at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:394)
>>  at org.apache.hadoop.mapred.MapTask.run(MapTask.java:332)
>>  at org.apache.hadoop.mapred.Child$4.run(Child.java:268)
>>  at java.security.AccessController.doPrivileged(Native Method)
>>  at javax.security.auth.Subject.doAs(Subject.java:396)
>>  at 
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
>>  at org.apache.hadoop.mapred.Child.main(Child.java:262)
>> Caused by: java.lang.NullPointerException
>>  at 
>> org.apache.hadoop.hive.ql.exec.Utilities.getMapRedWork(Utilities.java:215)
>>  ... 11 more
>> 2013-08-23 13:07:07,871 INFO org.apache.hadoop.mapred.Task: Runnning cleanup 
>> for the task
>> 
>> Any thoughts? 
>> 
>> Thanks again, 
>> 
>> JC
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> On Aug 22, 2013, at 11:15 PM, Justin Workman <justinjwork...@gmail.com> 
>> wrote:
>> 
>>> For anyone interested, I have finished my blog on getting Hive-0.11.0 
>>> working with CDH4.3.0. It is a hack and can be made better but works like a 
>>> charm.
>>> 
>>> http://www.justinjworkman.com/big-data/hive-0-11-0-on-cloudera/
>>> 
>>> 
>>> 
>>> On Thu, Aug 22, 2013 at 8:29 AM, Jim Colestock <j...@ramblingredneck.com> 
>>> wrote:
>>> Justin, 
>>> 
>>> That would be much appreciated! 
>>> 
>>> Thanks Again, 
>>> 
>>> JC
>>> 
>>> On Aug 22, 2013, at 9:18 AM, Justin Workman <justinjwork...@gmail.com> 
>>> wrote:
>>> 
>>>> I do have this working on CHD4.3.0 with MR1 and Kerberos.  I am finishing 
>>>> up a blog today explaining what was needed to get it working. I will reply 
>>>> with the link later today. 
>>>> 
>>>> Justin 
>>>> 
>>>> Sent from my iPhone
>>>> 
>>>> On Aug 22, 2013, at 6:17 AM, Jim Colestock <j...@ramblingredneck.com> 
>>>> wrote:
>>>> 
>>>>> Hello All, 
>>>>> 
>>>>> Sorry for not replying to the thread, but I never received any of your 
>>>>> responses..  Hopefully this info will help: 
>>>>> 
>>>>> @kulkarni.swarnim
>>>>> 
>>>>> Here is the error I'm getting on task tracker for the mapper: 
>>>>> 
>>>>> 2013-08-21 15:34:32,491 INFO org.apache.hadoop.mapred.TaskLogsTruncater: 
>>>>> Initializing logs' truncater with mapRetainSize=-1 and reduceRetainSize=-1
>>>>> 2013-08-21 15:34:32,494 WARN org.apache.hadoop.mapred.Child: Error 
>>>>> running child
>>>>> java.lang.RuntimeException: java.lang.NullPointerException
>>>>>   at 
>>>>> org.apache.hadoop.hive.ql.exec.Utilities.getMapRedWork(Utilities.java:230)
>>>>>   at 
>>>>> org.apache.hadoop.hive.ql.io.HiveInputFormat.init(HiveInputFormat.java:255)
>>>>>   at 
>>>>> org.apache.hadoop.hive.ql.io.HiveInputFormat.pushProjectionsAndFilters(HiveInputFormat.java:381)
>>>>>   at 
>>>>> org.apache.hadoop.hive.ql.io.HiveInputFormat.pushProjectionsAndFilters(HiveInputFormat.java:374)
>>>>>   at 
>>>>> org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getRecordReader(CombineHiveInputFormat.java:540)
>>>>>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:394)
>>>>>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:332)
>>>>>   at org.apache.hadoop.mapred.Child$4.run(Child.java:268)
>>>>>   at java.security.AccessController.doPrivileged(Native Method)
>>>>>   at javax.security.auth.Subject.doAs(Subject.java:396)
>>>>>   at 
>>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
>>>>>   at org.apache.hadoop.mapred.Child.main(Child.java:262)
>>>>> Caused by: java.lang.NullPointerException
>>>>>   at 
>>>>> org.apache.hadoop.hive.ql.exec.Utilities.getMapRedWork(Utilities.java:215)
>>>>>   ... 11 more
>>>>> 2013-08-21 15:34:32,498 INFO org.apache.hadoop.mapred.Task: Runnning 
>>>>> cleanup for the task
>>>>> 
>>>>> @Srinivas Surasani
>>>>>   Hopefully the above helps.. 
>>>>> 
>>>>> @Lars Francke
>>>>>   Thanks, I'll take a peek at those 2 patches.. But I've just recently 
>>>>> pulled down fresh code and recompiled everything and I'm still getting 
>>>>> the above error. 
>>>>> 
>>>>> And for reference, below is my original email to the list..  Thanks all 
>>>>> for you help! 
>>>>> 
>>>>> 
>>>>> JC
>>>>> 
>>>>> 
>>>>> 
>>>>> Hello All, 
>>>>> 
>>>>> Has anyone been successful at running hive 0.11 with Cloudera CDH 4.3?   
>>>>> 
>>>>> I've been able to get hive to connect to my metadb (which is in 
>>>>> Postgres).  Verified by doing a show tables..  I can run explain and 
>>>>> describes on tables, but when I try to run anything that fires off an M/R 
>>>>> job, I get the following error: 
>>>>> 
>>>>> hive>select count(*) from tableA;
>>>>> Total MapReduce jobs = 1
>>>>> Launching Job 1 out of 1
>>>>> Number of reduce tasks determined at compile time: 1
>>>>> In order to change the average load for a reducer (in bytes):
>>>>>   set hive.exec.reducers.bytes.per.reducer=<number>
>>>>> In order to limit the maximum number of reducers:
>>>>>   set hive.exec.reducers.max=<number>
>>>>> In order to set a constant number of reducers:
>>>>>   set mapred.reduce.tasks=<number>
>>>>> Starting Job = job_201307112247_13816, Tracking URL = 
>>>>> http://master:50030/jobdetails.jsp?jobid=job_201307112247_13816
>>>>> Kill Command = /usr/lib/hadoop/bin/hadoop job  -kill 
>>>>> job_201307112247_13816
>>>>> Hadoop job information for Stage-1: number of mappers: 1; number of 
>>>>> reducers: 1
>>>>> 2013-07-12 02:11:42,829 Stage-1 map = 0%,  reduce = 0%
>>>>> 2013-07-12 02:12:08,173 Stage-1 map = 100%,  reduce = 100%
>>>>> Ended Job = job_201307112247_13816 with errors
>>>>> Error during job, obtaining debugging information...
>>>>> Job Tracking URL: 
>>>>> http://master:50030/jobdetails.jsp?jobid=job_201307112247_13816
>>>>> Examining task ID: task_201307112247_13816_m_000002 (and more) from job 
>>>>> job_201307112247_13816
>>>>> Exception in thread "Thread-19" java.lang.NoClassDefFoundError: 
>>>>> org/apache/hadoop/mapreduce/util/HostUtil
>>>>> at 
>>>>> org.apache.hadoop.hive.shims.Hadoop23Shims.getTaskAttemptLogUrl(Hadoop23Shims.java:61)
>>>>> at 
>>>>> org.apache.hadoop.hive.ql.exec.JobDebugger$TaskInfoGrabber.getTaskInfos(JobDebugger.java:186)
>>>>> at 
>>>>> org.apache.hadoop.hive.ql.exec.JobDebugger$TaskInfoGrabber.run(JobDebugger.java:142)
>>>>> at java.lang.Thread.run(Thread.java:619)
>>>>> Caused by: java.lang.ClassNotFoundException: 
>>>>> org.apache.hadoop.mapreduce.util.HostUtil
>>>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
>>>>> at java.security.AccessController.doPrivileged(Native Method)
>>>>> at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:307)
>>>>> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:248)
>>>>> ... 4 more
>>>>> FAILED: Execution Error, return code 2 from 
>>>>> org.apache.hadoop.hive.ql.exec.MapRedTask
>>>>> MapReduce Jobs Launched:
>>>>> Job 0: Map: 1  Reduce: 1   HDFS Read: 0 HDFS Write: 0 FAIL
>>>>> Total MapReduce CPU Time Spent: 0 msec
>>>>> 
>>>>> I'm using my configs from hive 0.10, which works with no issues and this 
>>>>> was pretty much a drop in replacement on the machine that hadoop 0.10 was 
>>>>> running on.. 
>>>>> 
>>>>> I've done a bunch of googling around and have found a bunch of other 
>>>>> folks that have have had the same issue, but no solid answers.. 
>>>>> 
>>>>> Thanks in advance for any help.. 
>>>>> 
>>>>> JC
>>>>> 
>>>>> 
>>> 
>>> 
>> 

Reply via email to